Daniel Shawul wrote:... but your method does tend to give bad means too just to lower its variance because of assumptions it makes for the tree.
Not true. The method I and Michel propose will not give "bad means". Their mean is
exactly equal to the true size of the tree, no matter how wrong the assumptions were that went into deriving the p_i. Wrong assumptions just mean that you converge to that correct mean more slowly than with correct assumptions. But you cannot avoid making assumptions. Making 'no assumptions' on the relative sub-tree sizes is equivalent to the assumption that they are on averageof equal size(for which the optimum would be homogeneous sampling.) That assumption can be just as wrong as any other. And inparticular with LMR'ed trees it is a very inferior assumption.
At ad infinitum the variance of all methods will be zero, so surely we are comparing at fixed number of games.
The
variance will be zero, but not necessarily the
error. You are proposing methods (with p_i*w_i != 1) that in general will give the wrong result even after infinite sampling.
What good is that? It is a fair question because at any time t when you stop the iteration the reliablity of our mean has dropped to get a result with smaller variance.
That is why I don't say your method is biased or unbiased, but it can really blunder because it assumes a whole lot about the tree and will have difficulties to revive from extreme cases.. Its estimate of the mean are going to be bad because of too many assuumptions, solely to get the variance down. You are assuming the heuristic is going to be perfect but it won't , and when it fails its estimate will be worse than a biased estimator like mine.