> > What we can say from experiments is that the scaling with time is very
> > good
> > with few simulations, but becomes less interesting with a lot of
> > simulations.
>
> This is typical for statistical sampling. The variance of the sample mean
> is a function of the square-root of the sampling size.
> [....]
> In case of UCT the sample-bias is on purpose. I do not know if this
> accelerates convergence or introduces a systematic bias. A better
> performance on a finite sample sizes does not mean that the sampling
> converges in the limit to the true value. I also do not know if the result
> measured at the end-position has some bias.

Here we don't have to forget that MC estimation is only at the leaf of the UCT 
tree. If there were no tree, of course, only few thousands of simulations 
would be enough to estimate the average.
Here we don't try to estimate the expectation of a random variable using more 
sample, because after each simulation, the random variable is not the same.
So, using UCT, the important thing is how deep is the tree, and if the 
predicted moves are the good ones, much more than variance analysis. If the 
MC returns bad values, the tree will go deeper, but will be meaningless. 
As said Magnus, at the end of the game, MC is quite good and then the tree 
gives good moves and with a lot of simulations becomes very good. At the 
beginning, if you give a lot of simulations, the tree can predict bad moves 
due to the bad MC evaluation. Of course, asymptotically it will converge 
towards the very best move, but we all know that we never are at the 
asymptotic :-).

Sylvain

_______________________________________________
computer-go mailing list
computer-go@computer-go.org
http://www.computer-go.org/mailman/listinfo/computer-go/

Reply via email to