Maybe I did no explain my point well enough.

The problem with infinite is that we get a better approximation to a wrong value.

With few simulations we get that value with, say 1/10 error. With an astronomical amount of simulations we get the same value with an error of 1e-200, but it's still wrong!. It is proved that simulating a go position converges, but it does not converge to the same value as if the position was played by perfect players, it only converges to the asymptotic
limit of random play.

I am not an MC developer, but as far as I know, UCT keeps a limited (i.e. n-ply) tree in memory and intentionally unbiasses the nodes to make the convergence faster, that
does not change anything, assuming constant tree size.

A simple test :
1: after 100 simulations, choose the highest number in (0.96, 2.1, 3.18)
2: after 1e9 simulations, choose it in (0.9999999, 2.0000001, 3.000001)
You chose the same value (= same move).

That's why, I insist, if you don't increase the size of the tree and only get a better approximation to a wishful but frequently misconceived value (the limit of random play) witch is *not* a good evaluation of the game, you don't significantly improve your play. Of course, if you increase the tree, you reach perfect play, that's not
the point.

Jacques.
_______________________________________________
computer-go mailing list
computer-go@computer-go.org
http://www.computer-go.org/mailman/listinfo/computer-go/

Reply via email to