Hi,

In the paper you only present results of UCT_RAVE with the MoGo
default policy. Did you run tests with UCT_RAVE using "pure" random
playouts too?

I'm curious because I've tried millions ( well, it feels that way ) of
uses for AMAF in my code... but so far all of them have been proven
useless, often yielding worse results.

/Christian Nilsson


On 6/23/07, Sylvain Gelly <[EMAIL PROTECTED]> wrote:
Hello all,

We just presented our paper describing MoGo's improvements at ICML,
and we thought we would pass on some of the feedback and corrections
we have received.
(http://www.machinelearning.org/proceedings/icml2007/papers/387.pdf)

The way that we incorporate prior knowledge in UCT can be seen as a
bayesian prior, and corresponds exactly to the dirichlet prior (more
precisely to the beta prior as we here get binomials).

The cumulative result is only given using the prior knowledge on top
of RAVE, but it could have been done the other way round and give the
same type of results. Each particular improvement is somehow
independent of the others.

On figure 5, the legend of horizontal axis should be m_prior rather
than n_prior.

All experiments (except the default policy) were played against GnuGo
level 10, not level 8.

Any other comments are welcome!
Sylvain & David
_______________________________________________
computer-go mailing list
computer-go@computer-go.org
http://www.computer-go.org/mailman/listinfo/computer-go/

_______________________________________________
computer-go mailing list
computer-go@computer-go.org
http://www.computer-go.org/mailman/listinfo/computer-go/

Reply via email to