On Tue, 2009-02-17 at 23:29 -0800, David Fotland wrote:
> It's not true that MCTS only goes a few ply.  In 19x19 games on 32 CPU
> cores, searching about 3 million play outs per move, Many Faces of Go
> typically goes over 15 ply in the PV in the UCT tree.

That's what I meant when I said a few ply.  Lazarus doesn't get that
deep but it also goes several ply.   A few years ago I got the
impression that many people felt that anything other than local search
was completely worthless because you couldn't go the 100+ ply you needed
to understand a lot of things.    Perhaps I exaggerate,  but that was
the gist.  

It's pretty clear of course that you cannot cover these 15 ply (or
whatever) with anything resembling a brute force search, where other
than alpha/beta pruning you look at every move to some goal depth with
perhaps some kind of brief quies search.   That's probably what the
previous list member meant.


> 
> I agree that it is much easier to reliably prune bad moves in go than
> it is
> in chess.
> 
> Many Faces (pre MCTS) was a selective alpha-beta searcher.  Switching
> to
> MCTS made it much stronger, so I think MCTS is better than alpha-beta
> for
> go.  The reason is that to get any depth out of alpha-beta I had to
> hard-prune most moves, and misses too many good moves.  MCTS is full
> width,
> so misses nothing, and still searches twice as deep as the pruned
> alpha-beta.  Perhaps a dumber but faster evaluation function could do
> better
> than I did though.

This may be the case.   I believe that MCTS is just easier to think
about and manage and I just have a hunch that a very similar thing could
be formulated within the alpha/beta framework.   Even if I'm right,
there may be no benefit to doing this if they are (as I tentatively
believe) more or less equivalent.    But alpha/beta has one advantage,
that perhaps could be taken advantage of,  and that is the hard pruning
you talk about.   Even though it's "hard pruning" in some sense, it's
fully admissible pruning and there are parts of the tree that MCTS
always looks at, but that alpha/beta knows can be correctly rejected.
Of course I appreciate the MCTS doesn't waste much time on those, but
maybe it wastes enough time to make a difference,  I don't know. 

I think of a proper alpha/beta search as full width too, it is just
iterative.    There is no move that gets un-admissibly and permanently
pruned in alpha/beta even with modern techniques like null move and late
move reductions.    Those moves are effectively picked up and explored
to greater depths on the NEXT iterations.    Unless of course they are
outside the alpha/beta window in which case it doesn't matter.  So
modern alpha/beta to me is looking more and more like the TS in MCTS,
without the MC with at least one possible advantage.   

MCTS is really easy to reason about and control however and in
comparison alpha/beta is temperamental.  If some move in the tree that
looks bad is really good,  MCTS explores it gradually and in a
controlled way and alpha/beta has to wait much longer to modify it's
decision.   And it's so well integrated with Monte Carlo and it's not
clear how to do that so nicely with alpha/beta.   

So I am only open to the  possibility that they are equivalent in some
implementable sense, but I may be wrong about this.

- Don
 



> 
> Davdi

_______________________________________________
computer-go mailing list
computer-go@computer-go.org
http://www.computer-go.org/mailman/listinfo/computer-go/

Reply via email to