Hi All,

I wanted to share an update to a post I wrote last year about using the AlphaGo 
Zero algorithm on small boards (7x7). I train for approximately 2 months on a 
single desktop PC with 2 GPU cards.

In the article I was getting mediocre performance from the networks. Now, I've 
found that there was a bug in the way that I was evaluating the networks and 
that what I've been training seems to be matching GNU Go's level of performance.

Anyway, I'm aware I'm not exactly pushing the bounds of what's been done 
before, but I thought some might be interested to see how one can still get 
decent performance, at least in my opinion, on extremely limited hardware 
setups -- orders of magnitude less than what DeepMind (and Leela) have used.

The post where I talk about the model's performance, training, and setup:
https://medium.com/@cody2007.2/how-i-trained-a-self-supervised-neural-network-to-beat-gnugo-on-small-7x7-boards-6b5b418895b7

A video where I play the network and show some of its move probabilities during 
self-play games:
https://www.youtube.com/watch?v=a5vq1OjZrCU

The model weights and tensorflow code:
https://github.com/cody2007/alpha_go_zero_implementation

-Cody
_______________________________________________
Computer-go mailing list
Computer-go@computer-go.org
http://computer-go.org/mailman/listinfo/computer-go

Reply via email to