Re: [Computer-go] 30% faster with a batch size of 63 instead of 64!

2020-05-09 Thread David Wu
Very nice. :) And thanks for the note about batch sizing. Specifically tuning parameters for this level of strength on 9x9 seems like it could be quite valuable, Kata definitely hasn't done that either. But it feels like bots are very very close to optimal on 9x9. With some dedicated work, more mo

Re: [Computer-go] 30% faster with a batch size of 63 instead of 64!

2020-05-09 Thread uurtamo .
Nice job! And the graph makes it super clear how the edge effects work. s. On Sat, May 9, 2020, 2:19 PM Rémi Coulom wrote: > Hi, > > I am probably not the only one who made this mistake: it is usually very > bad to use a power of 2 for the batch size! > > Relevant documentation by NVIDIA: > > h

Re: [Computer-go] 30% faster with a batch size of 63 instead of 64!

2020-05-09 Thread Rémi Coulom
Yeh! first win against Kata! http://www.yss-aya.com/cgos/viewer.cgi?9x9/SGF/2020/05/09/999849.sgf In addition to the optimized batch size, I did two other things: - I use two batches of 63 instead of one, with double buffering, so that the GPU is kept 100% busy. About 14k nodes per second now

[Computer-go] 30% faster with a batch size of 63 instead of 64!

2020-05-09 Thread Rémi Coulom
Hi, I am probably not the only one who made this mistake: it is usually very bad to use a power of 2 for the batch size! Relevant documentation by NVIDIA: https://docs.nvidia.com/deeplearning/performance/dl-performance-convolutional/index.html#quant-effects The documentation is not extremely cle