Hi,
Our program Natsukaze also used Leela Zero recent 700000 selfplay games
to train DNN.
Ladder escape moves(4% of total games) are removed, and chasing not
ladder(0.3%) also removed. But its DNN policy was weak, around CGOS 2100.
Maybe it is because current LZ selfplay use t=1 not first 30 moves but all
moves.
I did not know this. I think this makes selfplay weaker +1000 Elo.
Switch to t=1 for all self-play moves, i.e., randomcnt=999
https://github.com/gcp/leela-zero-server/pull/81
Thanks,
Hiroshi Yamashita
On 2018/12/19 2:01, Gian-Carlo Pascutto wrote:
On 17/12/18 01:53, Hiroshi Yamashita wrote:
Hi,
AI Ryusei 2018 was held on 15,16th December in Nihon-kiin, Japan.
14 programs played preliminary swiss 7 round, and top 6 programs
played round-robin final. Then, Golaxy won.
Result
https://www.igoshogi.net/ai_ryusei/01/en/result.html
It appears the 2nd place finisher after Golaxy was a hybrid of Rn and
Leela Zero, using rollouts to compensate for Leela's network being
trained with the "wrong" komi for this competition:
https://github.com/zakki/Ray/issues/171#issuecomment-447637052
https://img.igoshogi.net/ai_ryusei/01/data/11.pdf
_______________________________________________
Computer-go mailing list
Computer-go@computer-go.org
http://computer-go.org/mailman/listinfo/computer-go