The mini study so far ... games against gnugo-3.7.11 PLAYER TIME/GME RATING GAMES WIN% ------------ -------- ------- ----- ------- Mogo_01 0.10 1002.4 189 0.53 Mogo at 64 play-outs Mogo_02 0.14 1197.0 156 2.56 Mogo at 128 play-outs Mogo_03 0.17 1445.0 150 11.33 Mogo at 256 play-outs
Rémi Coulom wrote: > I believe the main problem is that the Elo-rating model is wrong for > bots. The phenomenon with Mogo is probably the same as Crazy Stone: if > there are enough strong MC bots playing to shield the top MC programs > from playing against GNU, then they'll get a high rating because they > are efficient at beating other MC bots. Otherwise, they are forced to > play against GNU, and lose points. > > For instance: > http://www.lri.fr/~teytaud/cross/CS-9-17-2CPU.html > GNU 1946 22 / 27 81.48 > GnuCvs-10 1969 26 / 31 83.87 > AyaMC637_4CPU 2108 18 / 19 94.74 > > > A very easy way to get over-evaluated on CGOS is to have two versions > of the same program that play each other. For instance, if I connect > CS-2CPU and CS-8CPU, they will play most of their games against each > other, ans CS-8CPU will get an incredible rating. > > Just incorporate GNU in Don's scalability study, and the rating range > will shrink a lot. > > Rémi > _______________________________________________ > computer-go mailing list > computer-go@computer-go.org > http://www.computer-go.org/mailman/listinfo/computer-go/ > _______________________________________________ computer-go mailing list computer-go@computer-go.org http://www.computer-go.org/mailman/listinfo/computer-go/