The mini study so far ...  games against gnugo-3.7.11

PLAYER        TIME/GME   RATING  GAMES     WIN%  
------------  --------  -------  -----  -------
Mogo_01           0.10   1002.4    189     0.53   Mogo at 64 play-outs
Mogo_02           0.14   1197.0    156     2.56   Mogo at 128 play-outs
Mogo_03           0.17   1445.0    150    11.33   Mogo at 256 play-outs




Rémi Coulom wrote:
> I believe the main problem is that the Elo-rating model is wrong for
> bots. The phenomenon with Mogo is probably the same as Crazy Stone: if
> there are enough strong MC bots playing to shield the top MC programs
> from playing against GNU, then they'll get a high rating because they
> are efficient at beating other MC bots. Otherwise, they are forced to
> play against GNU, and lose points.
>
> For instance:
> http://www.lri.fr/~teytaud/cross/CS-9-17-2CPU.html
> GNU     1946     22 / 27     81.48
> GnuCvs-10     1969     26 / 31     83.87
> AyaMC637_4CPU     2108     18 / 19     94.74
>
>
> A very easy way to get over-evaluated on CGOS is to have two versions
> of the same program that play each other. For instance, if I connect
> CS-2CPU and CS-8CPU, they will play most of their games against each
> other, ans CS-8CPU will get an incredible rating.
>
> Just incorporate GNU in Don's scalability study, and the rating range
> will shrink a lot.
>
> Rémi
> _______________________________________________
> computer-go mailing list
> computer-go@computer-go.org
> http://www.computer-go.org/mailman/listinfo/computer-go/
>
_______________________________________________
computer-go mailing list
computer-go@computer-go.org
http://www.computer-go.org/mailman/listinfo/computer-go/

Reply via email to