This is somewhat moot - if any moves had been significantly and obviously
weak to any observers, the results wouldn't have been 4-1.

I.e. One bad move out of 5 games would give roughly the same strength
information as one loss out of 5 games; consider that the kibitzing was
being done in real time.

s.
On Mar 22, 2016 11:08 AM, "Jim O'Flaherty" <jim.oflaherty...@gmail.com>
wrote:

> I think you are reinforcing Simon's original point; i.e. using a more fine
> grained approach to statically approximate AlphaGo's ELO where fine grained
> is degree of vetting per move and/or a series of moves. That is a
> substantially larger sample size and each sample will have a pretty high
> degree of quality (given the vetting is being done by top level
> professionals).
> On Mar 22, 2016 1:04 PM, "Jeffrey Greenberg" <je...@inventivity.com>
> wrote:
>
>> Given the minimal sample size, bothering over this question won't amount
>> to much. I think the proper response is that no one thought we'd see this
>> level of play at this point in our AI efforts and point to the fact that we
>> witnessed hundreds of moves vetted by 9dan players, especially Michael
>> Redmond's, where each move was vetted. In other words "was the level of
>> play very high?" versus the question "have we beat all humans". The answer
>> is more or less, yes.
>>
>> On Tuesday, March 22, 2016, Lucas, Simon M <s...@essex.ac.uk> wrote:
>>
>>> Hi all,
>>>
>>> I was discussing the results with a colleague outside
>>> of the Game AI area the other day when he raised
>>> the question (which applies to nearly all sporting events,
>>> given the small sample size involved)
>>> of statistical significance - suggesting that on another week
>>> the result might have been 4-1 to Lee Sedol.
>>>
>>> I pointed out that in games of skill there's much more to judge than
>>> just the final
>>> outcome of each game, but wondered if anyone had any better (or worse :)
>>> arguments - or had even engaged in the same type of
>>> conversation.
>>>
>>> With AlphaGo winning 4 games to 1, from a simplistic
>>> stats point of view (with the prior assumption of a fair
>>> coin toss) you'd not be able to claim much statistical
>>> significance, yet most (me included) believe that
>>> AlphaGo is a genuinely better Go player than Lee Sedol.
>>>
>>> From a stats viewpoint you can use this approach:
>>> http://www.inference.phy.cam.ac.uk/itprnn/book.pdf
>>> (see section 3.2 on page 51)
>>>
>>> but given even priors it won't tell you much.
>>>
>>> Anyone know any good references for refuting this
>>> type of argument - the fact is of course that a game of Go
>>> is nothing like a coin toss.  Games of skill tend to base their
>>> outcomes on the result of many (in the case of Go many hundreds of)
>>> individual actions.
>>>
>>> Best wishes,
>>>
>>>   Simon
>>>
>>>
>>> _______________________________________________
>>> Computer-go mailing list
>>> Computer-go@computer-go.org
>>> http://computer-go.org/mailman/listinfo/computer-go
>>
>>
>> _______________________________________________
>> Computer-go mailing list
>> Computer-go@computer-go.org
>> http://computer-go.org/mailman/listinfo/computer-go
>>
>
> _______________________________________________
> Computer-go mailing list
> Computer-go@computer-go.org
> http://computer-go.org/mailman/listinfo/computer-go
>
_______________________________________________
Computer-go mailing list
Computer-go@computer-go.org
http://computer-go.org/mailman/listinfo/computer-go

Reply via email to