On 21/03/2017 21:08, David Ongaro wrote:
>> But how would you fix it? Isn't that you'd need to retrain your value
>> network from the scratch?
> 
> I would think so as well. But I some months ago I already made a
> proposal in this list to mitigate that problem: instead of training a
> different value network for each Komi, add a “Komi adjustment” value as
> input during the training phase. That should be much more effective,
> since the “win/lost” evaluation shouldn’t change for many (most?)
> positions for small adjustments but the resulting value network (when
> trained for different Komi adjustments) has a much greater range of
> applicability.

The problem is not the training of the network itself (~2-4 weeks of
letting a program someone else wrote run in the background, easiest
thing ever in computer go), or whether you use a komi input or a
separate network, the problem is getting data for the different komi values.

Note that if getting data is not a problem, then a separate network
would perform better than your proposal.

-- 
GCP
_______________________________________________
Computer-go mailing list
Computer-go@computer-go.org
http://computer-go.org/mailman/listinfo/computer-go

Reply via email to