Given a game state S and the result of the game R in { -1.0 (loss) , 0.0 (tie), 1.0 (win) }, is the value head trained to predict R?brianr wrote: ↑Sun Feb 03, 2019 1:45 amIn terms of the actual net training process, Leela learns pretty much the same way with PGNs. The major practical difference is that the PGN game positions must first be converted into the input bit planes which are fed to the Leela net training code (Python and TensorFlow). Fortunately, the sample dataset (11GB compressed) includes both the PGNs and the bit plane files (2MM, IIRC, one file for each game), so it is much easier. The early versions of the SL code to convert PGN games to the bit plane format were a bit quirky, although I think there are some forks that might have fixed it. With the sample dataset that part can be skipped.
Generally, when learning from self-play games it is reinforcement learning (RL), and from PGNs it is supervised learning (SL). Which PGN games are selected and the order in which they are fed to the training process can have a significant impact on the resulting net strength. SL can train a net faster, but there is some debate about the ultimate result being as strong as RL. The massive crowd-sourced Leela training effort is primarily to provide self-play games for RL, and some match testing. Unfortunately, early on that was also called "training", which is not the same thing as using Tensorflow to actually train the weights in the net. Quite a few folks are also experimenting with SL.
So, how many of you are working on neural networks for chess?
Moderators: hgm, Rebel, chrisw
-
- Posts: 25
- Joined: Mon Dec 10, 2018 2:45 pm
- Full name: jack d.
Re: So, how many of you are working on neural networks for chess?
-
- Posts: 536
- Joined: Thu Mar 09, 2006 3:01 pm
Re: So, how many of you are working on neural networks for chess?
I think so, yes.
And the policy head predicts the move with the best winning chances.
How the value and policy heads work together is something I need to understand better.
-
- Posts: 2658
- Joined: Wed Mar 10, 2010 10:18 pm
- Location: Hamburg, Germany
- Full name: Srdja Matovic
Re: So, how many of you are working on neural networks for chess?
Hi Daniel, can you elaborate a bit on your qsearch?Daniel Shawul wrote: ↑Sat Feb 02, 2019 4:15 pm
I have added policy networks now which seems to have helped quite a bit.
They seem to add a lot of positional knowledge besides making my inference faster (used to do qsearch for policy before)
...
Do you think LC0 could profit tactically by adding a qsearch?
--
Srdja