So, how many of you are working on neural networks for chess?

Discussion of chess software programming and technical issues.

Moderators: hgm, Rebel, chrisw

So, how many of you are working on neural networks for chess?

Poll ended at Mon Mar 04, 2019 7:03 pm

Yep - i am in
0
No votes
No - nothing for me
2
100%
Hm - still waiting for...
0
No votes
 
Total votes: 2

jackd
Posts: 25
Joined: Mon Dec 10, 2018 2:45 pm
Full name: jack d.

Re: So, how many of you are working on neural networks for chess?

Post by jackd »

brianr wrote: Sun Feb 03, 2019 1:45 am
jackd wrote: Sat Feb 02, 2019 7:52 pm Could you summarize how Leela trains off pgns?
In terms of the actual net training process, Leela learns pretty much the same way with PGNs. The major practical difference is that the PGN game positions must first be converted into the input bit planes which are fed to the Leela net training code (Python and TensorFlow). Fortunately, the sample dataset (11GB compressed) includes both the PGNs and the bit plane files (2MM, IIRC, one file for each game), so it is much easier. The early versions of the SL code to convert PGN games to the bit plane format were a bit quirky, although I think there are some forks that might have fixed it. With the sample dataset that part can be skipped.

Generally, when learning from self-play games it is reinforcement learning (RL), and from PGNs it is supervised learning (SL). Which PGN games are selected and the order in which they are fed to the training process can have a significant impact on the resulting net strength. SL can train a net faster, but there is some debate about the ultimate result being as strong as RL. The massive crowd-sourced Leela training effort is primarily to provide self-play games for RL, and some match testing. Unfortunately, early on that was also called "training", which is not the same thing as using Tensorflow to actually train the weights in the net. Quite a few folks are also experimenting with SL.
Given a game state S and the result of the game R in { -1.0 (loss) , 0.0 (tie), 1.0 (win) }, is the value head trained to predict R?
brianr
Posts: 536
Joined: Thu Mar 09, 2006 3:01 pm

Re: So, how many of you are working on neural networks for chess?

Post by brianr »

jackd wrote: Sun Feb 03, 2019 4:20 am Given a game state S and the result of the game R in { -1.0 (loss) , 0.0 (tie), 1.0 (win) }, is the value head trained to predict R?
I think so, yes.
And the policy head predicts the move with the best winning chances.
How the value and policy heads work together is something I need to understand better.
smatovic
Posts: 2645
Joined: Wed Mar 10, 2010 10:18 pm
Location: Hamburg, Germany
Full name: Srdja Matovic

Re: So, how many of you are working on neural networks for chess?

Post by smatovic »

Daniel Shawul wrote: Sat Feb 02, 2019 4:15 pm
I have added policy networks now which seems to have helped quite a bit.
They seem to add a lot of positional knowledge besides making my inference faster (used to do qsearch for policy before)
...
Hi Daniel, can you elaborate a bit on your qsearch?

Do you think LC0 could profit tactically by adding a qsearch?

--
Srdja