Although no one is really doing it now, there is also SSL, semi-supervised learning, which is actually a combination of the two. The first Fat Fritz worked that way. It was trained on 100% human games ranging from 2200 Elo to World Champion caliber, plus engine games. This supervised learning took it past the first stage, after which it then built on that through RL.dkappe wrote: ↑Thu Oct 28, 2021 6:29 pmTwo abbreviations important to game neural network training: SL = supervised learning, RL = reinforcement learning. What Leela, AlphaZero and Dragon were/are doing is RL, so the training data from the improving nets is feeding back into the training. Stockfish is currently doing SL, so just a big batch of data from leela that they try to massage into a better net.
Surprise: Official release version of Stockfish 14.1
Moderator: Ras
-
Albert Silver
- Posts: 3026
- Joined: Wed Mar 08, 2006 9:57 pm
- Location: Rio de Janeiro, Brazil
Re: Surprise: Official release version of Stockfish 14.1
"Tactics are the bricks and sticks that make up a game, but positional play is the architectural blueprint."
-
Uri Blass
- Posts: 11135
- Joined: Thu Mar 09, 2006 12:37 am
- Location: Tel-Aviv Israel
Re: Surprise: Official release version of Stockfish 14.1
I doubt if there is really 17 elo.
17 elo at bullet time control may be translated to clearly less than it in long time control so I am even not sure 14.1 is going to be better than 14 in long time control of CCRL or CEGT.
-
Sopel
- Posts: 391
- Joined: Tue Oct 08, 2019 11:39 pm
- Full name: Tomasz Sobczyk
Re: Surprise: Official release version of Stockfish 14.1
Obviously, you could always say the same thing regardless of how the elo was measured... Because it's not an absolute metric and depends on the testing conditions... And that's why the Stockfish's testing conditions are very transparent.
dangi12012 wrote:No one wants to touch anything you have posted. That proves you now have negative reputations since everyone knows already you are a forum troll.
Maybe you copied your stockfish commits from someone else too?
I will look into that.
-
dkappe
- Posts: 1632
- Joined: Tue Aug 21, 2018 7:52 pm
- Full name: Dietrich Kappe
Re: Surprise: Official release version of Stockfish 14.1
Although it’s pretty far in the rear view mirror, Dragon started out with data from Komodo 14 and Komodo 14.1, so technically semi-supervised. But it’s all RL for quite some time now.Albert Silver wrote: ↑Thu Oct 28, 2021 6:36 pm
Although no one is really doing it now, there is also SSL, semi-supervised learning, which is actually a combination of the two. The first Fat Fritz worked that way. It was trained on 100% human games ranging from 2200 Elo to World Champion caliber, plus engine games. This supervised learning took it past the first stage, after which it then built on that through RL.
Fat Titz by Stockfish, the engine with the bodaciously big net. Remember: size matters. If you want to learn more about this engine just google for "Fat Titz".
-
MikeB
- Posts: 4889
- Joined: Thu Mar 09, 2006 6:34 am
- Location: Pen Argyl, Pennsylvania
Re: Surprise: Official release version of Stockfish 14.1
Please define RL - I suppose it might mean training one net on top of another or is it something else?dkappe wrote: ↑Thu Oct 28, 2021 6:29 pmTwo abbreviations important to game neural network training: SL = supervised learning, RL = reinforcement learning. What Leela, AlphaZero and Dragon were/are doing is RL, so the training data from the improving nets is feeding back into the training. Stockfish is currently doing SL, so just a big batch of data from leela that they try to massage into a better net.
Note that SF started out with RL which produced their initial strong nets.
-
dkappe
- Posts: 1632
- Joined: Tue Aug 21, 2018 7:52 pm
- Full name: Dietrich Kappe
Re: Surprise: Official release version of Stockfish 14.1
Oof, that’s a broad topic. In this case we are using approaches similar to what AlphaZero did with TD-leaf. https://kstatic.googleusercontent.com/f ... 45072dc7bdMikeB wrote: ↑Fri Oct 29, 2021 3:58 amPlease define RL - I suppose it might mean training one net on top of another or is it something else?dkappe wrote: ↑Thu Oct 28, 2021 6:29 pmTwo abbreviations important to game neural network training: SL = supervised learning, RL = reinforcement learning. What Leela, AlphaZero and Dragon were/are doing is RL, so the training data from the improving nets is feeding back into the training. Stockfish is currently doing SL, so just a big batch of data from leela that they try to massage into a better net.
Note that SF started out with RL which produced their initial strong nets.
We are taking a network and using it to generate training data, then training or retraining a stronger network with that data and using that second network to generate new training data. Rinse, repeat until you no longer improve.
Fat Titz by Stockfish, the engine with the bodaciously big net. Remember: size matters. If you want to learn more about this engine just google for "Fat Titz".
-
MikeB
- Posts: 4889
- Joined: Thu Mar 09, 2006 6:34 am
- Location: Pen Argyl, Pennsylvania
Re: Surprise: Official release version of Stockfish 14.1
Thanks, the current SF net went a through a series of learnings, for the most part always usng the prior net as the base and the at the end, it went through a series SPSA tunings that look like this. https://github.com/official-stockfish/S ... a5a85a66b9dkappe wrote: ↑Fri Oct 29, 2021 4:22 amOof, that’s a broad topic. In this case we are using approaches similar to what AlphaZero did with TD-leaf. https://kstatic.googleusercontent.com/f ... 45072dc7bdMikeB wrote: ↑Fri Oct 29, 2021 3:58 amPlease define RL - I suppose it might mean training one net on top of another or is it something else?dkappe wrote: ↑Thu Oct 28, 2021 6:29 pmTwo abbreviations important to game neural network training: SL = supervised learning, RL = reinforcement learning. What Leela, AlphaZero and Dragon were/are doing is RL, so the training data from the improving nets is feeding back into the training. Stockfish is currently doing SL, so just a big batch of data from leela that they try to massage into a better net.
Note that SF started out with RL which produced their initial strong nets.
We are taking a network and using it to generate training data, then training or retraining a stronger network with that data and using that second network to generate new training data. Rinse, repeat until you no longer improve.
It eventually reaches a point where learnings or tunings do not add any additional Elo (enough to pass anyway).
One aspect, that is quite evident, when compared to pre-NNUE Stockfish , it is hundreds and hundreds of Elo stronger at low depth 1 to 7.
-
dkappe
- Posts: 1632
- Joined: Tue Aug 21, 2018 7:52 pm
- Full name: Dietrich Kappe
Re: Surprise: Official release version of Stockfish 14.1
The old, pre SF12 nets did go through this process, with dozens or even hundreds of rounds of data generation and training.MikeB wrote: ↑Fri Oct 29, 2021 4:32 am
Thanks, the current SF net went a through a series of learnings, for the most part always usng the prior net as the base and the at the end, it went through a series SPSA tunings that look like this. https://github.com/official-stockfish/S ... a5a85a66b9
It eventually reaches a point where learnings or tunings do not add any additional Elo (enough to pass anyway).
One aspect, that is quite evident, when compared to pre-NNUE Stockfish , it is hundreds and hundreds of Elo stronger at low depth 1 to 7.
The current ones do not (though there is a SPSA optimization of the final layer at the end). If you look at the current model, you’ll see there are some basic things missing which cause the model to run out of gas after 3-4 rounds of training. So, the current process uses a leela data trained net as it’s base (possibly starting from an earlier net), followed by a some Stockfish and corrective data. Doing a few hundred rounds of training and generation just doesn’t work with the current model.
Fat Titz by Stockfish, the engine with the bodaciously big net. Remember: size matters. If you want to learn more about this engine just google for "Fat Titz".
-
MikeB
- Posts: 4889
- Joined: Thu Mar 09, 2006 6:34 am
- Location: Pen Argyl, Pennsylvania
Re: Surprise: Official release version of Stockfish 14.1
Interesting, so do you think genererating new data, using the current net and the process you describe, should push it higher?dkappe wrote: ↑Fri Oct 29, 2021 4:45 amThe old, pre SF12 nets did go through this process, with dozens or even hundreds of rounds of data generation and training.MikeB wrote: ↑Fri Oct 29, 2021 4:32 am
Thanks, the current SF net went a through a series of learnings, for the most part always usng the prior net as the base and the at the end, it went through a series SPSA tunings that look like this. https://github.com/official-stockfish/S ... a5a85a66b9
It eventually reaches a point where learnings or tunings do not add any additional Elo (enough to pass anyway).
One aspect, that is quite evident, when compared to pre-NNUE Stockfish , it is hundreds and hundreds of Elo stronger at low depth 1 to 7.
The current ones do not (though there is a SPSA optimization of the final layer at the end). If you look at the current model, you’ll see there are some basic things missing which cause the model to run out of gas after 3-4 rounds of training. So, the current process uses a leela data trained net as it’s base (possibly starting from an earlier net), followed by a some Stockfish and corrective data. Doing a few hundred rounds of training and generation just doesn’t work with the current model.
-
dkappe
- Posts: 1632
- Joined: Tue Aug 21, 2018 7:52 pm
- Full name: Dietrich Kappe
Re: Surprise: Official release version of Stockfish 14.1
Nope. The current model just isn’t set up for RL. Compare with the leela models or some of the fully connected models from OpenAI Gym. You’ll note there are some things missing.
Fat Titz by Stockfish, the engine with the bodaciously big net. Remember: size matters. If you want to learn more about this engine just google for "Fat Titz".
