That is a start for sure -- proving a NN evaluation could be competive or even much better than a hand crafted evaluation function. The latency of evaluating the NN can be countered with a combination hardware (GPU/TPU) and software (async evaluations) which is what Google did for AlphaGo. Giraffe used only three layers of NN with chess specific inputs such as attack maps while AlphaZero used many more layers of CNN with just the rules of the game as input. Texel actually replaced its evaluation function with Giraffe's NN and showed that the eval is actually better but it would need a time odds to be competitive on the same hardware.
I expected that they would use alpha-beta but the fact that they did it with MCTS speeks volumes about the power of their NN (maybe i am wrong here and MCTS could be better than AB as they seem to claim). The fact that AZ started tuning itself from scratch doesn't surpize me one bit because all it does is slow down the convergence. I think they also claimed it could start from learning the rules of the game by itself and become a grandmaster as well -- but it would take even more training time. I am also able to train "scorpioZero" 's evaluation by starting from all parameters set to 0 -- not the same thing as AlphaZero's but still...
Once Remi started asking why is not anyone here trying an MCTS + CNN combo for chess, it raised my suspicion and for sure a week later AlphaZero dropped the bomb.
Daniel
Google's AlphaGo team has been working on chess
Moderators: hgm, Rebel, chrisw
-
- Posts: 4185
- Joined: Tue Mar 14, 2006 11:34 am
- Location: Ethiopia
-
- Posts: 273
- Joined: Wed Aug 24, 2016 9:49 pm
Re: Google's AlphaGo team has been working on chess
I think why they used mcts is pretty clear. They wanted a system that can play almost any (board) game by just knowing the rules. Mct's already has the property that you only need to konw the rules of the game. That's why I wonder whether alpha-beta-search+deepMind`s neural network might be better.Daniel Shawul wrote:That is a start for sure -- proving a NN evaluation could be competive or even much better than a hand crafted evaluation function. The latency of evaluating the NN can be countered with a combination hardware (GPU/TPU) and software (async evaluations) which is what Google did for AlphaGo. Giraffe used only three layers of NN with chess specific inputs such as attack maps while AlphaZero used many more layers of CNN with just the rules of the game as input. Texel actually replaced its evaluation function with Giraffe's NN and showed that the eval is actually better but it would need a time odds to be competitive on the same hardware.
I expected that they would use alpha-beta but the fact that they did it with MCTS speeks volumes about the power of their NN (maybe i am wrong here and MCTS could be better than AB as they seem to claim). The fact that AZ started tuning itself from scratch doesn't surpize me one bit because all it does is slow down the convergence. I think they also claimed it could start from learning the rules of the game by itself and become a grandmaster as well -- but it would take even more training time. I am also able to train "scorpioZero" 's evaluation by starting from all parameters set to 0 -- not the same thing as AlphaZero's but still...
Once Remi started asking why is not anyone here trying an MCTS + CNN combo for chess, it raised my suspicion and for sure a week later AlphaZero dropped the bomb.
Daniel
-
- Posts: 31
- Joined: Fri Nov 25, 2016 10:14 am
- Location: Singapore
Re: Google's AlphaGo team has been working on chess
In his NIPS2017 presentation, David Silver mentioned 3 academic
work the mosr recent is
Thinking Fast and Slow with Deep Learning and Tree Search
Thomas Anthony, Zheng Tian, David Barber
https://arxiv.org/abs/1705.08439
Worth a read to further understand AlphaZero
work the mosr recent is
Thinking Fast and Slow with Deep Learning and Tree Search
Thomas Anthony, Zheng Tian, David Barber
https://arxiv.org/abs/1705.08439
Worth a read to further understand AlphaZero
Brahim HAMADICHAREF
Singapore
Singapore
-
- Posts: 7000
- Joined: Thu Aug 18, 2011 12:04 pm
- Full name: Ed Schröder
Re: Google's AlphaGo team has been working on chess
Statements like these could make me a believer.Daniel Shawul wrote:That is a start for sure -- proving a NN evaluation could be competive or even much better than a hand crafted evaluation function. The latency of evaluating the NN can be countered with a combination hardware (GPU/TPU) and software (async evaluations) which is what Google did for AlphaGo. Giraffe used only three layers of NN with chess specific inputs such as attack maps while AlphaZero used many more layers of CNN with just the rules of the game as input. Texel actually replaced its evaluation function with Giraffe's NN and showed that the eval is actually better but it would need a time odds to be competitive on the same hardware.
-
- Posts: 13447
- Joined: Wed Mar 08, 2006 9:02 pm
- Location: Dallas, Texas
- Full name: Matthew Hull
Re: Google's AlphaGo team has been working on chess
As for mitigating latency there used to be significant optimizations available on MMX capable chips in the very old Brainmaker NN application (I still have several copies of it):Rebel wrote:Statements like these could make me a believer.Daniel Shawul wrote:That is a start for sure -- proving a NN evaluation could be competive or even much better than a hand crafted evaluation function. The latency of evaluating the NN can be countered with a combination hardware (GPU/TPU) and software (async evaluations) which is what Google did for AlphaGo. Giraffe used only three layers of NN with chess specific inputs such as attack maps while AlphaZero used many more layers of CNN with just the rules of the game as input. Texel actually replaced its evaluation function with Giraffe's NN and showed that the eval is actually better but it would need a time odds to be competitive on the same hardware.
https://calsci.com/MMX.html
It seems this software is still for sale. At the time (25 years ago) Brainmaker also sold an EISA card ($4,150) that could speed up training and processing by even more.
Matthew Hull
-
- Posts: 7000
- Joined: Thu Aug 18, 2011 12:04 pm
- Full name: Ed Schröder
Re: Google's AlphaGo team has been working on chess
It's still sold.
BrainMaker Professional for Windows $795
BrainMaker Professional for Windows $795
-
- Posts: 690
- Joined: Mon Apr 19, 2010 7:07 pm
- Location: Sweden
- Full name: Peter Osterlund
Re: Google's AlphaGo team has been working on chess
The post describing this test is here.Rebel wrote:Statements like these could make me a believer.Daniel Shawul wrote:That is a start for sure -- proving a NN evaluation could be competive or even much better than a hand crafted evaluation function. The latency of evaluating the NN can be countered with a combination hardware (GPU/TPU) and software (async evaluations) which is what Google did for AlphaGo. Giraffe used only three layers of NN with chess specific inputs such as attack maps while AlphaZero used many more layers of CNN with just the rules of the game as input. Texel actually replaced its evaluation function with Giraffe's NN and showed that the eval is actually better but it would need a time odds to be competitive on the same hardware.
-
- Posts: 2272
- Joined: Mon Sep 29, 2008 1:50 am
Re: Google's AlphaGo team has been working on chess
Just in case this isn't widely known, Marcel Van Kervinck did a similar experiment grafting SF's eval onto Crafty's search (by actually making Crafty call the SF binary).
http://rybkaforum.net/cgi-bin/rybkaforu ... ?tid=30107
In order to purely measure the quality of the eval Marcel used a clever trick to offset the difference in evaluation speed. He made both engines evaluate both evaluation functions and then pick their own. So for both engines evaluation took exactly the same time. This trick could also be used in the Texel-Giraffe experiment so that it is not necessary to translate an NPS difference into an elo differences (which always involves some guessing).
http://rybkaforum.net/cgi-bin/rybkaforu ... ?tid=30107
In order to purely measure the quality of the eval Marcel used a clever trick to offset the difference in evaluation speed. He made both engines evaluate both evaluation functions and then pick their own. So for both engines evaluation took exactly the same time. This trick could also be used in the Texel-Giraffe experiment so that it is not necessary to translate an NPS difference into an elo differences (which always involves some guessing).
Ideas=science. Simplification=engineering.
Without ideas there is nothing to simplify.
Without ideas there is nothing to simplify.
-
- Posts: 7000
- Joined: Thu Aug 18, 2011 12:04 pm
- Full name: Ed Schröder
Re: Google's AlphaGo team has been working on chess
I am try to train Giraffe exactly how it is described here.
But it crashes, see screen dump.
Anyone with experience? I am using Win7.
But it crashes, see screen dump.
Anyone with experience? I am using Win7.
-
- Posts: 2272
- Joined: Mon Sep 29, 2008 1:50 am
Re: Google's AlphaGo team has been working on chess
Other ways of offsetting the differences in evaluation speed are of course fixed depth or fixed nodes searches.Michel wrote:Just in case this isn't widely known, Marcel Van Kervinck did a similar experiment grafting SF's eval onto Crafty's search (by actually making Crafty call the SF binary).
http://rybkaforum.net/cgi-bin/rybkaforu ... ?tid=30107
In order to purely measure the quality of the eval Marcel used a clever trick to offset the difference in evaluation speed. He made both engines evaluate both evaluation functions and then pick their own. So for both engines evaluation took exactly the same time. This trick could also be used in the Texel-Giraffe experiment so that it is not necessary to translate an NPS difference into an elo differences (which always involves some guessing).
Ideas=science. Simplification=engineering.
Without ideas there is nothing to simplify.
Without ideas there is nothing to simplify.