Alphazero news

Discussion of anything and everything relating to chess playing software and machines.

Moderators: hgm, Harvey Williamson, bob

Forum rules
This textbox is used to restore diagrams posted with the [d] tag before the upgrade.
User avatar
cdani
Posts: 2104
Joined: Sat Jan 18, 2014 9:24 am
Location: Andorra
Contact:

Re: Alphazero news

Post by cdani » Sat Dec 08, 2018 7:56 am

matthewlai wrote:
Sat Dec 08, 2018 3:17 am
glennsamuel32 wrote:
Sat Dec 08, 2018 2:21 am
Hello Matthew, nice to see you back after so long !!

Does this mean Giraffe will get some updates in the future ? :D
Thanks!

Afraid not! AlphaZero is taking up all my time these days, and it's a very exciting project with lots of uncharted territory ahead :). AlphaZero is basically what I ever wanted Giraffe to become... and then a lot more. I have never been this excited about computer chess my whole life.
Congratulations!!! I don't understand anything about A0, but I will try that Andscacs finds ways to play better against it :-)

Alexander Schmidt
Posts: 1086
Joined: Thu May 10, 2007 12:49 pm

Re: Alphazero news

Post by Alexander Schmidt » Sat Dec 08, 2018 11:29 am

Astatos wrote:
Fri Dec 07, 2018 12:01 pm
OK what we know :
1) Stockfish is the best engine in the world
2) LC0 guys did manage to reverse engineer A0 successfully
3) LC0 and A0 roughly at the same strength
4) NN are not less resource hungry than Alpha Beta
5) Scalability is about the same in both methods
6) Google has unacceptable behaviour, hiding data, obfuscating opponents and hyping results
I would rather say what we know is:
1) We have real new kind of chess engines
2) Computer chess enthusiasts should be happy and excited
3) We are at the very beginning of a new age of computer chess, nn will dominate computer chess in the future
4) Google shares the knowledge (what they don't have to do)
5) Some people make us a gift in reproducing the work (what they don't have to do)
6) Instead of saying "thank you" some fanboys of the common engines have an unacceptable behavior.

The only thing I really dislike is, that Google don't want to build AI to solve Chess, Go or whatever games. They want to learn how to use AI for other purposes and I think this will be dangerous for mankind. At the moment an AI decides which commercials we see, which news we read. That leads to radicalization and division of the society. One day an AI will decide if someone goes to jail or not. Maybe people will go to jail because an AI thinks he will someday do a crime. One day autonomous robots will decide which person to kill, on a battlefield or to prevent a possible crime. Maybe one day an AI will press the red button.

nabildanial
Posts: 104
Joined: Thu Jun 05, 2014 3:29 am
Location: Malaysia

Re: Alphazero news

Post by nabildanial » Sat Dec 08, 2018 11:39 am

Alexander Schmidt wrote:
Sat Dec 08, 2018 11:29 am
Maybe people will go to jail because an AI thinks he will someday do a crime. One day autonomous robots will decide which person to kill, on a battlefield or to prevent a possible crime. Maybe one day an AI will press the red button.
We have the so-called "Ethics of artificial intelligence" to prevent those things from happening.

matthewlai
Posts: 791
Joined: Sun Aug 03, 2014 2:48 am
Location: London, UK
Contact:

Re: Alphazero news

Post by matthewlai » Sat Dec 08, 2018 11:45 am

glennsamuel32 wrote:
Sat Dec 08, 2018 5:27 am
Matthew, could you divulge the size of the network file that A0 used ?
The details are in supplementary materials:
Architecture
Apart from the representation of positions and actions described above, AlphaZero uses the
same network architecture as AlphaGo Zero (9), briefly recapitulated here.
The neural network consists of a “body” followed by both policy and value “heads”. The
body consists of a rectified batch-normalized convolutional layer followed by 19 residual blocks (48).
Each such block consists of two rectified batch-normalized convolutional layers with a skip connection.
Each convolution applies 256 filters of kernel size 3 ⇥ 3 with stride 1. The policy head
applies an additional rectified, batch-normalized convolutional layer, followed by a final convolution
of 73 filters for chess or 139 filters for shogi, or a linear layer of size 362 for Go,
representing the logits of the respective policies described above. The value head applies an
additional rectified, batch-normalized convolution of 1 filter of kernel size 1 ⇥ 1 with stride 1,
followed by a rectified linear layer of size 256 and a tanh-linear layer of size 1.
Disclosure: I work for DeepMind on the AlphaZero project, but everything I say here is personal opinion and does not reflect the views of DeepMind / Alphabet.

Rein Halbersma
Posts: 685
Joined: Tue May 22, 2007 9:13 am

Re: Alphazero news

Post by Rein Halbersma » Sat Dec 08, 2018 12:16 pm

So why is it that A0's learning curve seems to flatten to almost no progress beyond it's current level? If e.g. the number of layers or channels is expanded, would you expect that a few hundred Elo more could be obtained? Or is A0 approaching perfection with its current network and is an absolute upperbound of Elo in sight?

nabildanial
Posts: 104
Joined: Thu Jun 05, 2014 3:29 am
Location: Malaysia

Re: Alphazero news

Post by nabildanial » Sat Dec 08, 2018 12:38 pm

matthewlai wrote:
Sat Dec 08, 2018 11:45 am
glennsamuel32 wrote:
Sat Dec 08, 2018 5:27 am
Matthew, could you divulge the size of the network file that A0 used ?
The details are in supplementary materials:
Architecture
Apart from the representation of positions and actions described above, AlphaZero uses the
same network architecture as AlphaGo Zero (9), briefly recapitulated here.
The neural network consists of a “body” followed by both policy and value “heads”. The
body consists of a rectified batch-normalized convolutional layer followed by 19 residual blocks (48).
Each such block consists of two rectified batch-normalized convolutional layers with a skip connection.
Each convolution applies 256 filters of kernel size 3 ⇥ 3 with stride 1. The policy head
applies an additional rectified, batch-normalized convolutional layer, followed by a final convolution
of 73 filters for chess or 139 filters for shogi, or a linear layer of size 362 for Go,
representing the logits of the respective policies described above. The value head applies an
additional rectified, batch-normalized convolution of 1 filter of kernel size 1 ⇥ 1 with stride 1,
followed by a rectified linear layer of size 256 and a tanh-linear layer of size 1.
I think what glenn meant by the question is how big the filesize is as in MB.

Alexander Schmidt
Posts: 1086
Joined: Thu May 10, 2007 12:49 pm

Re: Alphazero news

Post by Alexander Schmidt » Sat Dec 08, 2018 12:48 pm

nabildanial wrote:
Sat Dec 08, 2018 11:39 am
Alexander Schmidt wrote:
Sat Dec 08, 2018 11:29 am
Maybe people will go to jail because an AI thinks he will someday do a crime. One day autonomous robots will decide which person to kill, on a battlefield or to prevent a possible crime. Maybe one day an AI will press the red button.
We have the so-called "Ethics of artificial intelligence" to prevent those things from happening.
These things already happen:
http://www.israeltoday.co.il/NewsItem/t ... fault.aspx
https://www.datanami.com/2017/07/17/neu ... ion-banks/

jp
Posts: 748
Joined: Mon Apr 23, 2018 5:54 am

Re: Alphazero news

Post by jp » Sat Dec 08, 2018 12:54 pm

nabildanial wrote:
Sat Dec 08, 2018 12:38 pm
matthewlai wrote:
Sat Dec 08, 2018 11:45 am
glennsamuel32 wrote:
Sat Dec 08, 2018 5:27 am
Matthew, could you divulge the size of the network file that A0 used ?
The details are in supplementary materials
I think what glenn meant by the question is how big the filesize is as in MB.
Yep. How big is the filesize in MB?

shrapnel
Posts: 1153
Joined: Fri Nov 02, 2012 8:43 am
Location: New Delhi, India

Re: Alphazero news

Post by shrapnel » Sat Dec 08, 2018 12:57 pm

Any Plans to Commercialize AlphaZero ?
Can't wait to get my hands on AlphaZero engine no matter what the Price.
i7 5960X @ 4.1 Ghz, 64 GB G.Skill RipJaws RAM, Asus ROG Strix 11 GB Geforce 1080 Ti and AMD Ryzen 7 1800X @4.0 GHz, 32 GB DDR4-2400 G.Skill RAM, ASUS Prime x370-PRO, Noctua NH-D15 SE-AM4 Cooler.

matthewlai
Posts: 791
Joined: Sun Aug 03, 2014 2:48 am
Location: London, UK
Contact:

Re: Alphazero news

Post by matthewlai » Sat Dec 08, 2018 1:55 pm

jp wrote:
Sat Dec 08, 2018 12:54 pm
nabildanial wrote:
Sat Dec 08, 2018 12:38 pm
matthewlai wrote:
Sat Dec 08, 2018 11:45 am
glennsamuel32 wrote:
Sat Dec 08, 2018 5:27 am
Matthew, could you divulge the size of the network file that A0 used ?
The details are in supplementary materials
I think what glenn meant by the question is how big the filesize is as in MB.
Yep. How big is the filesize in MB?
Not sure. We don't store the networks locally. They are just TensorFlow SavedModels. If you construct the same network and save it you'll get the same size. https://www.tensorflow.org/guide/saved_model
Disclosure: I work for DeepMind on the AlphaZero project, but everything I say here is personal opinion and does not reflect the views of DeepMind / Alphabet.

Post Reply