Lee Sedol vs. AlphaGo [link to live feed]

Discussion of anything and everything relating to chess playing software and machines.

Moderators: hgm, Rebel, chrisw

Uri Blass
Posts: 10282
Joined: Thu Mar 09, 2006 12:37 am
Location: Tel-Aviv Israel

Re: Lee Sedol vs. AlphaGo [link to live feed]

Post by Uri Blass »

Milos wrote:
Uri Blass wrote:Define reasonable evaluation function.
I do not consider piece square table+material as a reasonable evaluation function in chess and I believe that a good evaluation function can beat piece square table+material even with 1,000,000 nodes per move against 100,000,000 nodes per move assuming both programs use the same search rules.
That's because you are thinking as a human chess player and not as a computer chess programmer. You are imagining some "godly" (EGTB like) evaluation function that is actually including search in itself, because as a human you cannot so easily decompose evaluation from search that does actually happen in your brain when you "evaluate" the position. Anything that you as a human "see" on the board and that you cannot clearly formulate as a rule is actually including a search that happens on subconscious level and that ppl like to call ... intuition ;).
No

I think about important factors like mobility and king safety and pawn structure that are not in piece square table evaluation.
peter
Posts: 3186
Joined: Sat Feb 16, 2008 7:38 am
Full name: Peter Martan

Re: Lee Sedol vs. AlphaGo [link to live feed]

Post by peter »

Hi!
Lee Sedol has won his first game in the match.
3-1
Peter.
whereagles
Posts: 565
Joined: Thu Nov 13, 2014 12:03 pm

Re: Lee Sedol vs. AlphaGo [link to live feed]

Post by whereagles »

Milos wrote:ML based evaluation chess will probably never beat classical approach.
I wouldn't be so quick to jump into that conclusion, but yeah you might be right.
whereagles
Posts: 565
Joined: Thu Nov 13, 2014 12:03 pm

Re: Lee Sedol vs. AlphaGo [link to live feed]

Post by whereagles »

peter wrote:Hi!
Lee Sedol has won his first game in the match.
3-1
there is still hope for mankind :D
Dirt
Posts: 2851
Joined: Wed Mar 08, 2006 10:01 pm
Location: Irvine, CA, USA

Re: Lee Sedol vs. AlphaGo [link to live feed]

Post by Dirt »

peter wrote:Hi!
Lee Sedol has won his first game in the match.
3-1
With the match victory in hand AlphaGo probably got lazy. :-)
Deasil is the right way to go.
matthewlai
Posts: 793
Joined: Sun Aug 03, 2014 4:48 am
Location: London, UK

Re: Lee Sedol vs. AlphaGo [link to live feed]

Post by matthewlai »

Milos wrote: Chess patterns are much simpler but chess is also much simpler game and therefore quality of chess play is for many orders of magnitude higher than quality of go play, or if you like much, much closer to perfection that level of go play.
Yes, you programmed a nice program using pattern recognition in evaluation, but take SF, strip it completely of its evaluation and replace it always with only lazy eval (material, pst and mobility) and it would still destroy your Giraffe. So the question is what is the point of some "advanced" evaluation when it is so inaccurate (and by inaccurate I mean it cannot discern reliably between couple of cp score difference that might make huge difference in selection a winning or losing move) ? How much is a slow down of your program due to calculation of the evaluation function.
I am guessing that if you took Giraffe, strip it completely of its "smart eval", use only material, pst and mobility and some endgame knowledge and pawn structure knowledge (so effectively extremely light evaluation) that it would beat current version of Giraffe comfortably.
So again, what is the purpose of that type of ML evaluation in chess, except as a student exercise? Deepmind is not working on chess not because it is easy, but because they know it's a lost cause. ML based evaluation chess will probably never beat classical approach.
Yes, that's exactly what I meant. Handwritten evaluation functions in chess are good enough that switching to a neural network (slightly better outputs but much slower) is a net negative.

I am not so sure about Giraffe with a simpler eval beating Giraffe. On CCRL Giraffe already beats many many engines that search 10-100x faster than it does (at around 2400 level). So there is certainly a lot of value in having a very good eval at that level, and it can certainly compensate for the speed loss to some extent, just not enough to get it to the top level.

Giraffe searches at about 140knps on my machine, compared to 1Mnps for Stockfish, and close to 3Mnps for Crafty (all single CPU).

I am still not entirely convinced that the AlphaGo/Giraffe approach (value and policy networks) cannot get to state of the art in chess, but I unfortunately don't have time to explore that anymore. Giraffe was just a crude first attempt. I really hope someone skilled in ML steps in and take it as far as it can go and see what happens.

Even if the AlphaGo/Giraffe approach doesn't work, there are still many possible machine learning approaches that can be explored.

Quoting myself from another thread -
Today's chess engines are terrible at generalizing and reusing results.

A human can see that a chess position is comprised of different patterns, many of which are independent, and results can be reused across different branches to massively cut down on branching factor.

Transposition tables are a very crude attempt to exploit this, but they require exact matches, which are very rare.

IMHO the next big breakthrough in computer chess will come when we have transposition tables (or something like transposition tables) that can do "fuzzy matches" to reuse results from different branches that don't have exactly matching positions.

This is obviously very difficult because of all the exceptions and special cases, and I don't really see any way of solving this without machine learning. We need a system that can extract the "essence" of positions, and use that to do matching instead.

I have some pretty concrete ideas on how to do this based on non-public information. A lot of that will become public at ICML in a few months.

Here is one possible starting point, if anyone wants to take a crack at it: http://arxiv.org/abs/1410.5401
I think it's dangerous to say that chess nowadays is close to perfect. It's an easy illusion to have when we don't know any better. People thought the same with Rybka many years ago, and nowadays top engines are a few hundred ELOs stronger than early Rybka.
Disclosure: I work for DeepMind on the AlphaZero project, but everything I say here is personal opinion and does not reflect the views of DeepMind / Alphabet.
User avatar
towforce
Posts: 11572
Joined: Thu Mar 09, 2006 12:57 am
Location: Birmingham UK

Re: Lee Sedol vs. AlphaGo [link to live feed]

Post by towforce »

Dirt wrote:
peter wrote:Hi! Lee Sedol has won his first game in the match. 3-1
With the match victory in hand AlphaGo probably got lazy. :-)
From the commentary, it looks as though AlphaGo played some inexplicably bad moves. I wonder if Google are trying to trap Ke Jie, the world's top-rated player, into accepting a challenge after he rudely said he thought he was better than AlphaGo?
Writing is the antidote to confusion.
It's not "how smart you are", it's "how are you smart".
Your brain doesn't work the way you want, so train it!
mvk
Posts: 589
Joined: Tue Jun 04, 2013 10:15 pm

Re: Lee Sedol vs. AlphaGo [link to live feed]

Post by mvk »

Dirt wrote:
peter wrote:Hi!
Lee Sedol has won his first game in the match.
3-1
With the match victory in hand AlphaGo probably got lazy. :-)
It calculated the probability of winning the match by losing this game as 100% and then lost the game.
[Account deleted]
peter
Posts: 3186
Joined: Sat Feb 16, 2008 7:38 am
Full name: Peter Martan

Re: Lee Sedol vs. AlphaGo [link to live feed]

Post by peter »

mvk wrote: It calculated the probability of winning the match by losing this game as 100% and then lost the game.
:D
Peter.
User avatar
Laskos
Posts: 10948
Joined: Wed Jul 26, 2006 10:21 pm
Full name: Kai Laskos

Re: Lee Sedol vs. AlphaGo [link to live feed]

Post by Laskos »

Crazy Stone evaluation (percentages to win for Black) with both White and Black:

Game 4:
Image

Early on in the game, AlphaGo was doing great, like in the previous games. We know that the decisive blunder of AlphaGo occurred at move 79, Crazy Stone sees something at move 81, but still thinks AlphaGo is winning up to move 100 or so. From DeepMind tweets it seems AlphaGo was still thinking it wins up to move 87.