Well, claims without justification can be rejected without justification.gladius wrote: Agreed, it's certainly worth investigating other approaches, and the 700 ELO is no joke . This quote from the paper is interesting though (even if it doesn't really have any justification):
AlphaZero evaluates positions using non-linear function approximation based on a deep neural network, rather than the linear function approximation used in typical chess programs. This provides a much more powerful representation, but may also introduce spurious approximation errors. MCTS averages over these approximation errors, which therefore tend to cancel out when evaluating a large subtree. In contrast, alpha-beta search computes an explicit minimax, which propagates the biggest approximation errors to the root of the subtree. Using MCTS may allow AlphaZero to effectively combine its neural network representations with a powerful, domain-independent search.
To be honest, that paragraph to me reads as "hey, you know, we're different. Different isn't necessarily better or worse, but we should put some positive spin on it, so... ok, how's this: we're averaging over errors rather than propagating them, so let's say that's better and hope we can get it past the referee without doing more work."
Clearly, you can do more with a non-linear function than with a linear one (and obviously there's no reason "typical chess programs" can't use non-linear evaluation terms; some do) and clearly there are more pitfalls with non-linear functions as well. Calling these "approximation errors" is a simplification and claiming they will average out is not justified without proof (it's intuitively likely, but it depends on the nature of the deviation and a word or two to reassure the reader that yes, in this case they do average out, would not be out of place). You can't claim that backing up the largest score implies backing up the largest error without proof either, for that matter.