Matthias Gemuh wrote:melajara wrote:Thanks, so far half of the test is completed and I confess I'm disappointed by the result, a meager 10 ELO increase over v2.3.1
I'm even hoping something is wrong in this testing

You are disappointed because of an ELO increase of a +3000 ELO engine ?
No, I know it's hard to improve at this level BUT
1, there are a lot of people involved in the improvement of Stockfish, some with interesting lateral thinking too.
2, there is now an excellent testing framework to separate the wheat from the chaff in all those "improvements".
3, There has been approximately 5 months to pile up improvements.
4, Houdini 3 raised the bar so much that it could expose previously hidden eval flaws in the top contenders tier.
So, all in all, I was expecting something like 30-40 ELO. Of course, there is the confidence margin and others factors like perhaps not easily quantifiable improvements in "style" but of course this is debatable.
For me it makes Houdini 3 even more impressive as seemingly NOBODY from the top contenders is able to catch up (if not, maybe, the elusive Komodo 6 MP)
