lithander wrote: ↑Fri Jul 08, 2022 12:27 pm
Maybe as a intermediate step try to tune the HCE of Blunder with a training set exclusively derived from Blunder's own games. So forget all you learned from Zurichess' games and try to learn it again from selfplay games - starting with material only evaluation. At least that's what I'm going to try with Leorik before I transition to NNUEs...
Good idea! I like that a lot, actually.
Because if I were to start training a neural network, I suppose its data wouldn't really be original, due to using the Zurichess dataset for so much tuning. That definitely sounds like a fun challenge to try to reach the strength of the latest Blunder dev version using only purely self-trained data.
To do this, I would need to go all the way to Blunder 1.0.0, since for that version, all of the evaluation data was 100% original. All of it is likely very bad, but still original

I selected basic material values and wrote some very simple PSQT myself, so the engine plays a lot like me. That would be where I would start, downloading it and playing many self-play games.
I'm actually pretty curious if I can do the sort of cyclic tuning you describe, where each tuning session I can gain strength just from the new self-play games.
lithander wrote: ↑Fri Jul 08, 2022 12:27 pm
And I will also delay looking into NNUEs until at least version 3.0 (with a lot of minor versions in between) because once you switch I fear there's no coming back to tinker with the HCE. You should really feel like you're "done" with the HCE e.g. that all your ideas that you are curious about have been tried and tested.
True. I don't think many people would be interested in a switch back to HCE in a new version after you've already given a taste of the power of NNUE.
I'll probably start preliminary experiments at least with NNs right now, just because I find them fascinating in their own right. Might not even be something chess-related, but still something made from scratch using a custom train, not Tensorflow or something.
But don't worry, I still have some HCE ideas on the backlog

especially now that I fixed that bug. I actually have several test I need to go back and run since I'm not sure how that bug might have warped their improvement.
lithander wrote: ↑Fri Jul 08, 2022 12:27 pm
Leorik is scared!

How much Elo was it? And are you going to release a fixed version?
Well tell Leorik not to worry too much right now

after 2K games at 10+0.1s:
Code: Select all
Score of Blunder 8.2.1 vs Blunder 8.2.0: 569 - 550 - 881 [0.505] 2000
... Blunder 8.2.1 playing White: 306 - 255 - 440 [0.525] 1001
... Blunder 8.2.1 playing Black: 263 - 295 - 441 [0.484] 999
... White vs Black: 601 - 518 - 881 [0.521] 2000
Elo difference: 3.3 +/- 11.4, LOS: 71.5 %, DrawRatio: 44.0 %
SPRT: llr 0.0356 (1.2%), lbound -2.94, ubound 2.94
I'm going to try testing at longer TCs as well, something like 15+0.2s or 5+0.5s, but for now the gain isn't huge. Regardless, I'm glad Blunder is no longer making such bad...er...blunders.
Anyhow, I'll definitely be releasing a bug-fixed version soon. Since the current tournament Blunder 8.0.0 is so far along, I'll probably let it finish first, since I don't want Graham to feel like he has to re-start Blunder's games just because of me

but I'll release Blunder 8.2.0 before the next division tourney, which hopefully has around +50 Elo now on 8.0.0.