We just have a very different understanding of 'competitive'. According to CCRL, only 2 engines below #15 use NNs and only one of them uses NNUE (to my knowledge). These would be Winter (using classical eval and CNN iirc) and Halogen. So, Seer is not really competitive with NNUE engines but still easily beats purely classical engines. That NNUE is superior to HCE shouldn't be a question. I was one of the first to show that. So yes, if we consider all engines, Seer is competitive.connor_mcmonigle wrote: ↑Mon Oct 25, 2021 6:12 pmI guess you didn't read my post very carefully.ChickenLogic wrote: ↑Mon Oct 25, 2021 5:58 pm ...
And again it shows that people here talking about NNUE know very little about how much it takes to do properly. Generating your own data that is just about half way decent and confirming it as such takes Stockfish well over a couple of weeks, and that is with multiple V100s and threadrippers. You also need multiple full training runs to confirm your new method isn't a fluke. If you don't happen to have multiple PCs in the range of multiple thousand of dollars there is no way you produce anything close to top engines without taking 'their' data. There are people in the Stockfish project that exclusively train neural nets. You sir just see that Ethereal and Halogen progress quickly but fail to see the insane amount of hardware they have in the background aside from open bench. OpenBench is also 'sponsored' by noob. You really think a single engine author can compete with a guy who has his own data center? And even if you join openbench, you still need insane amounts of hardware for training on your own.
I've generated about 2B d8 self play games which I use for training Seer on a R5 3600. I train solely using a GTX 950. Seer is a top 15 engine.
Koivisto is trained on 500M mixed depth self play games generated using comparatively modest hardware. Koivisto's training code is CPU based. Koivisto is a top 10 engine (at least).
BitGenie is a 3100 elo engine. It's selfplay data was generated on 4 thread laptop and it was also trained on said laptop.
So no, it doesn't take weeks of compute with "multiple V100s and threadrippers" to obtain competitive results. Maybe that's what it would take to beat Stockfish, but that's not what I claimed.
For me HCE is dead - I don't see them as part of the competition at all. Either way, I'm glad that there are other ways than the "SF way" to surpass HCE with 'moderate amounts of hardware juice'. I don't mean to discourage you or belittle your work given that you wrote your own trainer and also generate your own data. This is very commendable - especially given how many people don't do either. So keep on going
(I'm sorry I didn't specify what I meant by 'doing [NNUE training] properly'. I am aware that this could be seen as shifting goalposts. As silly as it sounds, sometimes I just forget that reading someone's mind is not a thing, especially over the internet)