It’s just an experimental chess engine, and will never release to the public.
A lot of words. Where can I download and test your engine vs Stockfish 18?
The proposal sounds ambitious, but in its current form it reads more like
promotional AI language than a serious technical argument.
The central idea is easy to understand:
instead of one NNUE evaluation network, use several specialized NNUE style evaluators and let a gating mechanism decide which ones should contribute in a given position. That part is at least conceptually reasonable. However, the text repeatedly jumps from “interesting possibility” to “major breakthrough” without providing the evidence needed to justify such claims.
The first problem is that it assumes MoE can be transferred into a chess engine almost for free. That is not obvious at all. In a classical alpha beta engine,
evaluation must be extremely fast, cache friendly, and predictable. A routing network, multiple experts, additional memory traffic, and expert selection overhead may easily cancel out any theoretical gain in model capacity.
In chess engines, elegant ideas often fail because the engineering cost per node is too high.
The second problem is that the text treats specialization as automatically beneficial. It is true that opening play, tactical middlegames, positional play, and endgames differ greatly. But it does not follow that several experts will outperform one well trained network under real search conditions. The burden of proof is not on the skeptic. It is on the proposer. One would need strong experimental evidence showing better Elo at fixed time controls, not just a plausible story.
Third, the claim that this approach could scale to massive parameter counts while still keeping very high NPS is speculative. In chess, raw arithmetic is only part of the story. Memory locality, cache reuse, branching behavior, and incremental update structure matter enormously. A larger sparse model may look efficient on paper while being slower in practice on CPU hardware, which is still the main battlefield for top alpha beta engines.
Fourth, the training discussion is vague. Phrases such as “train the gate network with reinforcement learning” and “end to end fine tuning” sound impressive, but they avoid the hardest questions. How is expert collapse prevented? How is balanced usage maintained? How is consistency between experts ensured so that search does not become noisy or unstable? How is the incremental update mechanism preserved across multiple experts without creating unacceptable implementation complexity?
Fifth, the text overstates novelty. Specialized evaluation is not a magical new concept. Chess engines have long separated concerns implicitly or explicitly through handcrafted terms, phase dependent interpolation, endgame logic, and different search heuristics for different position types. Even within neural evaluation, the mere addition of routing and multiple subnetworks does not automatically create a new generation of engine design.
So the fair conclusion is this: the idea may be worth exploring as a research direction, but the text does not demonstrate a breakthrough. It offers a speculative architecture, wraps it in grand language, and skips the hard empirical questions. Until there are rigorous benchmarks showing clear Elo gains at practical time controls and acceptable speed on real hardware, this should be treated as an interesting hypothesis, not as proof of a new era in chess engine design.