I think you can link to the info here:
http://abrok.eu/stockfish/?page=6
The actual text follow:
As I was in the middle of some code rewrite, I decided to try this idea. So far, it is working fine. My implementation is as follows:Author: Joost VandeVondele
Date: Mon Dec 18 16:32:21 2017 +0100
Timestamp: 1513611141
Allow for general transposition table sizes. (#1341)
For efficiency reasons current master only allows for transposition table sizes that are N = 2^k in size, the index computation can be done efficiently as (hash % N) can be written instead as (hash & 2^k - 1). On a typical computer (with 4, 8... etc Gb of RAM), this implies roughly half the RAM is left unused in analysis.
This issue was mentioned on fishcooking by Mindbreaker:
http://tests.stockfishchess.org/tests/v ... ccbb8be04
Recently a neat trick was proposed to map a hash into the range [0,N[ more efficiently than (hash % N) for general N, nearly as efficiently as (hash % 2^k):
https://lemire.me/blog/2016/06/27/a-fas ... eduction/
namely computing (hash * N / 2^32) for 32 bit hashes. This patch implements this trick and now allows for general hash sizes. Note that for N = 2^k this just amounts to using a different subset of bits from the hash. Master will use the lower k bits, this trick will use the upper k bits (of the 32 bit hash).
There is no slowdown as measured with [-3, 1] test:
http://tests.stockfishchess.org/tests/v ... 0ccbb8be04
LLR: 2.96 (-2.94,2.94) [-3.00,1.00]
Total: 128498 W: 23332 L: 23395 D: 81771
There are two (smaller) caveats:
1) the patch is implemented for a 32 bit hash (so that a 64 bit multiply can be used), this effectively limits the number of clusters that can be used to 2^32 or to 128Gb of transpostion table. That's a change in the maximum allowed TT size, which could bother those using 256Gb or more regularly.
2) Already in master, an excluded move is hashed into the position key in rather simple way, essentially only affecting the lower 16 bits of the key. This is OK in master, since bits 0-15 end up in the index, but not in the new scheme, which picks the higher bits. This is 'fixed' by shifting the excluded move a few bits up. Eventually a better hashing scheme seems wise.
Despite these two caveats, I think this is a nice improvement in usability.
Bench: 5346341
Initialize the hash with:
Code: Select all
U32 m_NumHashEntries = (HashSize * MEGABYTE) / sizeof(TTMAIN);
m_Hash = (TTMAIN *)calloc(m_NumHashEntries, sizeof(TTMAIN));
TTMAIN is the hash entry structure
MEGABTYE = 1024 * 1024
To get the correct index from the key...
Code: Select all
U32 GetHashIndex(const U64 key){
return ((key & BITMASK32) * m_NumHashEntries) >> 32;
// from stockfish
}U64 BITMASK32 = 0xFFFFFFFF;
the 32 bit right shift is a very fast divide so the extra overhead of this method is not very much. The first mask gives use the lower 32bits of the 64bit key, and the multiply will not overflow a 64bit number. Then we shift 32 to the right to get the upper portion as our index.