I've been working on tactical testsuites for well over 10 years and lately it has become harder and harder to find challenging positions. But after much work over the last year, and with help from other testsuites
It's currently private, but you'll have seen 70% of the positions before.
Then, I got hold of 3 powerful machines:
1) Threadripper 3990X 64 cores / 128 threads (SF15 running on 60 threads, 70-90 MN/s averaged)
2) Cloud rental SF15 Cluster 512 cores (the company claims 1 BN/s, in reality 800 MN/s is closer)
3) Cloud rental SF15 Cluster 1024 cores (the company claims 2 BN/s, in reality 1.6 BN/s is closer)
Each contender got 2 mins/position on the Testsuite with the following results:
1. SF Cluster "2 BN/s" 96/100 average solve time 3.10 seconds
2. SF Cluster "1 BN/s" 92/100 average solve time 4.49 seconds
3. Threadripper 3990X 90/100 average solve time 5.86 seconds
The difficulty of a few positions disguises how far apart these scores are.
So far, nothing too surprising.
Then I decided to give the Threadripper 16x as much time (32 mins per position) and it scored
94/100 with an average solve time of 59 seconds.
The thing I'm surprised by is that despite the vast difference in threads (and with that associated search efficiency loss, even with Lazy SMP?) the results are surprisingly dependent on the total number of nodes searched! Threadripper on 16x more time searched a total number of nodes between the two cluster versions at 2 mins/position and its testsuite score was in line with that. I am quite surprised by this.
Also time to depth was pretty meaningless: the cluster versions were only a few plies higher and d=35 on a cluster means something much more than d=35 on a PC I think.