I agree with Don and I think that it is better to have weights that are too small and not weights that are too big.Richard Allbert wrote:Hi Don,
Can you give some pointers to this?
I've finally decided to start taking a disciplined approach to testing, have written a program to set up testing games using Cutechess.
The problem is knowing where to start.
For example, clean out the Eval function and tune just one Paramter
When the next parameter is tuned, it will be tuned vs the first value , and so on.
The values you end up with then depend on the order in which they are introduced? This doesn't seem a god way to do things.
If you get a result where VersionA has 2000elo +- 10 and VersionB 2020 +-10, then that is equal, correct? Unclear, as both fall within the error margin.
Can you give some tips on starting the testing, please?
I've tested a group of oppenents at 20s+0.2s for stabilty, all was ok, and I've used Bob Hyatt's openings.epd as starting positions.
Any help is appreciated. I don't have a huge cluster, unfortunately, rather 4 cores spare
Regards
Richard
I even think that deciding about weights that are too small when you know that the optimal weights are bigger may be productive for later improvement.
For example Initially when you do not have passed pawns evaluation
and tune your piece square table you may need a big bonus for pawns in the 6th or in the 7th .
Later with bonus for passed pawns(dependent on the rank of the pawn
and on the question if the pawn is blocked) you need smaller bonus for pushing passed pawns in the piece square table.
If you start with the "right" piece square table then you may have problems with improvement by adding other terms and you need to reduce your initial bonus so instead of increasing bonus only to reduce it later it seems to me faster to start with a bonus that is too small when hopefully later the bonus is not going to continue to be too small.
