Using scale=1 seemed to give me close to real Elos and not some abstract Bayeselos, but that was the IPON database. Also, Adam plots for CCRL (?) databse showed the same, he matched the Bayeselo predictions to the Elo logistic, if I am not wrong, and got very good fit with scale set to 1. But now, that that problem with pure and complete lists appeared, I do not know where the problem is (or if there is a problem at all, maybe we should not compare different lists and that's it).Daniel Shawul wrote:I do not know of the IPON problem. Neither did I ventured to guess the what caused the difference b/n pure/complete rating lists of CCRL. But I know not using the scale is worse than using scale = 1 to make comparisons between different lists. If you take the above example I did, mm calculated the scale to be around 0.7 which is why the elostat and bayeselo ratings numbers are more or less equal. If I used scale=1, bayeselo output would be magnified by 1/0.7 = 1.43 so a 100 elo difference maybe magnified to 140 elo. This definately would make comparisons difficult.The IPON problem was using default Bayeselo. I think using scale=1 eliminates the problem to compare with performances. I do not know whith what scale the real Elos are shown.
Kai
If you look at using scale=1, there isn't really any advantage. Staying true to the model ? why anyway because one can assume the model multiplies by a factor. What advantage does scale=1 bring? I know for sure using scaled rating atleast makes comparisons somewhat more acceptable.
Kai