Later.


A summary of results so far. Glenn-david and Rao-Kupper seem to overlap a lot


Moderator: Ras




The current system might work pretty well for human players; humans tend to also learn of course, yet they do not play long matches against another player usually.Rémi Coulom wrote:Hi Vincent,diep wrote:What seems very popular nowadays is that all sorts of engines do learning in a rather hard manner. Hard i mean: difficult to turn off.
Bayeselo assumes players of constant strength. Measuring changes in strength caused by learning is much more difficult. It may be possible to adapt WHR to do it:
http://remi.coulom.free.fr/WHR/
But if you want to measure accurately the change in strength caused by a change in your algorithm, it is considerably more efficient to use bayeselo with opponents that don't learn.
Rémi
Nonsensediep wrote:
In most engines you simply cannot turn off learning. They have learned simply they get higher at elolist if you can't turn it off.
Deleting of learnfiles helps indeed. No one is deleting them however.Uri Blass wrote:Nonsensediep wrote:
In most engines you simply cannot turn off learning. They have learned simply they get higher at elolist if you can't turn it off.
1)I believe that most engines do not have learning
2)It is easy to turn off learning for engines that learn.
Simply delete the engine and the files that it generated from your computer and download it again.
I don't expect this could improve performance much.Daniel Shawul wrote:Even though it is fast now, I would like to speed it up more by using a radial basis function to fit the likelihood function. I did some experiments in the past using a multi-quadric rbf function for a black box optimization. The results were good and it significantly sped up my objective function for the project I worked on. Anyway I don't know if a different rbf such as gaussian would be more appropriate here, but I will try it sometime in the future.
Ok after some break, I continued working on it. I did a goodness of fit test using pearson's chi squared test. The result shows Davidson is the best fit followed by Glenn David. Both seem to be significantly better than the default bayeselo model i.e Rao Kupper. I did the computation on two pgns , now including cegt blitz with about 1 million games.Rémi Coulom wrote:I don't expect this could improve performance much.Daniel Shawul wrote:Even though it is fast now, I would like to speed it up more by using a radial basis function to fit the likelihood function. I did some experiments in the past using a multi-quadric rbf function for a black box optimization. The results were good and it significantly sped up my objective function for the project I worked on. Anyway I don't know if a different rbf such as gaussian would be more appropriate here, but I will try it sometime in the future.
What may be important, if you are not alreadying doing it, is to perform CG in the space of ratings (log(gamma)) instead of the space of gamma. The posterior distribution should have an approximately Gaussian shape in the space of ratings, so its logarithm is approximately quadratic. Maybe you can you the Hessian computed for the LOSS to do the optimization. But the Hessian does not scale with a huge number of players, so CG might be better.
I hope you can show that Davidson fits the data better than a Gaussian
Rémi
Code: Select all
CCRL CEGT
Davidson 692 628
GlennDavid 748 777
RaoKupper 1079 1432
Great !Daniel Shawul wrote:Ok after some break, I continued working on it. I did a goodness of fit test using pearson't chi square. The result shows Davidson is the best fit followed by Glenn David. Both seem to be significantly better than the default bayeselo model i.e Rao Kupper. I did the computation on two pgns , now including cegt blitz with about 1 million games.Code: Select all
CCRL CEGT Davidson 692 628 GlennDavid 748 777 RaoKupper 1079 1432
I recommended half of the games, but a more usual statistical procedure is 10-fold cross-validation.Next I will randomly select half the games for each player to construct a model and then test for goodness of fit again on the rest of the data.
Sure that is what I had in mind. But you should know I am not so much knowledgeable about it so you have to tell me to do this and thatI am very glad you did all that work. I had been waiting to find someone to do it for a long time. Please tell me if you are interested in finishing the paper I started.
I had not noticed that question. Because the ratings can be offset by a constant without changing the log-likelihood, the Hessian is singular. Its eigenvalue for direction (1, 1, 1, ..., 1) is zero, which would mean an infinite variance. So, what I do is fix the rating of one player, and get the covariance of all the others.Daniel Shawul wrote:Question: I don't quite understand how the A matrix matrix was derived. Can you explain please? I thought the covariance matrix was just inverse of the negative hessian C = -H^-1 but doing that clearly fails most of the time. Later on at the end of the page you also tried to use cholesky but it didn't work for me probably because the matrix is not positive definite. What are the properties of the covariance matrix ?