Discussion about this post

User's avatar
Ben Johnson's avatar

Fascinating, as always Julian! Keep up the good work.

Expand full comment
dboing dboing's avatar

you:

The first striking thing for me is that the graphs for Leela and Maia mostly agree, which is kind of amazing since they work quite differently

me: well. I wonder if there is not some high level information re-circulation here. It might depends on how maia was trained. The principle of it is a model with a perfect play benchmark and then a binned category variable dependent error model with few non-board hyperparameters (if i got it), and maybe some position information dependency (???). It might be that in the conversion calibration from SF and using lichess data, their odds function definition might be so that your observation would follow.

As while the chess play itself or the NN training using the same LC0 input encoding, but has different purpose, (the error from SF I presume, or other perfect play benchmark definition), I smell somewhere in there, that we should check the flow of information. In the high level statistics about odds. I am not sure that those are independent. I don't even understand what it being trained. Fitting the licehss outcome and the error model at each position, giving the error model fit and its parameter estimation.

If I am readable, do you have yourself an understadning of what was the machine learning setup. The degsing of the training data matrix. We have lichess games. all their positions, the players ratings, and the game outcomes. We also have for each positions within that set of games (EPDs being same position or not, see my other comment on definition of position difficulty), the SF benchamark modulo or added error part (whatever the function of error and benchamark dictated by the model assumption) that would fit the outcome. IS that right? don'T have to answer if not making sense. don'T sweat. but if you share my questions... please help.

Expand full comment
7 more comments...

No posts