bob wrote:You have been harping on the idea that there is correlation between the games. Karl explained where it most likely comes from.
bob wrote:Karl pointed this out quite clearly and described how it could easily explain the rather odd results that had come up over and over in the past.
Again, wrong, wrong, wrong! Why don't you read the following sentence a couple of hundred times, until it sinks in? (Well, perhaps make that 25,000 times.... Whatever it takes!

)
Fritzlein wrote:I have not explained why the original two 25600-game test runs failed to get the same wrong answer twice with high precision. I'll let computer scientists duke it, and to me it is frankly less interesting because the mathematical mystery is gone.
I have always been claiming that there is more correlation for games within the same run, than there is in games between the runs. For only that could drive up variability of the runs. Karl does
nothing explain that. As should be obvious for anyone that knows statistics, and as he very clearly writes (see highlited sentence in my quote above).
That it was perhaps not obvious to you, because I did not explicitly make the distinction witin runs / between runs, then I apologize. But addig that qualification has only become necessary since Karl altered the definition of correlation. At the time I was writing it, correlation implied 'correlation of the result of games from the same position within the runs', because if games would correlate equally between runs and within runs, in my definition there would be zero correlation between all games.
By looking at correlations in different sets of games (not relevant for explaining run variability, but relevant for explaining run error), one gets of course a different value for the correlations. No mystery there.
It appears that you are not only incapable of understanding what Karl wrote, (not even after I highlight it for you...), but also what I wrote (no surprise there, I guess...). You keep focussing on the point that in my ancient post where I explain what the
consequence of using results that systematically deviate from the truth is. If one uses an erroneous signal as feedback to 'improve' the engine, a feedback signal that contains large noise due to the small number of positions (frozen in, because you use that same small set consistently), the effect you get (namely that the engine gets in fact weaker, but does better in your flawed test) is known as training.
Hint: you should focus on the part that explains
why the measure you use for feedback (i.e. which changes to keep, and which to discard) contained an error. Perhaps you will understand it then. (Who says I am not an incurable optimist?

)