CEGT - rating lists March 16th 2008

Discussion of computer chess matches and engine tournaments.

Moderator: Ras

Michael Sherwin
Posts: 3196
Joined: Fri May 26, 2006 3:00 am
Location: WY, USA
Full name: Michael Sherwin

Re: New Homer release

Post by Michael Sherwin »

Tony Thomas wrote:
Michael Sherwin wrote:If the ring that that creature coveted is the equivilent of ELO for the chess programmer then I am not that creature. Yet! :roll: But, to get ahead of and stay ahead of 'you know who' ... Lets just say that I sense my inner demon starting to awake! :twisted: :lol:
Its not the relative ELO, its the relative IQ.. :wink:
My IQ is the same as yours Tony :wink: , 140.

However, mine was measured by a three part two hour mensa test. But, the twenty minute iternet test that I took a couple of years ago also gave a 140 result.

All that means is that the both of us are just smart enough to also have a very high IQ of a different type. A super high idiot quotient! :twisted:

Just some humble pie King Leonidas! :P

It's 1AM, do you know where your goat is! :lol:
If you are on a sidewalk and the covid goes beep beep
Just step aside or you might have a bit of heat
Covid covid runs through the town all day
Can the people ever change their ways
Sherwin the covid's after you
Sherwin if it catches you you're through
Tony Thomas

Re: New Homer release

Post by Tony Thomas »

Michael Sherwin wrote:
Tony Thomas wrote:
Michael Sherwin wrote:If the ring that that creature coveted is the equivilent of ELO for the chess programmer then I am not that creature. Yet! :roll: But, to get ahead of and stay ahead of 'you know who' ... Lets just say that I sense my inner demon starting to awake! :twisted: :lol:
Its not the relative ELO, its the relative IQ.. :wink:
My IQ is the same as yours Tony :wink: , 140.

However, mine was measured by a three part two hour mensa test. But, the twenty minute iternet test that I took a couple of years ago also gave a 140 result.

All that means is that the both of us are just smart enough to also have a very high IQ of a different type. A super high idiot quotient! :twisted:

Just some humble pie King Leonidas! :P

It's 1AM, do you know where your goat is! :lol:
Haha, I have never taken the Mensa. However I did take a pattern recognition test with 30 questions without any time limits when I tried out for the police academy. I think I got 28/30, and I have no idea what the score was...On an internet test that I took I only scored 136 so I am not a genious like you are.
Michael Sherwin
Posts: 3196
Joined: Fri May 26, 2006 3:00 am
Location: WY, USA
Full name: Michael Sherwin

Re: New Homer release

Post by Michael Sherwin »

Daniel Mehrmann wrote:
Werner wrote:
Graham Banks wrote: I'm pretty sure that CEGT also test with learning off.
Regards, Graham.
Bingo!
btw: Can somebody tell me why the learning, specially here Romichess, should be so importent ? I think without learning is it just fair, like gerneral books.

But i can tell you that a guy, if i remember correctly his name was Rodolfo, did some tests with engine learning. He compared Romichess, TheBaron (private version), Homer and Spike with static opening positions.

I was very very suprised as he told me that Homer won this competition. Homer (version 2.0) beated Romichess without any problems as well.
Interesting was that at the beginning Romi was better with his learning, but as longer the match run Homer overrun Romi.

So, like "Vas" style, i'm claiming that Homer has the world strongest (best) learning feature and he just only needs 256kb filesize. :twisted:

Best,
Daniel
Rodolfo never said anything like that to me. Nor AFAIK did he publish those test. If someone wants to test your assertion and make a report then that is fine. My personal test done a long time ago gave a different result--Romi's learning was better!

256kb is enough for testing one static position, maybe. It would not be enough to cover an entire range of openings. I suppose that you could have many such files though if you wanted.

Anyway I was just showing an interesting way of doing learning and never made a claim that it was best. I posted results that showed that Spikes might be better!

If what you do is better, then I am happy for you! :D
If you are on a sidewalk and the covid goes beep beep
Just step aside or you might have a bit of heat
Covid covid runs through the town all day
Can the people ever change their ways
Sherwin the covid's after you
Sherwin if it catches you you're through
Tony Thomas

Re: New Homer release

Post by Tony Thomas »

Michael Sherwin wrote:
Daniel Mehrmann wrote:
Werner wrote:
Graham Banks wrote: I'm pretty sure that CEGT also test with learning off.
Regards, Graham.
Bingo!
btw: Can somebody tell me why the learning, specially here Romichess, should be so importent ? I think without learning is it just fair, like gerneral books.

But i can tell you that a guy, if i remember correctly his name was Rodolfo, did some tests with engine learning. He compared Romichess, TheBaron (private version), Homer and Spike with static opening positions.

I was very very suprised as he told me that Homer won this competition. Homer (version 2.0) beated Romichess without any problems as well.
Interesting was that at the beginning Romi was better with his learning, but as longer the match run Homer overrun Romi.

So, like "Vas" style, i'm claiming that Homer has the world strongest (best) learning feature and he just only needs 256kb filesize. :twisted:

Best,
Daniel
Rodolfo never said anything like that to me. Nor AFAIK did he publish those test. If someone wants to test your assertion and make a report then that is fine. My personal test done a long time ago gave a different result--Romi's learning was better!

256kb is enough for testing one static position, maybe. It would not be enough to cover an entire range of openings. I suppose that you could have many such files though if you wanted.

Anyway I was just showing an interesting way of doing learning and never made a claim that it was best. I posted results that showed that Spikes might be better!

If what you do is better, then I am happy for you! :D
Another engine of interest to you might be Frenzee, as a winboard engine it keeps a learn file for each opponent it plays against. Spike also have a positional learning system if I am not mistaken.
User avatar
Daniel Mehrmann
Posts: 858
Joined: Wed Mar 08, 2006 9:24 pm
Location: Germany
Full name: Daniel Mehrmann

Re: New Homer release

Post by Daniel Mehrmann »

Michael Sherwin wrote:
Daniel Mehrmann wrote:
Werner wrote:
Graham Banks wrote: I'm pretty sure that CEGT also test with learning off.
Regards, Graham.
Bingo!
btw: Can somebody tell me why the learning, specially here Romichess, should be so importent ? I think without learning is it just fair, like gerneral books.

But i can tell you that a guy, if i remember correctly his name was Rodolfo, did some tests with engine learning. He compared Romichess, TheBaron (private version), Homer and Spike with static opening positions.

I was very very suprised as he told me that Homer won this competition. Homer (version 2.0) beated Romichess without any problems as well.
Interesting was that at the beginning Romi was better with his learning, but as longer the match run Homer overrun Romi.

So, like "Vas" style, i'm claiming that Homer has the world strongest (best) learning feature and he just only needs 256kb filesize. :twisted:

Best,
Daniel
Rodolfo never said anything like that to me. Nor AFAIK did he publish those test. If someone wants to test your assertion and make a report then that is fine. My personal test done a long time ago gave a different result--Romi's learning was better!

256kb is enough for testing one static position, maybe. It would not be enough to cover an entire range of openings. I suppose that you could have many such files though if you wanted.

Anyway I was just showing an interesting way of doing learning and never made a claim that it was best. I posted results that showed that Spikes might be better!

If what you do is better, then I am happy for you! :D
Of course i was kidding with the learning claim.

As i wrote i was realy suprised and never expect that result. However, there is a lot of improvment space in my learning and of course 256kb is not enought space to handle thousands of different openings in thousands games each other. But it was never my goal to do that. The learning should support the chessplayer with analysis and that works very well. :D

I never looked how your learn stuff works, but what i readed here was interesting, but my learing works completly different.

Best,
Daniel
bob
Posts: 20943
Joined: Mon Feb 27, 2006 7:30 pm
Location: Birmingham, AL

Re: SMP effectivity from Fruit

Post by bob »

Uri Blass wrote:
bob wrote:
Uri Blass wrote:
bob wrote:
Werner wrote:Hi,
I took 10 positions from WM-Test and had following results:
1->2CPU: 142% and
2->4CPU: 146% average
What does this mean?

For example, if it takes 2 minutes to solve N positions, and it take 6 minutes to solve the other N, than a 2 minute search with 2x the processors is going to have no effect on the results.

The right way to measure SMP performance is to take a large set of positions, and search with 1, 2, 4, ..., N processors, and measure the time required to reach a specific depth for each position.

A _much_ less accurate way is to play a one cpu program against a bunch of opponents, then keep everything the same but use two cpus for that program and see if it performs better. It is much less accurate because the variance/randomness in a basic game of chess is extremely high. And when you factor in the randomness produced by a parallel search, it goes even higher. You need thousands of games (at a bare minimum) to get a reasonable estimate on improvement. Most can't pull that off.
You assume that depth means the same with 1 processor and more processors.

It is not obvious.
For example it may be possible that the program does less late move reductions with more processors.

Uri
I don't see why. But if you search the _same_ depth, why would there be any difference with respect to number of processors used? That gives an accurate SMP speedup number. Which is a different number than what you get if you try the same program against the same opponents, but you vary the number of processors. now, rather than measuring exact speedup, you are measuring the effect of searching faster with respect to Elo improvement. Which is also an interesting number, but is independent of the parallel speedup as it is generally defined.
My poiint is that parallel search may change the tree so the same depth may not be equivalent.

if your late move reduction has a rule not to reduce the first 3 moves then with parallel search you may get different moves as the first 3 moves.

Uri
If you implement a parallel algorithm like that, then those of us teaching this stuff would call that "badly flawed".

The first principle of parallelizing something is to ask the question "If I do these "things" in different (or random) order, will it change the result?" If the answer is YES, then we can't do a parallel operation here. If someone implements reductions as you describe, then they have an algorithm flaw that is going to cause problems. Crafty does not behave like this at all. It doesn't consider the first N as non-reducible, rather, it doesn't start reducing until the first few move selection passes are completed to remove the captures, killers, etc, and then it doesn't reduce checks and such. So The same move would not be reduced if I search in one order and be reduced if I search in a different order. To do so would violate an important parallel programming principle.

Now, that said, all parallel chess searches violate the above in a way that can cause an issue here and there, because the pure order the moves are searched in can influence the final result because of the transposition table. This is unsolvable unless the hash table is removed, which is an unreasonable solution, so we just accept the fact that sometimes the search will not return exactly the same value (or best move) if the same search is repeated several times.