Uri Blass wrote:bob wrote:Dirt wrote:bob wrote:First let's settle on a 10 year hardware period. The q6600 is two years old. If you want to use that as a basis, we need to return to early 1997 to choose the older hardware. The Pentium 2 (Klamath) came out around the middle of 1997, which probably means the best was the pentium pro 200. I suspect we are _still_ talking about 200:1
This is not about simple clock frequency improvements, more modern architectures are faster for other reasons such as better speculative execution, more pipelines, register renaming, etc...
Correct me if I'm wrong, but in moving to a time handicap you seem to be ignoring the parallel search inefficiency we were both just explaining to Louis Zulli. Shouldn't that be taken into account?
I don't see why. I used the same parallel search 10 years ago that I use today, the overhead has not changed.
The main point both Don and I have _tried_ to make is that given a certain class of hardware, one is willing or able to do things that are not possible on slower hardware. In Cray Blitz, we specifically made use of vectors, and that gave us some features we could use that would be too costly in a normal scalar architecture. So there are three kinds of improvements over the past 10 years.
1. pure hardware
2. pure software
3. hybrid improvements where improved hardware gave us the ability to do things in software we could not do with previous generations of hardware due to speed issues...
Maybe you use the same parallel search 10 years ago that you use today
but I think that other improved their parallel search so I guess that better parallel search is software improvement unless Crafty is the best software of the beginning of 1999.
I also wonder how you can be sure that you used efficient parallel search for more than 8 cores with Crafty when you even had not the possibility to use 8 cores to test in 1999.
The reason that I suggested to use 8 cores for software with equivalent strength to Fritz of january 1999(when you suggested to use top hardware of today) is that I thought that software of that time could not use more than 8 cores but if you insist on more than 8 cores then I have no objection in case that you also use software of the same time and not something that I consider to be equivalent on 1 core(but not with more than 8 cores).
If you think that old Crafty of january 1999 is the best software of january 1999 on big hardware because it can use many processors efficiently then I have no problem with using it in the test or even with using better hardware for it
in case that it can use it.
I do not believe _any_ commercial program has a better parallel search than what is in Crafty, and what has been in it for 10+ years. There have been changes made, but in 1998 there was no NUMA hardware (AMD started this in the X86 world) so the more recent NUMA-related stuff is completely irrelevant to the 1998 discussion or even todays intel core-3 (I7) processor... In that light, Crafty's parallel search today is almost identical to what it was in 1998.
Again, I have suggested (a) P2/300 single-chip since that was the best available at the end of 1998. And for today, the latest is the Intel Core-3 (I7). the raw speed difference between the two, using Crafty as a benchmark, is a factor of 200. If you want to measure a set of positions with time-to-solution used rather than raw NPS, then today's hardware is around 170-175 times faster. Note this is a single-chip discussion, where in actuality, I can put together far larger systems today than I could in 1998. I elected to keep this simple by using a single chip, which is actually a little unfair to the "hardware side".
The best chip of 1998 was a single core. The best of today is a quad-core. In 1998 you could buy a dual-chip p2/300 if you wanted, I had one. Today you can easily buy a quad core-3 box with 16 cores, which is yet _another_ factor of two. So we should maybe go with a factor of 340:1 rather than 170:1. And larger configurations are available from places like Sun, etc. So we could make that 1,000:1 if you want..
You are talking yourself into one hell of a deep hole here. I believe that with 340:1 time odds, we could take an old buggy gnuchess and give Rybka absolute fits...
I have not studied this much, but I started a test earlier today giving glaurung 1+1, and crafty 100+100 for the time controls. I completed 450 games before quitting, and the result was one draw, the rest wins. with 450 games, if crafty was less than 600 Elo better than Glaurung 2, I would have expected about 1 loss out of every 64 games. For 800 I would expect 1 loss for every 256 games. one loss out of 900 games gives an idea of just what 100:1 does... I'm not sure this experiment will really be that interesting. And considering it should be either 170 or 340 depending on which hardware we consider, it only gets worse.
I have not checked the SSDF to see where G2 (most recent version) compares to Rybka. But it had better be at least 600 worse or this is not going to be so interesting IMHO.
But I do think it would be interesting to assess Elo gain for hardware vs for software, just so we would know. I know the programmers want to take credit for most of the gains. But I'll still bet that the engineers are responsible for the biggest gain...