A paper about parameter tuning

Discussion of chess software programming and technical issues.

Moderators: hgm, Rebel, chrisw

Rémi Coulom
Posts: 438
Joined: Mon Apr 24, 2006 8:06 pm

A paper about parameter tuning

Post by Rémi Coulom »

http://books.nips.cc/papers/files/nips2 ... 9_0508.pdf

I did not read carefully. Some of you might be interested.

Rémi
User avatar
beachknight
Posts: 3533
Joined: Tue Jan 09, 2007 8:33 pm
Location: Antalya, Turkey

Re: A paper about parameter tuning

Post by beachknight »

Quite interesting resource.

http://books.nips.cc/

A query with "chess" resulted with 32 items.

Best,
hi, merhaba, hallo HT
MattieShoes
Posts: 718
Joined: Fri Mar 20, 2009 8:59 pm

Re: A paper about parameter tuning

Post by MattieShoes »

I read it carefully but I still couldn't make sense of the equations. In a general sense, they were doing something similar to what knightcap did, but on more nodes of the search tree than knightcap, resulting in faster learning and better results using self-play...

Bleh, I should back up and learn more about TD learning first, then re-read this.
zamar
Posts: 613
Joined: Sun Jan 18, 2009 7:03 am

Re: A paper about parameter tuning

Post by zamar »

There are two things I don't like in these kind of papers:

1) Simple things are expressed in complicated math symbols. Writer uses great effort to write his ideas in very abstract form and reader uses great effort to extract these ideas from that abstract form. What's the point? I could get ideas much more easily from well documented C-like pseudo-code.

2) What's the point of testing these new kind of learning algorithms using completely random values as starting point? Any bad learning algorithm can improve from there! Pick up some strong existing chess program (Crafty, Fruit, Glaurung) and try to improve from there. Then I believe that this learning algorithm has some potential!
Joona Kiiski
Aaron Becker
Posts: 292
Joined: Tue Jul 07, 2009 4:56 am

Re: A paper about parameter tuning

Post by Aaron Becker »

This looks like a smarter approach than KnightCap, based on the description of KnightCap that they provide. The biggest differences that I see are that 1) all of the parameter updates occur inside a single timestep, so it eliminates a lot of the dependence on opponent strength, and 2) updates aren't limited to root and pv-leaf nodes, so you do a lot more updates per search, and you hit positions that will never occur on the board.

One downside is that since all the updates depend on the fidelity of your own search routine, you can screw up your own learning by doing nullmove search, forward pruning, and possibly to a lesser extend extensions and reductions. The authors get away from this by turning all that stuff off while they do their learning. It's not clear to me how much of a problem this is for the actual learning.

Another downside is that it depends on your evaluation being expressed as a linear combination of weighted binary features. This is fine as a mathematical abstraction, but it would be a real pain to actually express your evaluation that way, which you would need to do in order to do the updates correctly. You're also extremely likely to get stuck in local maxima using a plain gradient descent algorithm like this in a complicated search space.
Michel
Posts: 2272
Joined: Mon Sep 29, 2008 1:50 am

Re: A paper about parameter tuning

Post by Michel »

1) Simple things are expressed in complicated math symbols. Writer uses great effort to write his ideas in very abstract form and reader uses great effort to extract these ideas from that abstract form. What's the point? I could get ideas much more easily from well documented C-like pseudo-code.
It depends on your background. Some people find formulas clearer than text or pseudocode.
mcostalba
Posts: 2684
Joined: Sat Jun 14, 2008 9:17 pm

Re: A paper about parameter tuning

Post by mcostalba »

Michel wrote:
1) Simple things are expressed in complicated math symbols. Writer uses great effort to write his ideas in very abstract form and reader uses great effort to extract these ideas from that abstract form. What's the point? I could get ideas much more easily from well documented C-like pseudo-code.
It depends on your background. Some people find formulas clearer than text or pseudocode.
In a graph is drawn that after 100 games of training ELO increase of 500 points !

This is total crap, of corse, becasue you cannot increase by training with less games it takes to measure the ELO gain itself, and in 100 games you measure nothing, so this is just another piece of crap :-(

Just to be more clear, if the engine increased by 500 ELO points in 100 games it is _theoretically_ only for pure luck.
User avatar
michiguel
Posts: 6401
Joined: Thu Mar 09, 2006 8:30 pm
Location: Chicago, Illinois, USA

Re: A paper about parameter tuning

Post by michiguel »

zamar wrote:
There are two things I don't like in these kind of papers:

1) Simple things are expressed in complicated math symbols. Writer uses great effort to write his ideas in very abstract form and reader uses great effort to extract these ideas from that abstract form. What's the point? I could get ideas much more easily from well documented C-like pseudo-code.

2) What's the point of testing these new kind of learning algorithms using completely random values as starting point? Any bad learning algorithm can improve from there! Pick up some strong existing chess program (Crafty, Fruit, Glaurung) and try to improve from there. Then I believe that this learning algorithm has some potential!
A scientific paper should try to tackle a problem of artificial intelligence, not how to make a competitive chess engine. Of course, you should start from scratch if that is your goal.

Miguel
Last edited by michiguel on Wed Jan 13, 2010 11:38 am, edited 1 time in total.
Gian-Carlo Pascutto
Posts: 1243
Joined: Sat Dec 13, 2008 7:00 pm

Re: A paper about parameter tuning

Post by Gian-Carlo Pascutto »

This is total crap, of corse, becasue you cannot increase by training with less games it takes to measure the ELO gain itself, and in 100 games you measure nothing, so this is just another piece of crap.

Just to be more clear, if the engine increased by 500 ELO points in 100 games it is _theoretically_ only for pure luck.
While there are some things in the paper that can be criticized, your criticism itself is just as flawed.

ELO gains cannot be measured exactly without an infinite number of games, so "measuring ELO gain" can realistically only be done with some error margin.

Measuring an improvement of 500 ELO in 100 games is going to be statistically significant. The error margin might be 300 ELO, but this means worst case there is still an improvement of 200 ELO. An ELO improvement of 500 ELO over 100 games means a score of something like 5 - 95!
Last edited by Gian-Carlo Pascutto on Wed Jan 13, 2010 11:42 am, edited 1 time in total.
User avatar
michiguel
Posts: 6401
Joined: Thu Mar 09, 2006 8:30 pm
Location: Chicago, Illinois, USA

Re: A paper about parameter tuning

Post by michiguel »

Michel wrote:
1) Simple things are expressed in complicated math symbols. Writer uses great effort to write his ideas in very abstract form and reader uses great effort to extract these ideas from that abstract form. What's the point? I could get ideas much more easily from well documented C-like pseudo-code.
It depends on your background. Some people find formulas clearer than text or pseudocode.
Which narrows the audience. I understand that the formalism is most of the time important, but many times it is completely unnecessary.

Science tries to fulfill several goals, among them, accuracy and dissemination. Balance between both is important.

Miguel