http://books.nips.cc/papers/files/nips2 ... 9_0508.pdf
I did not read carefully. Some of you might be interested.
Rémi
A paper about parameter tuning
Moderator: Ras
-
- Posts: 3533
- Joined: Tue Jan 09, 2007 8:33 pm
- Location: Antalya, Turkey
Re: A paper about parameter tuning
Quite interesting resource.
http://books.nips.cc/
A query with "chess" resulted with 32 items.
Best,
http://books.nips.cc/
A query with "chess" resulted with 32 items.
Best,
hi, merhaba, hallo HT
-
- Posts: 718
- Joined: Fri Mar 20, 2009 8:59 pm
Re: A paper about parameter tuning
I read it carefully but I still couldn't make sense of the equations. In a general sense, they were doing something similar to what knightcap did, but on more nodes of the search tree than knightcap, resulting in faster learning and better results using self-play...
Bleh, I should back up and learn more about TD learning first, then re-read this.
Bleh, I should back up and learn more about TD learning first, then re-read this.
-
- Posts: 613
- Joined: Sun Jan 18, 2009 7:03 am
Re: A paper about parameter tuning
There are two things I don't like in these kind of papers:Rémi Coulom wrote:http://books.nips.cc/papers/files/nips2 ... 9_0508.pdf
1) Simple things are expressed in complicated math symbols. Writer uses great effort to write his ideas in very abstract form and reader uses great effort to extract these ideas from that abstract form. What's the point? I could get ideas much more easily from well documented C-like pseudo-code.
2) What's the point of testing these new kind of learning algorithms using completely random values as starting point? Any bad learning algorithm can improve from there! Pick up some strong existing chess program (Crafty, Fruit, Glaurung) and try to improve from there. Then I believe that this learning algorithm has some potential!
Joona Kiiski
-
- Posts: 292
- Joined: Tue Jul 07, 2009 4:56 am
Re: A paper about parameter tuning
This looks like a smarter approach than KnightCap, based on the description of KnightCap that they provide. The biggest differences that I see are that 1) all of the parameter updates occur inside a single timestep, so it eliminates a lot of the dependence on opponent strength, and 2) updates aren't limited to root and pv-leaf nodes, so you do a lot more updates per search, and you hit positions that will never occur on the board.
One downside is that since all the updates depend on the fidelity of your own search routine, you can screw up your own learning by doing nullmove search, forward pruning, and possibly to a lesser extend extensions and reductions. The authors get away from this by turning all that stuff off while they do their learning. It's not clear to me how much of a problem this is for the actual learning.
Another downside is that it depends on your evaluation being expressed as a linear combination of weighted binary features. This is fine as a mathematical abstraction, but it would be a real pain to actually express your evaluation that way, which you would need to do in order to do the updates correctly. You're also extremely likely to get stuck in local maxima using a plain gradient descent algorithm like this in a complicated search space.
One downside is that since all the updates depend on the fidelity of your own search routine, you can screw up your own learning by doing nullmove search, forward pruning, and possibly to a lesser extend extensions and reductions. The authors get away from this by turning all that stuff off while they do their learning. It's not clear to me how much of a problem this is for the actual learning.
Another downside is that it depends on your evaluation being expressed as a linear combination of weighted binary features. This is fine as a mathematical abstraction, but it would be a real pain to actually express your evaluation that way, which you would need to do in order to do the updates correctly. You're also extremely likely to get stuck in local maxima using a plain gradient descent algorithm like this in a complicated search space.
-
- Posts: 2292
- Joined: Mon Sep 29, 2008 1:50 am
Re: A paper about parameter tuning
It depends on your background. Some people find formulas clearer than text or pseudocode.1) Simple things are expressed in complicated math symbols. Writer uses great effort to write his ideas in very abstract form and reader uses great effort to extract these ideas from that abstract form. What's the point? I could get ideas much more easily from well documented C-like pseudo-code.
-
- Posts: 2684
- Joined: Sat Jun 14, 2008 9:17 pm
Re: A paper about parameter tuning
Michel wrote:In a graph is drawn that after 100 games of training ELO increase of 500 points !It depends on your background. Some people find formulas clearer than text or pseudocode.1) Simple things are expressed in complicated math symbols. Writer uses great effort to write his ideas in very abstract form and reader uses great effort to extract these ideas from that abstract form. What's the point? I could get ideas much more easily from well documented C-like pseudo-code.
This is total crap, of corse, becasue you cannot increase by training with less games it takes to measure the ELO gain itself, and in 100 games you measure nothing, so this is just another piece of crap
Just to be more clear, if the engine increased by 500 ELO points in 100 games it is _theoretically_ only for pure luck.
-
- Posts: 6401
- Joined: Thu Mar 09, 2006 8:30 pm
- Location: Chicago, Illinois, USA
Re: A paper about parameter tuning
A scientific paper should try to tackle a problem of artificial intelligence, not how to make a competitive chess engine. Of course, you should start from scratch if that is your goal.zamar wrote:There are two things I don't like in these kind of papers:Rémi Coulom wrote:http://books.nips.cc/papers/files/nips2 ... 9_0508.pdf
1) Simple things are expressed in complicated math symbols. Writer uses great effort to write his ideas in very abstract form and reader uses great effort to extract these ideas from that abstract form. What's the point? I could get ideas much more easily from well documented C-like pseudo-code.
2) What's the point of testing these new kind of learning algorithms using completely random values as starting point? Any bad learning algorithm can improve from there! Pick up some strong existing chess program (Crafty, Fruit, Glaurung) and try to improve from there. Then I believe that this learning algorithm has some potential!
Miguel
Last edited by michiguel on Wed Jan 13, 2010 11:38 am, edited 1 time in total.
-
- Posts: 1260
- Joined: Sat Dec 13, 2008 7:00 pm
Re: A paper about parameter tuning
While there are some things in the paper that can be criticized, your criticism itself is just as flawed.This is total crap, of corse, becasue you cannot increase by training with less games it takes to measure the ELO gain itself, and in 100 games you measure nothing, so this is just another piece of crap.
Just to be more clear, if the engine increased by 500 ELO points in 100 games it is _theoretically_ only for pure luck.
ELO gains cannot be measured exactly without an infinite number of games, so "measuring ELO gain" can realistically only be done with some error margin.
Measuring an improvement of 500 ELO in 100 games is going to be statistically significant. The error margin might be 300 ELO, but this means worst case there is still an improvement of 200 ELO. An ELO improvement of 500 ELO over 100 games means a score of something like 5 - 95!
Last edited by Gian-Carlo Pascutto on Wed Jan 13, 2010 11:42 am, edited 1 time in total.
-
- Posts: 6401
- Joined: Thu Mar 09, 2006 8:30 pm
- Location: Chicago, Illinois, USA
Re: A paper about parameter tuning
Michel wrote:Which narrows the audience. I understand that the formalism is most of the time important, but many times it is completely unnecessary.It depends on your background. Some people find formulas clearer than text or pseudocode.1) Simple things are expressed in complicated math symbols. Writer uses great effort to write his ideas in very abstract form and reader uses great effort to extract these ideas from that abstract form. What's the point? I could get ideas much more easily from well documented C-like pseudo-code.
Science tries to fulfill several goals, among them, accuracy and dissemination. Balance between both is important.
Miguel