Why so gloomy? Cost of hardware will go down, and the curent A0 is no doubt over-engineered (burning a few TPU years in parallel must be cheaper than a dozen man years of careful algorithm design). Apart from distributed open source efforts, there are also things like transfer learning and network compression that might greatly reduce the burden of replicating (and then surpasing) the current state of the art. After all, Deep Blue has also been surpassed by today's engines. But first, there needs to be a full paper with enough details to replicate AlphaZero-Chess.Laskos wrote: Yes, but Tord also said the most important thing, IMO. In maybe 5 years, the times of low budget single brilliant author or 1-2-3 brilliant authors of close to top engines will end. Massive hardware will be needed for training the networks, and such giants as Google or whatever have a decisive advantage. Maybe at top will be 2-3 large distributed projects aside from some large companies. Seems a pretty gloomy future.
Reactions about AlphaZero from top GMs...
Moderators: hgm, Rebel, chrisw
-
- Posts: 741
- Joined: Tue May 22, 2007 11:13 am
Re: Reactions about AlphaZero from top GMs...
-
- Posts: 10948
- Joined: Wed Jul 26, 2006 10:21 pm
- Full name: Kai Laskos
Re: Reactions about AlphaZero from top GMs...
The paradigm shifts, nevertheless. Mark Lefler was improving on Komodo using his laptop while on an air flight. These days will be over. And, again, hardware advantage will be decisive, not in running these engines (we will all run them on comparable hardware anyway), but in building that software.Rein Halbersma wrote:Why so gloomy? Cost of hardware will go down, and the curent A0 is no doubt over-engineered (burning a few TPU years in parallel must be cheaper than a dozen man years of careful algorithm design). Apart from distributed open source efforts, there are also things like transfer learning and network compression that might greatly reduce the burden of replicating (and then surpasing) the current state of the art. After all, Deep Blue has also been surpassed by today's engines. But first, there needs to be a full paper with enough details to replicate AlphaZero-Chess.Laskos wrote: Yes, but Tord also said the most important thing, IMO. In maybe 5 years, the times of low budget single brilliant author or 1-2-3 brilliant authors of close to top engines will end. Massive hardware will be needed for training the networks, and such giants as Google or whatever have a decisive advantage. Maybe at top will be 2-3 large distributed projects aside from some large companies. Seems a pretty gloomy future.
-
- Posts: 1346
- Joined: Sat Apr 19, 2014 1:47 pm
Re: Reactions about AlphaZero from top GMs...
Meanwhile, I d like to see more game played and alphazero trained longer ! Anyone knows if more game are coming or if alphazero will be more trained at chess ?
-
- Posts: 1339
- Joined: Fri Nov 02, 2012 9:43 am
- Location: New Delhi, India
Re: Reactions about AlphaZero from top GMs...
+1.JJJ wrote:Meanwhile, I d like to see more game played and alphazero trained longer !
That should happen, otherwise if DeepMind goes quiet, it will only encourage the detractors of AlphaZero to find their voices again.
i7 5960X @ 4.1 Ghz, 64 GB G.Skill RipJaws RAM, Twin Asus ROG Strix OC 11 GB Geforce 2080 Tis
-
- Posts: 10282
- Joined: Thu Mar 09, 2006 12:37 am
- Location: Tel-Aviv Israel
Re: Reactions about AlphaZero from top GMs...
I am not sure if hardware advantage will be decisive in building that software.Laskos wrote:The paradigm shifts, nevertheless. Mark Lefler was improving on Komodo using his laptop while on an air flight. These days will be over. And, again, hardware advantage will be decisive, not in running these engines (we will all run them on comparable hardware anyway), but in building that software.Rein Halbersma wrote:Why so gloomy? Cost of hardware will go down, and the curent A0 is no doubt over-engineered (burning a few TPU years in parallel must be cheaper than a dozen man years of careful algorithm design). Apart from distributed open source efforts, there are also things like transfer learning and network compression that might greatly reduce the burden of replicating (and then surpasing) the current state of the art. After all, Deep Blue has also been surpassed by today's engines. But first, there needs to be a full paper with enough details to replicate AlphaZero-Chess.Laskos wrote: Yes, but Tord also said the most important thing, IMO. In maybe 5 years, the times of low budget single brilliant author or 1-2-3 brilliant authors of close to top engines will end. Massive hardware will be needed for training the networks, and such giants as Google or whatever have a decisive advantage. Maybe at top will be 2-3 large distributed projects aside from some large companies. Seems a pretty gloomy future.
People may come with new ideas how to use slower hardware and still be competitive.
-
- Posts: 273
- Joined: Wed Aug 24, 2016 9:49 pm
Re: Reactions about AlphaZero from top GMs...
I think it's not unlikely that one can improve upon the algorithm deepMind used with some additional domain knowledge. Remember, they wanted to show that their algorithm works for any (board) game but if we are only intrested in chess there should be ways to speed up the training processUri Blass wrote:I am not sure if hardware advantage will be decisive in building that software.Laskos wrote:The paradigm shifts, nevertheless. Mark Lefler was improving on Komodo using his laptop while on an air flight. These days will be over. And, again, hardware advantage will be decisive, not in running these engines (we will all run them on comparable hardware anyway), but in building that software.Rein Halbersma wrote:Why so gloomy? Cost of hardware will go down, and the curent A0 is no doubt over-engineered (burning a few TPU years in parallel must be cheaper than a dozen man years of careful algorithm design). Apart from distributed open source efforts, there are also things like transfer learning and network compression that might greatly reduce the burden of replicating (and then surpasing) the current state of the art. After all, Deep Blue has also been surpassed by today's engines. But first, there needs to be a full paper with enough details to replicate AlphaZero-Chess.Laskos wrote: Yes, but Tord also said the most important thing, IMO. In maybe 5 years, the times of low budget single brilliant author or 1-2-3 brilliant authors of close to top engines will end. Massive hardware will be needed for training the networks, and such giants as Google or whatever have a decisive advantage. Maybe at top will be 2-3 large distributed projects aside from some large companies. Seems a pretty gloomy future.
People may come with new ideas how to use slower hardware and still be competitive.
-
- Posts: 4190
- Joined: Wed Nov 25, 2009 1:47 am
Re: Reactions about AlphaZero from top GMs...
MCTS used in A0 you can probably somewhat improve without having monstrous hardware. However, improving actual NN is really out of reach for regular ppl, at least for next 5-10 years if not longer.CheckersGuy wrote:I think it's not unlikely that one can improve upon the algorithm deepMind used with some additional domain knowledge. Remember, they wanted to show that their algorithm works for any (board) game but if we are only intrested in chess there should be ways to speed up the training processUri Blass wrote:I am not sure if hardware advantage will be decisive in building that software.Laskos wrote:The paradigm shifts, nevertheless. Mark Lefler was improving on Komodo using his laptop while on an air flight. These days will be over. And, again, hardware advantage will be decisive, not in running these engines (we will all run them on comparable hardware anyway), but in building that software.Rein Halbersma wrote:Why so gloomy? Cost of hardware will go down, and the curent A0 is no doubt over-engineered (burning a few TPU years in parallel must be cheaper than a dozen man years of careful algorithm design). Apart from distributed open source efforts, there are also things like transfer learning and network compression that might greatly reduce the burden of replicating (and then surpasing) the current state of the art. After all, Deep Blue has also been surpassed by today's engines. But first, there needs to be a full paper with enough details to replicate AlphaZero-Chess.Laskos wrote: Yes, but Tord also said the most important thing, IMO. In maybe 5 years, the times of low budget single brilliant author or 1-2-3 brilliant authors of close to top engines will end. Massive hardware will be needed for training the networks, and such giants as Google or whatever have a decisive advantage. Maybe at top will be 2-3 large distributed projects aside from some large companies. Seems a pretty gloomy future.
People may come with new ideas how to use slower hardware and still be competitive.
Basically already after 4 hours of training (around 20 million self-played games) current NN performance was saturated.
After that most of games in playouts were draws.
To improve this more one would need larger NN (there are indications used NN was 20 blocks one, not 40), but that would only mean much longer duration of self-play games and more hardware for actual training, so even harder for regular developer without access to monstrous hardware.