MuZero: Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model
Summary
Take an AI that learns to play Go and other board games better than any human using only self-play, now make it work for ATARI games. That is exactly what the authors of MuZero have achieved, and in doing so they beat the state-of-the-art on many of the ATARI games. We covered their algorithm in more detail and compare it to AlphaZero.References
- Multiplayer AlphaZero (2019)
Petosa, Nick and Balch, Tucker
- Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model (2019)
Schrittwieser, Julian and Antonoglou, Ioannis and Hubert, Thomas and Simonyan, Karen and Sifre, Laurent and Schmitt, Simon and Guez, Arthur and Lockhart, Edward and Hassabis, Demis and Graepel, Thore and Lillicrap, Timothy and Silver, David
- Mastering Chess and Shogi by Self-Play with a General Reinforcement Learning Algorithm (2017)
Silver, David and Hubert, Thomas and Schrittwieser, Julian and Antonoglou, Ioannis and Lai, Matthew and Guez, Arthur and Lanctot, Marc and Sifre, Laurent and Kumaran, Dharshan and Graepel, Thore and Lillicrap, Timothy and Simonyan, Karen and Hassabis, Demis
- Mastering the Game of Go with Deep Neural Networks and Tree Search (2016)
Silver, David and Huang, Aja and Maddison, Chris J. and Guez, Arthur and Sifre, Laurent and van den Driessche, George and Schrittwieser, Julian and Antonoglou, Ioannis and Panneershelvam, Veda and Lanctot, Marc and Dieleman, Sander and Grewe, Dominik and Nham, John and Kalchbrenner, Nal and Sutskever, Ilya and Lillicrap, Timothy and Leach, Madeleine and Kavukcuoglu, Koray and Graepel, Thore and Hassabis, Demis