Mastering Chess and Shogi by Self-Play with a General Reinforcement Learning Algorithm
About
The game of chess is the most widely-studied domain in the history of artificial intelligence. The strongest programs are based on a combination of sophisticated search techniques, domain-specific adaptations, and handcrafted evaluation functions that have been refined by human experts over several decades. In contrast, the AlphaGo Zero program recently achieved superhuman performance in the game of Go, by tabula rasa reinforcement learning from games of self-play. In this paper, we generalise this approach into a single AlphaZero algorithm that can achieve, tabula rasa, superhuman performance in many challenging domains. Starting from random play, and given no domain knowledge except the game rules, AlphaZero achieved within 24 hours a superhuman level of play in the games of chess and shogi (Japanese chess) as well as Go, and convincingly defeated a world-champion program in each case.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Chess Puzzle Solving | Chess Puzzles Novel boards | Puzzle Accuracy95.6 | 12 | |
| Chess Playing | Internal Chess Tournament Large-scale main results | Tournament Elo2.47e+3 | 11 | |
| Move prediction | Maia Master (test) | Move Prediction Accuracy48.69 | 11 | |
| Move prediction | Maia Skilled (test) | Move Prediction Accuracy40.46 | 11 | |
| Move prediction | Maia Advanced (test) | Move Prediction Accuracy44.45 | 11 | |
| Chess | Chess Match vs Stockfish 100 games | Win Rate25 | 2 | |
| Go | Go Match vs AlphaGo Zero 3-day 100 games (train) | Wins31 | 2 | |
| Shogi | Shogi Match vs Elmo 100 games | Win Count47 | 2 |