← Home

Quick answer

AI Summary: Presents MuZero, an algorithm that achieves superhuman performance across board games and Atari by learning an internal model of the environment's dynamics, without needing the rules beforehand.

Claim

Mastering Atari, Go, chess and shogi by planning with a learned model

Julian Schrittwieser·
Ioannis Antonoglou·
Thomas Hubert·
Karen Simonyan·
Laurent Sifre·
David Silver

ABSTRACT

Constructing agents with planning capabilities has long been one of the main challenges in the pursuit of artificial intelligence. Tree-based planning methods have enjoyed huge success in challenging domains, such as chess and Go, where a perfect simulator is available. However, in real-world problems the dynamics governing the environment are often complex and unknown. Here we present the MuZero algorithm, which, by combining a tree-based search with a learned model, achieves superhuman performance in a range of visually complex domains, without any knowledge of their underlying dynamics. MuZero learns a model that, when applied iteratively, predicts the quantities most directly relevant to planning: the reward, the action-selection policy, and the value function.

Review Snapshot

Explore ratings

4.6
★★★★★
5 ratings
5 star
60%
4 star
40%
3 star
0%
2 star
0%
1 star
0%

Recommendation

100%

recommend this content.

Review this content

Share your opinion to help other learners triage faster.

Write a review

Invite a reviewer

Invite someone by email to share an invited review for Mastering Atari, Go, chess and shogi by planning with a learned model.

Author Inquiries

Public questions about this content. Attendemia will route your question to the author. Vote on the most important ones. No guarantee of response.
Post an inquiry
Sort by: Most helpful