Quick answer
AI Summary: Presents MuZero, an algorithm that achieves superhuman performance across board games and Atari by learning an internal model of the environment's dynamics, without needing the rules beforehand.
AI Summary: Presents MuZero, an algorithm that achieves superhuman performance across board games and Atari by learning an internal model of the environment's dynamics, without needing the rules beforehand.
Constructing agents with planning capabilities has long been one of the main challenges in the pursuit of artificial intelligence. Tree-based planning methods have enjoyed huge success in challenging domains, such as chess and Go, where a perfect simulator is available. However, in real-world problems the dynamics governing the environment are often complex and unknown. Here we present the MuZero algorithm, which, by combining a tree-based search with a learned model, achieves superhuman performance in a range of visually complex domains, without any knowledge of their underlying dynamics. MuZero learns a model that, when applied iteratively, predicts the quantities most directly relevant to planning: the reward, the action-selection policy, and the value function.
Share your opinion to help other learners triage faster.
Write a reviewInvite someone by email to share an invited review for Mastering Atari, Go, chess and shogi by planning with a learned model.