← Home

Quick answer

AI Summary: Introduces RT-1, a Robotics Transformer that leverages a massive 130k-episode real-world dataset to achieve a 97% success rate on robotic manipulation tasks, demonstrating unprecedented zero-shot generalization.

Claim

RT-1: Robotics Transformer for Real-World Control at Scale

Anthony Brohan·
Noah Brown·
Justice Carbajal·
Yevgen Chebotar·
Joseph Dabis·
Chelsea Finn·
Keerthana Gopalakrishnan·
Karol Hausman·
Alex Herzog·
Jasmine Hsu·
Julian Ibarz·
Brian Ichter·
Alex Irpan·
Tomas Jackson·
Google DeepMind

ABSTRACT

By transferring knowledge from large, diverse, task-agnostic datasets, modern machine learning models can solve specific downstream tasks either zero-shot or with small task-specific datasets. We investigate whether this paradigm can be applied to robotics. We present RT-1, a Robotics Transformer model trained on a massive real-world dataset of 130k episodes, collected over 17 months by a fleet of 13 robots. RT-1 can execute over 700 linguistic instructions with a 97% success rate. The model demonstrates unprecedented generalization capabilities, executing novel instructions, operating in unseen backgrounds, and manipulating unobserved objects. We show that simply scaling data and model size systematically improves the robustness of end-to-end robotic control.

Review Snapshot

Explore ratings

4.6
★★★★★
5 ratings
5 star
60%
4 star
40%
3 star
0%
2 star
0%
1 star
0%

Recommendation

100%

recommend this content.

Review this content

Share your opinion to help other learners triage faster.

Write a review

Invite a reviewer

Invite someone by email to share an invited review for RT-1: Robotics Transformer for Real-World Control at Scale.

Author Inquiries

Public questions about this content. Attendemia will route your question to the author. Vote on the most important ones. No guarantee of response.
Post an inquiry
Sort by: Most helpful
RT-1: Robotics Transformer for Real-World Control at Scale | Attendemia