One-shot imitation learning
Summary
OpenAI proposes a meta-learning framework for one-shot imitation learning that enables robots to learn new tasks from a single demonstration and generalize to new instances without task-specific engineering. The approach uses soft attention mechanisms to allow neural networks trained on diverse task pairs to perform well on unseen tasks at test time.
View Cached Full Text
Cached at: 04/20/26, 02:43 PM
Similar Articles
Robots that learn
OpenAI describes a robot learning system powered by two neural networks — a vision network trained on simulated images and an imitation network that generalizes task demonstrations to new configurations. The system is applied to block-stacking tasks, learning to infer and replicate task intent from paired demonstration examples.
Third-person imitation learning
OpenAI presents a method for unsupervised third-person imitation learning that enables agents to learn from demonstrations taken from different viewpoints without explicit state correspondence, using domain confusion techniques to learn viewpoint-agnostic features.
Generalizing from simulation
OpenAI describes challenges with conventional RL on robotics tasks and introduces Hindsight Experience Replay (HER), a new RL algorithm that enables agents to learn from binary rewards by reframing failures as intended outcomes, combined with domain randomization for sim-to-real transfer.
Gotta Learn Fast: A new benchmark for generalization in RL
OpenAI presents a new reinforcement learning benchmark based on Sonic the Hedgehog to measure transfer learning and few-shot learning performance in RL agents, along with baseline algorithm evaluations.
On first-order meta-learning algorithms
This paper analyzes first-order meta-learning algorithms for few-shot learning, introducing Reptile and providing theoretical insights into why these computationally efficient methods work well on established benchmarks.