Transfer from simulation to real world through learning deep inverse dynamics model

OpenAI Blog Papers

Summary

This paper proposes a method to bridge the simulation-to-real-world gap in robotics by learning a deep inverse dynamics model that maps desired next states (from simulation) to appropriate real-world actions. The approach is evaluated against baselines like output error control and Gaussian dynamics adaptation.

No content available
Original Article Export to Word Export to PDF
View Cached Full Text

Cached at: 04/20/26, 02:45 PM

# Transfer from simulation to real world through learning deep inverse dynamics model Source: [https://openai.com/index/transfer-from-simulation-to-real-world-through-learning-deep-inverse-dynamics-model/](https://openai.com/index/transfer-from-simulation-to-real-world-through-learning-deep-inverse-dynamics-model/) ## Abstract Developing control policies in simulation is often more practical and safer than directly running experiments in the real world\. This applies to policies obtained from planning and optimization, and even more so to policies obtained from reinforcement learning, which is often very data demanding\. However, a policy that succeeds in simulation often doesn't work when deployed on a real robot\. Nevertheless, often the overall gist of what the policy does in simulation remains valid in the real world\. In this paper we investigate such settings, where the sequence of states traversed in simulation remains reasonable for the real world, even if the details of the controls are not, as could be the case when the key differences lie in detailed friction, contact, mass and geometry properties\. During execution, at each time step our approach computes what the simulation\-based control policy would do, but then, rather than executing these controls on the real robot, our approach computes what the simulation expects the resulting next state\(s\) will be, and then relies on a learned deep inverse dynamics model to decide which real\-world action is most suitable to achieve those next states\. Deep models are only as good as their training data, and we also propose an approach for data collection to \(incrementally\) learn the deep inverse dynamics model\. Our experiments shows our approach compares favorably with various baselines that have been developed for dealing with simulation to real world model discrepancy, including output error control and Gaussian dynamics adaptation\.

Similar Articles

Sim-to-real transfer of robotic control with dynamics randomization

OpenAI Blog

OpenAI researchers demonstrate a method to bridge the reality gap in robotic control by training policies with randomized simulator dynamics, enabling robots trained purely in simulation to successfully transfer to real-world tasks like object manipulation without physical training.

World Model for Robot Learning: A Comprehensive Survey

Hugging Face Daily Papers

This comprehensive survey reviews the literature on world models for robot learning, covering their roles in policy learning, planning, and simulation. It highlights key paradigms, benchmarks, and future directions for predictive modeling in embodied agents.

Physics-based Digital Twins for Integrated Thermal Energy Systems Using Active Learning

arXiv cs.LG

This paper proposes an active learning framework to couple high-fidelity Modelica simulations with simpler surrogate models (SINDyC, FNN, GRU) for creating efficient digital twins of thermal energy distribution systems. The approach significantly reduces the number of simulation trajectories needed while maintaining predictive accuracy and enabling uncertainty quantification.

Generalizing from simulation

OpenAI Blog

OpenAI describes challenges with conventional RL on robotics tasks and introduces Hindsight Experience Replay (HER), a new RL algorithm that enables agents to learn from binary rewards by reframing failures as intended outcomes, combined with domain randomization for sim-to-real transfer.