IROS 2025
Adam Wei, Abhinav Agarwal, Boyuan Chen, Rohan Bosworth, Nicholas Pfaff, Russ Tedrake
Cotrain from both simulated and real-world robot data to maximize a real-world performance objective
50 real demos
50 real demos
2000 sim demos
Success rate: 10/20
Success rate: 19/20
2x
2x
Policy:
Performance Objective:
Success rate on a planar pushing task
Diffusion Policy [2]
[2] Chi et. al "Diffusion Policy: Visuomotor Policy Learning via Action Diffusion"
[1] Graedsal et. al, "Towards Tight Convex Relaxations For Contact-Rich Manipulation"
Low Real Data Regime
Med Real Data Regime
High Real Data Regime
Low Real Data Regime
Med Real Data Regime
High Real Data Regime
How do visual, physics, and task shift on performance?
Reducing all shifts improves performance
Paradoxically, some visual shift is required for good performance!
Cotraining performance is more sensitive to physics and task shifts.
Real-World Demo
Policy Rollout
(Cotrained)
Simulated Demo
High-performing policies must learn to identify sim vs real
since the physics of each environment requires different actions
2x
2x
2x
Sim data improves dataset coverage
Training on sim reduces test loss in real (power law)
Please refer to the paper for more experiments and details.