Robot learning requires a considerable amount of data to realize the promise of generalization. However, it can be challenging to actually collect the magnitude of high-quality data necessary for generalization entirely in the real world. Simulation can serve as a source of plentiful data, wherein techniques such as reinforcement learning can obtain broad coverage over states and actions. However, high-fidelity physics simulators are fundamentally misspecified approximations to reality, making direct zero-shot transfer challenging, especially in tasks where precise and forceful manipulation is necessary. This makes real-world fine-tuning of policies pretrained in simulation an attractive approach to robot learning. However, exploring the real-world dynamics with standard RL fine-tuning techniques is to inefficient for many real-world applications. This paper introduces Simulation-Guided Fine-Tuning (SGFT), a general framework which leverages the structure of the simulator to guide exploration, substantially accelerating adaptation to the real-world. We demonstrate our approach across several manipulation tasks in the real world, learning successful policies for problems that are challenging to learn using purely real-world data. We further provide theoretical backing for the paradigm.
How do we solve contact-rich manipulation in situations where sim2real transfer fails?
Problem:
Main Idea:
*Note: The RL objective is now biased. Theoretical analysis in the paper shows this bias is acceptable.
*For fine-tuned insertion policy video, we roll in with the pretrained policy to grasp and switch to the fine-tuned insertion policy
@inproceedings{yin2025sgft,
author = {Yin, Patrick and Westenbroek, Tyler and Bagaria, Simran and Huang, Kevin and Cheng, Ching-An and Kolobov, Andrey and Gupta, Abhishek},
title = {Rapidly Adapting Policies to the Real-World via Simulation-Guided Fine-Tuning},
booktitle = {International Conference on Learning Representations (ICLR)},
year = {2025},
}