Search papers, labs, and topics across Lattice.
The paper introduces SPARR, a hybrid approach for robotic assembly that combines a simulation-trained base policy with a real-world residual policy to address the sim-to-real gap. The simulation policy uses low-level state observations and dense rewards, while the real-world policy uses visual observations and sparse rewards to compensate for discrepancies in dynamics and sensor noise. Real-world experiments show that SPARR achieves near-perfect success rates across diverse two-part assembly tasks, improving success rates by 38.4% and reducing cycle time by 29.7% compared to state-of-the-art zero-shot sim-to-real methods.
By combining simulation-trained priors with real-world adaptation, SPARR achieves near-perfect success in robotic assembly without human supervision, outperforming both sim-to-real and real-world RL baselines.
Robotic assembly presents a long-standing challenge due to its requirement for precise, contact-rich manipulation. While simulation-based learning has enabled the development of robust assembly policies, their performance often degrades when deployed in real-world settings due to the sim-to-real gap. Conversely, real-world reinforcement learning (RL) methods avoid the sim-to-real gap, but rely heavily on human supervision and lack generalization ability to environmental changes. In this work, we propose a hybrid approach that combines a simulation-trained base policy with a real-world residual policy to efficiently adapt to real-world variations. The base policy, trained in simulation using low-level state observations and dense rewards, provides strong priors for initial behavior. The residual policy, learned in the real world using visual observations and sparse rewards, compensates for discrepancies in dynamics and sensor noise. Extensive real-world experiments demonstrate that our method, SPARR, achieves near-perfect success rates across diverse two-part assembly tasks. Compared to the state-of-the-art zero-shot sim-to-real methods, SPARR improves success rates by 38.4% while reducing cycle time by 29.7%. Moreover, SPARR requires no human expertise, in contrast to the state-of-the-art real-world RL approaches that depend heavily on human supervision.