Search papers, labs, and topics across Lattice.
This dissertation focuses on improving the deployment-time reliability of learned robot policies by developing mechanisms that operate around them. It introduces runtime monitoring methods to detect impending failures, a data-centric framework for policy interpretability using influence functions, and a method for reliable long-horizon task execution via success probability estimation and feasibility-aware task planning. The research demonstrates advancements in addressing distribution shift, compounding errors, and complex task dependencies, ultimately enhancing the practical use of learned robot policies in real-world scenarios.
Make your robot policies more reliable at deployment time with runtime monitoring, interpretable failure tracing, and success-probability-aware task planning.
Recent advances in learning-based robot manipulation have produced policies with remarkable capabilities. Yet, reliability at deployment remains a fundamental barrier to real-world use, where distribution shift, compounding errors, and complex task dependencies collectively undermine system performance. This dissertation investigates how the reliability of learned robot policies can be improved at deployment time through mechanisms that operate around them. We develop three complementary classes of deployment-time mechanisms. First, we introduce runtime monitoring methods that detect impending failures by identifying inconsistencies in closed-loop policy behavior and deviations in task progress, without requiring failure data or task-specific supervision. Second, we propose a data-centric framework for policy interpretability that traces deployment-time successes and failures to influential training demonstrations using influence functions, enabling principled diagnosis and dataset curation. Third, we address reliable long-horizon task execution by formulating policy coordination as the problem of estimating and maximizing the success probability of behavior sequences, and we extend this formulation to open-ended, language-specified tasks through feasibility-aware task planning. By centering on core challenges of deployment, these contributions advance practical foundations for the reliable, real-world use of learned robot policies. Continued progress on these foundations will be essential for enabling trustworthy and scalable robot autonomy in the future.