Conference on Robot Learning • CoRL 2025
An active learning strategy for contextual failure discovery using limited data with expert in loop
Example of a contextual failure discovered using our method
Failure examples for Diffusion Policy trained in Sim for Push-T task on UR3E.
Failure examples for YOLO self-driving object detection in CARLA.
Ensuring the robustness of robotic systems is crucial for their de- ployment in safety-critical domains. Failure discovery, or falsification, is a widely used approach for evaluating robustness, with recent advancements fo- cusing on improving sample efficiency and generalization through probabilistic sampling techniques and learning-theoretic approaches. However, existing methods rely on explicitly defined analytical cost functions to characterize failures, often overlooking the underlying causes and diversity of discovered failure scenarios. In this work, we propose a novel failure discovery framework that integrates contextual reasoning in the falsification process, specifically tailored for high evaluation-cost applications. Our method incorporates expert-in-the-loop feedback to construct a probabilistic surrogate model of failures using Bayesian inference. This model is iteratively refined and leveraged to guide an active learning strategy that prioritizes the discovery of diverse failure cases. We empirically validate our approach across a range of tasks for high-cost contextual falsification in robotic manipulation and autonomous driving.
Reproduce our CARLA and Push-T (sim) results
git clone https://github.com/MIT-REALM/contextual_failure.git
cd contextual_failure
@inproceedings{parashar2025paper,
title={Paper Title},
author={Parashar, Anjali and Coauthors},
booktitle={Proceedings of the Conference on Robot Learning (CoRL)},
year={2025}
}