Distributionally Robust Imitation Learning: Layered Control Architecture for Certifiable Autonomy

Generative AI & LLMs
Published: arXiv: 2512.17899v1
Authors

Aditya Gahlawat Ahmed Aboudonia Sandeep Banik Naira Hovakimyan Nikolai Matni Aaron D. Ames Gioele Zardini Alberto Speranzon

Abstract

Imitation learning (IL) enables autonomous behavior by learning from expert demonstrations. While more sample-efficient than comparative alternatives like reinforcement learning, IL is sensitive to compounding errors induced by distribution shifts. There are two significant sources of distribution shifts when using IL-based feedback laws on systems: distribution shifts caused by policy error and distribution shifts due to exogenous disturbances and endogenous model errors due to lack of learning. Our previously developed approaches, Taylor Series Imitation Learning (TaSIL) and $\mathcal{L}_1$ -Distributionally Robust Adaptive Control (\ellonedrac), address the challenge of distribution shifts in complementary ways. While TaSIL offers robustness against policy error-induced distribution shifts, \ellonedrac offers robustness against distribution shifts due to aleatoric and epistemic uncertainties. To enable certifiable IL for learned and/or uncertain dynamical systems, we formulate \textit{Distributionally Robust Imitation Policy (DRIP)} architecture, a Layered Control Architecture (LCA) that integrates TaSIL and~\ellonedrac. By judiciously designing individual layer-centric input and output requirements, we show how we can guarantee certificates for the entire control pipeline. Our solution paves the path for designing fully certifiable autonomy pipelines, by integrating learning-based components, such as perception, with certifiable model-based decision-making through the proposed LCA approach.

Paper Summary

Problem
Learning from expert demonstrations, a powerful framework for synthesizing control policies directly from data, has a fundamental limitation: its dependence on training data. Any deviation between the deployment states and those in the expert demonstrations can lead to compounding errors, resulting in degraded performance.
Key Innovation
The researchers introduce a new approach called Distributionally Robust Imitation Policy (DRIP), which combines two existing methods: TaSIL and L1-DRAC. TaSIL learns policies exploiting input-to-state stability properties of the known system, while L1-DRAC guarantees robustness against uncertainty-induced distribution shifts. By integrating these two methods, DRIP achieves guaranteed robustness against both policy and uncertainty-induced distribution shifts.
Practical Impact
This research has significant practical implications for applications where learning from expert demonstrations is crucial, such as robotics, autonomous driving, and aerial vehicles. By providing a certifiably robust approach, DRIP enables safer and more reliable control policies, reducing the risk of compounding errors and improving overall performance.
Analogy / Intuitive Explanation
Imagine you're learning to drive a car from an experienced driver. If you're only shown how to drive on a straight road, but then you're asked to drive on a winding road, you might make mistakes because you haven't learned how to handle the unexpected turns. DRIP is like a training program that not only teaches you how to drive on a straight road but also prepares you for the unexpected turns by learning from the expert driver's experience and incorporating robustness against uncertainty-induced distribution shifts.
Paper Information
Categories:
eess.SY cs.LG
Published Date:

arXiv ID:

2512.17899v1

Quick Actions