Vehicle Controls & Behaviors
Annual PlanHybrid Learning-based Policy for High-Speed Autonomous Off-Road Navigation
Project Team
Government
Paramsothy Jayakumar, Calvin Cheung, U.S. Army GVSC
Faculty
Elena Shrestha, U. of Michigan
Industry
Michael McCullough, BAE
Cory Winter, Ahmed Mekky, MathWorks
Student
Jonathan Michaux, U. of Michigan
Project Summary
Project begin 2024.
The overarching goal is to develop a hybrid policy that will enable unmanned autonomous vehicles to adapt to dynamic off-road environments and extend the operational design domains (ODD). The envisioned autonomy stack should be able to use data collected online to compute trajectories that safely navigate over an unknown terrain without operator intervention. Overall, this project aims to address the following question: How can a hybrid policy be updated to account for new data collected from the environment while safely adhering to probabilistic constraints on the system dynamics?
To construct the hybrid policy, we leverage an off-policy reinforcement learning (RL) algorithm (Task 1) that optimizes the policy using an ensemble of experts (Task 2) composed of both physics-based and data-driven models and explores additional states using a risk-aware safety layer that incorporates reachability analysis (Task 3). The key objectives are:
- Set up an off-policy RL framework capable of using data collected both online and offline.
- Develop a hierarchical architecture that learns a modular set of behavior sub-policies using an ensemble of physics-based and data-driven models.
- Develop a reachability-based planning layer that can adjust high-risk trajectories and enable safe exploration of new policies.
The proposed framework facilitates training with data collected online from simulations and with expert demonstration data collected offline from field tests of the ARC MRZR Vehicle Platforms.
#1.42