By quantifying perception uncertainty and embedding it into robust multi‑agent RL, the approach enables safe, reliable operation of autonomous robots at scale, reducing costly failures and accelerating real‑world adoption.
The keynote by Fei Miao focused on advancing uncertainty understanding and safe, robust reinforcement learning for multi‑agent robotic systems, with autonomous driving as a primary example. Miao highlighted the gap between high‑performance perception models and their lack of calibrated uncertainty, which hampers safety guarantees in real‑time decision making.
The research introduces a novel pipeline that predicts both mean and covariance for perception outputs, then applies statistical calibration—such as moving‑block bootstrap and conformal prediction—to produce reliable uncertainty estimates. This approach boosts detection accuracy, improves tracking of occluded objects, and enhances 3D occupancy predictions, especially for rare, safety‑critical classes like pedestrians and bicycles.
Building on calibrated perception, the team formulates a state‑adversarial Markov game to model multi‑agent reinforcement learning under uncertain observations. They prove that robust Nash equilibria rarely exist, so they adopt worst‑case optimization and integrate adversarial training into policy learning. The resulting algorithm couples discrete RL actions with a continuous Model Predictive Control layer constrained by Control Barrier Functions that incorporate perception uncertainty, achieving 100% safety in simulated intersection and highway scenarios.
Hardware experiments demonstrate that the robust RL‑MPC‑CBF framework transfers zero‑shot to real robots, outperforming traditional domain‑randomization methods and delivering consistent safety margins even under aggressive perturbations. This work signals a shift toward provably safe, uncertainty‑aware AI for large‑scale robotic fleets, paving the way for reliable deployment in warehouses, manufacturing, and autonomous vehicles.
Comments
Want to join the conversation?
Loading comments...