IROS 2025 Keynotes - Learning and Embodied Control: Abhinav Valada
Why It Matters
By enabling label‑efficient perception, continual adaptation, and offline policy refinement, these methods bring truly autonomous, everyday robots closer to commercial viability, reducing deployment costs and expanding their functional scope.
Key Takeaways
- •Open‑world robot autonomy demands continual learning from diverse data.
- •Foundation models enable label‑efficient perception with minimal annotations.
- •Continual SLAM balances adaptation and memory across environments.
- •Diffusion policy adaptation uses world models for offline skill improvement.
- •Neural navigation integrates base motion and manipulation intent for mobile robots.
Summary
Abhinav Valada’s IROS 2025 keynote outlines a roadmap toward open‑world autonomy for everyday robots, emphasizing that true utility requires systems that can learn continuously across heterogeneous environments. He frames the challenge with a data pyramid—ranging from scarce, high‑quality tele‑operated robot data to abundant video streams—and asks how to fuse these tiers into robust policies. The talk then surveys concrete advances: foundation‑model‑driven perception that matches fully supervised performance with only ten labeled images, open‑set segmentation of unseen objects, and 3‑D scene‑graph representations for planning. He introduces continual SLAM, a dual‑network architecture that retains knowledge from prior scenes while adapting online, and demonstrates superior odometry across multiple city datasets. Further, Valada presents Artipoint for training‑free articulation reasoning from human demonstrations, and D.VA, which fine‑tunes diffusion policies entirely offline inside a learned world model, achieving drawer‑opening performance without any real‑world interactions. Finally, he showcases neural navigation for mobile manipulation, where a reinforcement‑learning base controller is conditioned on end‑effector intent, enabling zero‑shot task generalization across robots and dynamic environments. The overarching message is that perception must be label‑efficient, domain‑aware, and continuously improving, while policies should adapt offline using simulated rollouts and integrate navigation with manipulation intent. These innovations collectively lower the data‑collection burden, close the sim‑to‑real gap, and make robots more reliable in human‑centric settings, paving the way for service and industrial agents that can operate safely and autonomously throughout an entire day. The significance lies in demonstrating that robots can acquire and refine complex skills without extensive real‑world trial‑and‑error, dramatically accelerating deployment of adaptable, autonomous agents in homes, factories, and public spaces.
Comments
Want to join the conversation?
Loading comments...