Teaser Video

Abstract

This paper introduces Discrete-time Hybrid Automata Learning (DHAL), a framework using on-policy Reinforcement Learning to identify and execute mode-switching without trajectory segmentation or event function learning. Hybrid dynamical systems, which include continuous flow and discrete mode switching, can model robotics tasks like legged robot locomotion. Model-based methods depend on predefined gaits, while model-free approaches lack explicit mode-switching knowledge. Current methods identify discrete modes via segmentation before regressing continuous flow, but learning high-dimensional complex rigid body dynamics without trajectory labels or segmentation is a challenging open problem. Our approach incorporates a beta policy distribution and a multi-critic architecture to model contact-guided motions, exemplified by a challenging quadrupedal robot skateboard task. We validate our method through simulations and real-world tests, demonstrating robust performance in hybrid dynamical systems.

Skateboard Park

Wild

Indoor

Learning Hybrid Automata

We use different LED lights to indicate transitions between dynamic modes in the automata. Similar to segmentation techniques in computer vision, the learned hybrid modes can help us analyze motion patterns more systematically, improve interpretability in decision-making, and refine control strategies for enhanced adaptability.

Failure Case

Acknowledgements

We appreciate the valuable discussions, hardware guidance and constructive feedback from Yulun Zhuang and Yi Cheng. We also extend our gratitude to Linqi Ye for the initial brainstorming and insightful suggestions, which were inspired by the invaluable time I spent at SHU.