People Publications Research Robots Join

Research Talks
  • 2024 Oct: CMU RI Seminar Talk "Building Generalist Robots with Agility via Learning and Control: Humanoids and Beyond" (one hour) [ recording] [ 2025 April version at ETH Zürich] [ slides]
  • 2024 Sep: Georgia Tech IRIM Seminar Talk "Unifying Semantic and Physical Intelligence for Generalist Humanoid Robots" (one hour) [ recording]

Learning and Control for Humanoid Locomotion and Loco-Manipulation

Humanoid robots offer two unparalleled advantages in general-purpose embodied intelligence. First, humanoids are built as generalist robots that can potentially do all the tasks humans can do in complex environments. Second, the embodiment alignment between humans and humanoids allows for the seamless integration of human cognitive skills with versatile humanoid capabilities, making humanoids the most promising physical embodiment for AI.

Humanoid control in a whole-body manner is challenging, due to its high degrees of freedom and contact-rich nature (see this survey). We aim to develop learning-based whole-body control methods for humanoid locomotion and loco-manipulation problems, enabling humanoids to interact with the physical world, adapt quickly to many tasks and environments, and integrate with high-level decision-making layers such as VLMs.

Selected papers in this topic:
ASAP: Aligning Simulation and Real-World Physics for Learning Agile Humanoid Whole-Body Skills
Tairan He*, Jiawei Gao*, Wenli Xiao*, Yuanhang Zhang*, Zi Wang, Jiashun Wang, Zhengyi Luo, Guanqi He, Nikhil Sobanbab, Chaoyi Pan, Zeji Yi, Guannan Qu, Kris Kitani, Jessica Hodgins, Linxi "Jim" Fan, Yuke Zhu, Changliu Liu, Guanya Shi
paper   website   code

TL;DR: ASAP learns agile whole-body humanoid motions via learning a residual action model from the real world to align sim and real physics.

OmniH2O: Universal and Dexterous Human-to-Humanoid Whole-Body Teleoperation and Learning
Tairan He*, Zhengyi Luo*, Xialin He*, Wenli Xiao, Chong Zhang, Weinan Zhang, Kris Kitani, Changliu Liu, Guanya Shi
Conference on Robot Learning (CoRL), 2024
paper   website   dataset   code   IEEE Spectrum

TL;DR: OmniH2O provides a universal whole-body humanoid control interface that enables diverse teleoperation and autonomy methods.

WoCoCo: Learning Whole-Body Humanoid Control with Sequential Contacts
Chong Zhang*, Wenli Xiao*, Tairan He, Guanya Shi
Conference on Robot Learning (CoRL), 2024

(Oral Presentation)

paper   website   IEEE Spectrum

TL;DR: WoCoCo is a task-agnostic skill learning framework without any motion priors, by decomposing long-horizon tasks into contact sequences.



Improve Offline Learned Policies via Online Adaptation

Offline learned policies for robotic control have shown great success in the robot learning community. For example, learning manipulation policies from demonstrations using imitation learning with generative models; learning locomotion policies in simulation using sim2real reinforcement learning. However, those policies are frozen in test time, and cannot efficiently adapt to new environments or tasks. We aim to develop methods that can effectively learn "adaptable" representation from offline data and efficiently adapt in real time.

Selected papers in this topic:
Neural-Fly Enables Rapid Learning for Agile Flight in Strong Winds
Michael O'Connell*, Guanya Shi*, Xichen Shi, Kamyar Azizzadenesheli, Animashree Anandkumar, Yisong Yue, Soon-Jo Chung
Science Robotics
paper   video   Caltech news   Reuters   CNN   code

TL;DR: Neural-Fly uses adaptive control to online fine-tune a meta-pretrained DNN representation, enabling rapid adaptation in strong winds.

DATT: Deep Adaptive Trajectory Tracking for Quadrotor Control
Kevin Huang, Rwik Rana, Alexander Spitzer, Guanya Shi, Byron Boots
Conference on Robot Learning (CoRL), 2023

(Oral presentation, 6.6%)

paper   website   code

TL;DR: DATT can precisely track arbitrary, potentially infeasible trajectories in the presence of large disturbances.

AnyCar to Anywhere: Learning Universal Dynamics Model for Agile and Adaptive Mobility
Wenli Xiao*, Haoru Xue*, Tony Tao, Dvij Kalaria, John M. Dolan, Guanya Shi
Intertional Conference on Robotics and Automation (ICRA), 2025
paper   website   code   IEEE Spectrum

TL;DR: AnyCar is a transformer-based dynamics model that can adapt to various vehicles, environments, state estimators, and tasks.



Real2Sim2Real Reinforcement Learning

Sim2real Reinforcement Learning (RL) has become the dominant method for many locomotion problems, especially for humanoids, legged robots, drones, and ground vehicles. However, the large sim2real gap, tedious reward tuning, and the lack of diversity hinder its applications in other problems such as loco-manipulation and dexterous manipulation. We aim to address these limitations via enhancing the simulation training environment using real-world data, which is often referred to as real2sim2real.

Selected papers in this topic:
ASAP: Aligning Simulation and Real-World Physics for Learning Agile Humanoid Whole-Body Skills
Tairan He*, Jiawei Gao*, Wenli Xiao*, Yuanhang Zhang*, Zi Wang, Jiashun Wang, Zhengyi Luo, Guanqi He, Nikhil Sobanbab, Chaoyi Pan, Zeji Yi, Guannan Qu, Kris Kitani, Jessica Hodgins, Linxi "Jim" Fan, Yuke Zhu, Changliu Liu, Guanya Shi
paper   website   code

TL;DR: ASAP learns agile whole-body humanoid motions via learning a residual action model from the real world to align sim and real physics.



Model-Based RL, World Model, and Sampling-Based Optimal Control

Model-based RL (MBRL) first learns a dynamics model (or a "world" model) and then generates a policy via planning/optimization, policy learning, or search. The most compelling part of MBRL is that the learned model is task-agnostic. We are interested in all aspects of MBRL, including model learning, theoretical framework, and planning using optimal control. In particular, recently, we have developed new sampling-based optimal control frameworks that enable efficient online decision-making.

Selected papers in this topic:
Full-Order Sampling-Based MPC for Torque-Level Locomotion Control via Diffusion-Style Annealing
Haoru Xue*, Chaoyi Pan*, Zeji Yi, Guannan Qu, Guanya Shi
Intertional Conference on Robotics and Automation (ICRA), 2025
paper   website   code

TL;DR: DIAL-MPC is the first training-free method achieving real-time whole-body torque control using full-order dynamics.

TD-M(PC)2: Improving Temporal Difference MPC Through Policy Constraint
Haotian Lin, Pengcheng Wang, Jeff Schneider, Guanya Shi
paper   website   code

TL;DR: We observe the value overestimation issue in planner-based MBRL and propose a policy constraint solution with SOTA performance.

Optimal Exploration for Model-based RL in Nonlinear Systems
Andrew Wagenmaker, Guanya Shi, Kevin Jamieson
Neural Information Processing Systems (NeurIPS), 2023

(Spotlight, 3.1%)

paper

TL;DR: Not all model parameters are equally important. We develop an instance-optimal exploration algorithm for MBRL in nonlinear systems.



Algorithmic and Theoretical Foundations for “Computational Control”

Compared to learning-based methods, one problem of traditional control methods is that their performance cannot effectively scale up with the amount of data and parallel computing. Our focus is around a new concept called computational control, i.e., control methods that can effectively scale up with the amount of data and parallel computing. Examples include sampling-based optimal control and "generative action model" (policy learning using generative models). The goal is to systematically understand, analyze, and enhance computational control methods and apply them to robotics. One of recent interests is to understand how diffusion/flow-based methods work in decision-making problems.

Selected papers in this topic:
Model-Based Diffusion for Trajectory Optimization
Chaoyi Pan*, Zeji Yi*, Guanya Shi, Guannan Qu
Neural Information Processing Systems (NeurIPS), 2024
paper   website   code

TL;DR: MBD is a diffusion-based traj optimization method that directly computes the score function using models without any external data.



Embodied Intelligence in the Air: General-purpose Aerial Manipulation

To some extent, existing works on aerial manipulation primarily focus on the aerial perspective, rather than the general-purpose manipulation perspective. The goal is to study aerial manipulation from the embodied intelligence perspective, building general hardware platforms and control methods. Similar to humanoids, we aim to solve the whole-body control problem for aerial manipulation. We are also interested in designing the high-level policy via vision-language-action (VLA) models or imitation learning.

Selected papers in this topic:
Flying Hand: End-Effector-Centric Framework for Versatile Aerial Manipulation Teleoperation and Policy Learning
Guanqi He*, Xiaofeng Guo*, Luyi Tang, Yuanhang Zhang, Mohammadreza Mousaei, Jiahe Xu, Junyi Geng, Sebastian Scherer, Guanya Shi
Robotics: Science and Systems (RSS), 2025
paper   website

TL;DR: A general-purpose aerial manipulation framework with an EE-centric interface that bridges whole-body control and policy learning.

Flying Calligrapher: Contact-Aware Motion and Force Planning and Control for Aerial Manipulation
Xiaofeng Guo*, Guanqi He*, Jiahe Xu, Mohammadreza Mousaei, Junyi Geng, Sebastian Scherer, Guanya Shi
IEEE Robotics and Automation Letters (RA-L), 2024
paper   website   IEEE Spectrum

TL;DR: Flying calligrapher enables precise hybrid motion and contact force control for an aerial manipulator in various drawing tasks.

Aerial Interaction with Tactile Sensing
Xiaofeng Guo, Guanqi He, Mohammadreza Mousaei, Junyi Geng, Guanya Shi, Sebastian Scherer
International Conference on Robotics and Automation (ICRA), 2024
paper   website

TL;DR: We introduce a new aerial manipulation system that leverages tactile feedback for accurate contact force control and texture detection.



Structured Reinforcement Learning and Control with Guarantees

Most RL algorithms are general for all tasks. In contrast, drastically different control methods are developed for different systems/tasks, and their successes highly rely on structures inside these systems/tasks. We seek to encode these structures and algorithmic principles into black-box RL algorithms, to make RL algorithms more data-efficient, robust, interpretable, and safe. Examples include hierarchical RL and optimal control methods, learning safety filter for RL policies, and learning-based nonlinear control with stability guarantees.

Selected papers in this topic:
Agile But Safe: Learning Collision-Free High-Speed Legged Locomotion
Tairan He*, Chong Zhang*, Wenli Xiao, Guanqi He, Changliu Liu, Guanya Shi
Robotics: Science and Systems (RSS), 2024

(Outstanding Student Paper Award Finalist)

paper   website   code   IEEE Spectrum   CMU News

TL;DR: ABS enables fully onboard, agile (>3m/s), and collision-free locomotion for quadrupedal robots in cluttered environments.

Agile Continuous Jumping in Discontinuous Terrains
Yuxiang Yang, Guanya Shi, Changyi Lin, Xiangyun Meng, Rosario Scalise, Mateo Guaman Castro, Wenhao Yu, Tingnan Zhang, Ding Zhao, Jie Tan, Byron Boots
Intertional Conference on Robotics and Automation (ICRA), 2025
paper   website   code

TL;DR: Continuous, agile, and autonomous quadrupedal jumping via hierarchical model-free RL and model-based control.

Self-Supervised Meta-Learning for All-Layer DNN-Based Adaptive Control with Stability Guarantees
Guanqi He, Yogita Choudhary, Guanya Shi
Intertional Conference on Robotics and Automation (ICRA), 2025
paper   website   code

TL;DR: Pretrain a residual dynamics DNN using meta-learning and fine-tune the whole DNN online using adaptive control with stability guarantees.