Our grand finale: A complex, long-horizon dynamic sequence, all driven by a proprioceptive-only policy (no vision/LIDAR)! In this task, the robot carries a chair to a platform, uses it as a step to climb up, then leaps off and performs a parkour-style roll to absorb the landing. This pushes the boundaries of agile, human-like loco-manipulation!
All robot videos are real-time, with RL policies that are trained with shared 5 rewards & 4 domain randomization terms, and rely only on proprioception.
A dominant paradigm for teaching humanoid robots complex skills is to retarget human motions as kinematic references to train reinforcement learning (RL) policies. However, existing retargeting pipelines often struggle with the significant embodiment gap between humans and robots, producing physically implausible artifacts like foot-skating and penetration. More importantly, common retargeting methods neglect the rich human-object and human-environment interactions essential for expressive locomotion and loco-manipulation. To address this, we introduce OmniRetarget, an interaction- preserving data generation engine based on an interaction mesh that explicitly models and preserves the crucial spatial and contact relationships between an agent, the terrain, and manipulated objects. By minimizing the Laplacian deformation between the human and robot meshes while enforcing kinematic constraints, OmniRetarget generates kinematically feasible trajectories. Moreover, preserving task-relevant interactions enables efficient data augmentation, from a single demonstration to different robot embodiments, terrains, and object configurations. We comprehensively evaluate OmniRetarget by retargeting motions from OMOMO, LAFAN1, and our in-house MoCap datasets, generating over 9-hour trajectories that achieve better kinematic constraint satisfaction and contact preservation than widely used baselines. Such high-quality data enables proprioceptive RL policies to successfully execute long- horizon (up to 30 seconds) parkour and loco-manipulation skills on a Unitree G1 humanoid, trained with only 5 reward terms and simple domain randomization shared by all tasks, without any learning curriculum.
Rolling
Platform Climbing 1
Platform Climbing 2
Platform Climbing 3
Platform Climbing 4
Platform Climbing and Sitting
Stepping
Crawling 1
Crawling 2
Crawling 3
Crawling 4
Box Carrying 1
Box Carrying 2
Box Carrying 3
Box Carrying 4
Box Carrying 5
Box Carrying 6
Box Carrying 7
Box Carrying 8
Each row comes from a single demonstration (middle column), with augmented initial object pose, object size and terrain height.
Explore how OmniRetarget preserves interactions across different object initial poses, sizes, terrains and robot embodiments.
(Transparent is the original demonstration. Open Controls for more options.)
Interactive 3D visualization with initial pose variations
Interactive 3D visualization with original box size
Interactive 3D visualization with original terrain height
Interactive 3D visualization with T1 robot embodiment
Explore how OmniRetarget obeys non-penetration constraints and prevents foot-skating.
Select baseline method and task to compare with OmniRetarget on the same task
GMR baseline on box task
OmniRetarget results on box task
Interactive 3D visualization of LAFAN1 sequences