Spatio-Temporal Motion Retargeting
for Quadruped Robots

🎉 Accepted to IEEE Transactions on Robotics (T-RO)

Real-world experiments

Backflip-35cm

Backflip-25cm

Hopturn-B2

Sidesteps-B2

Hopturn-Aliengo

Sidesteps-Aliengo

Hopturn-Go1

Sidesteps-Go1

Baseless motion from videos

Reconstruction and deployment

Abstract

This work presents a motion retargeting approach for legged robots, aimed at transferring the dynamic and agile movements to robots from source motions. In particular, we guide the imitation learning procedures by transferring motions from source to target, effectively bridging the morphological disparities while ensuring the physical feasibility of the target system. In the first stage, we focus on motion retargeting at the kinematic level by generating kinematically feasible whole-body motions from keypoint trajectories. Following this, we refine the motion at the dynamic level by adjusting it in the temporal domain while adhering to physical constraints. This process facilitates policy training via reinforcement learning, enabling precise and robust motion tracking. We demonstrate that our approach successfully transforms noisy motion sources, such as hand-held camera videos, into robot-specific motions that align with the morphology and physical properties of the target robots. Moreover, we demonstrate terrain-aware motion retargeting to perform BackFlip on top of a box. We successfully deployed these skills to four robots with different dimensions and physical properties in the real world through hardware experiments.

Q&A

Q: What distinguishes STMR from previous works on motion retargeting?
A: Spatio-Temporal Motion Retargeting (STMR) lets quadruped robots imitate agile motions from videos or handcrafted animations, even when base motion is missing. It reconstructs the full-body movement—including the base and legs—from only joint poses and contact booleans, enabling motion retargeting directly from video. STMR then refines the temporal aspect of the motion to ensure it is physically feasible for the robot, especially crucial for dynamic movements with flight phases, such as hoping or back flipping. This allows robots to perform complex motions in the real world without manual tuning or per-motion optimization.

BibTeX

Methods

Spatial Motion Retargeting (SMR)

SMR focuses on transferring motion to a target robot with a different morphology at the kinematic level. The motions generated by SMR are kinematically feasible and avoid common artifacts such as foot sliding and ground penetration. A key feature of SMR is its ability to reconstruct whole-body motion from baseless motion inputs.

Overview

SMR Problem
Baseline methods can generate inappropriate base trajectories, leading to foot sliding and penetration.

SMR Detail
Our proposed method generates kinematically feasible whole-body motions, with base trajectory adjustment being a key feature that distinguishes it from prior approaches.

Adjusted base trajectories

SMR adjusts the base trajectories to match the target robot's morphology.

Reconstruction from baseless motion

SMR can generate the whole-body motion from the baseless motion.

Adaptation to different morphologies

The reconstruction can be applied to robots with different morphologies, resulting in distinct base trajectories suited to each robot.

Temporal Motion Retargeting (TMR)

TMR focuses on dynamic motion retargeting, particularly for agile and time-critical behaviors such as jumping. It adjusts the timing of reference motions using model-based control to ensure physical feasibility on the target robot. This allows the motion to adapt to the robot's dynamics and morphology.

TMR Detail
Using model-based control as an internal process, we search for optimal temporal parameters.
Visualization of the temporal optimization process. We deform the motion in the temporal domain and attempt to track it with a model-based controller.

Ablation Study: Is Temporal Optimization Necessary?

Original Motion

Retargeting without temporal optimization: Fails

Retargeting with temporal optimization: Succeeds


Acknowledgements

This work was supported by the Institute of Information & Communications Technology Planning & Evaluation (IITP) grants funded by the Korea government (MSIT): (No. RS-2019-II190079, Artificial Intelligence Graduate School Program at Korea University, 12.5%), (No. RS-2024-00457882, AI Research Hub Project, 12.5%), (No. RS-2022-II220871, AI Autonomy and Knowledge Enhancement for AI Agent Collaboration, 12.5%), (No. RS-2022-II220480, Development of Training and Inference Methods for Goal-Oriented Artificial Intelligence Agent, 12.5%), and (No. RS-2024-00336738, Development of Complex Task Planning Technologies for Autonomous Agents, 50%). Additionally, this work utilized research resources sponsored by the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (Grant Agreement No. 866480).