Computer Graphics Laboratory ETH Zurich

ETH

VMP: Versatile Motion Priors for Robustly Tracking Motion on Physical Characters

A. Serifi, R. Grandia, E. Knoop, M. Gross, M. Bächer

Proceedings of the 2024 ACM SIGGRAPH/Eurographics Symposium on Computer Animation (Montreal, Canada, Aug 21-23, 2024), pp. 1-11

Abstract

Recent progress in physics-based character control has made it possible to learn policies from unstructured motion data. However, it remains challenging to train a single control policy that works with diverse and unseen motions, and can be deployed to real-world physical robots. In this paper, we propose a two-stage technique that enables the control of a character with a full-body kinematic motion reference, with a focus on imitation accuracy. In a first stage, we extract a latent space encoding by training a variational autoencoder, taking short windows of motion from unstructured data as input. We then use the embedding from the time-varying latent code to train a conditional policy in a second stage, providing a mapping from kinematic input to dynamics-aware output. By keeping the two stages separate, we benefit from self-supervised methods to get better latent codes and explicit imitation rewards to avoid mode collapse. We demonstrate the efficiency and robustness of our method in simulation, with unseen user-specified motions, and on a bipedal robot, where we bring dynamic motions to the real world.




Downloads

Download Paper
[PDF]
Download Paper
[BibTeX]