Perceiving Systems, Computer Vision

ACTOR: Action-Conditioned 3D Human Motion Synthesis with Transformer VAE

2021-10-13


ACTOR learns an action-aware latent representation for human motions by training a generative variational autoencoder (VAE). By sampling from this latent space and querying a certain duration through a series of positional encodings, we synthesize variable-length motion sequences conditioned on a categorical action. ACTOR uses a transformer-based architecture to encode and decode a sequence of parametric SMPL human body models estimated from action recognition datasets.

ACTOR learns an action-aware latent representation for human motions by training a generative variational autoencoder (VAE). By sampling from this latent space and querying a certain duration through a series of positional encodings, we synthesize variable-length motion sequences conditioned on a categorical action. ACTOR uses a transformer-based architecture to encode and decode a sequence of parametric SMPL human body models estimated from action recognition datasets.

Author(s): Mathis Petrovich, Michael J. Black, Gül Varol
Department(s): Perceiving Systems
Publication(s): Action-Conditioned {3D} Human Motion Synthesis with Transformer {VAE}
Authors: Mathis Petrovich, Michael J. Black, Gül Varol
Release Date: 2021-10-13
Repository: https://github.com/Mathux/ACTOR