metadata
license: apache-2.0
tags:
- text-to-motion
- text-to-image
datasets:
- HumanML3D
- KIT-ML
Model Description
These are model weights originally provided by the authors of the paper T2M-GPT: Generating Human Motion from Textual Descriptions with Discrete Representations.
Conditional generative framework based on Vector QuantisedVariational AutoEncoder (VQ-VAE) and Generative Pretrained Transformer (GPT) for human motion generation from textural descriptions.
A simple CNN-based VQ-VAE with commonly used training recipes (EMA and Code Reset) allows us to obtain high-quality discrete representations
The official code of this paper in here
Example
Datasets
HumanML3D and KIT-ML