This project accompanies the release of EMOTE. A temporal, transformer-based VAE called FLINT, is an essential component of EMOTE.
The pretrained model used in EMOTE is publicly available. Go to EMOTE and run download_assets.sh
to download it.
If you want to train your own FLINT, follow the data processing instructions in EMOTE's data processing.
Then run the following:
python training/train_flint.py
Please refer to training/train_flint.py
for additional settings.
If you design a new motion prior, please create a PR. We will merge it. :-)
If you use this work in your publication, please cite the following:
@inproceedings{EMOTE,
title = {Emotional Speech-Driven Animation with Content-Emotion Disentanglement},
author = {Daněček, Radek and Chhatre, Kiran and Tripathi, Shashank and Wen, Yandong and Black, Michael and Bolkart, Timo},
publisher = {ACM},
month = dec,
year = {2023},
doi = {10.1145/3610548.3618183},
url = {https://emote.is.tue.mpg.de/index.html},
month_numeric = {12}
}