๐ญ SadTalker๏ผ Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation

[](https://colab.research.google.com/github/Winfredy/SadTalker/blob/main/quick_demo.ipynb)
1 Xi'an Jiaotong University 2 Tencent AI Lab 3 Ant Group
CVPR 2023

TL;DR: A realistic and stylized talking head video generation method from a single image and audio.
## ๐ Changelog
- __2023.03.22__: Launch new feature: generating the 3d face animation from a single image. New applications about it will be updated.
- __2023.03.22__: Launch new feature: `still mode`, where only a small head pose will be produced via `python inference.py --still`.
- __2023.03.18__: Support `expression intensity`, now you can change the intensity of the generated motion: `python inference.py --expression_scale 1.3 (some value > 1)`.
- __2023.03.18__: Reconfig the data folders, now you can download the checkpoint automatically using `bash scripts/download_models.sh`.
- __2023.03.18__: We have offically integrate the [GFPGAN](https://github.com/TencentARC/GFPGAN) for face enhancement, using `python inference.py --enhancer gfpgan` for better visualization performance.
- __2023.03.14__: Specify the version of package `joblib` to remove the errors in using `librosa`, [](https://colab.research.google.com/github/Winfredy/SadTalker/blob/main/quick_demo.ipynb) is online!