Welcome to GeneFace v1.1.0
We have made GeneFace more practical for industrial usage!
What's new in this release:
- We implement RAD-NeRF renderer, which could infer in real-time and be trained in 10 hours.
- We turn to pytorch-based
deep3d_recon
module to extract 3DMM, which is easier to install and is 8x faster than the previous TF-based version. - We provide a pitch-aware audio2motion module, which could generate more lip-sync landmark.
- Fix some bugs that cause large memory usage.
- We will upload the paper about this release soon.
We release the pre-trained models of GeneFace:
lrs3.zip
includes the models trained on LRS3-TED dataset (alm3d_vae_sync
to perform the audio2motion transform and asyncnet
for measuring the lip-sync), which are generic for all possible target person videos.May.zip
includes the models trained on theMay.mp4
target person video (alm3d_postnet_sync
for refining the predicted 3d landmark, alm3d_radnerf
for rendering the head image, and alm3d_radnerf_torso
for rendering the torso part). For each target person video, you need to train these three models.- How to use the pretrained models: unzip the
lrs3.zip
andMay.zip
into thecheckpoints
directory, then follow the commandlines for inference inREADME.md