This is an unofficial implementation of HiFiFace, including some modifications.
Wang, Yuhan, Xu Chen, Junwei Zhu, Wenqing Chu, Ying Tai, Chengjie Wang, Jilin Li, Yongjian Wu, Feiyue Huang, and Rongrong Ji. “HifiFace: 3D Shape and Semantic Prior Guided High Fidelity Face Swapping.” arXiv, June 18, 2021. https://doi.org/10.48550/arXiv.2106.09965.
to better preserve gaze and mouth shape
The project depends on multiple exterior repositories whose codes have been integrated into this repo already. You need to download corresponding model files.
In configs/train_config.py
, modify the dict to the correct paths of your downloaded models:
identity_extractor_config = {
"f_3d_checkpoint_path": "/data/useful_ckpt/Deep3DFaceRecon/epoch_20_new.pth",
"f_id_checkpoint_path": "/data/useful_ckpt/arcface/ms1mv3_arcface_r100_fp16_backbone.pth",
"bfm_folder": "/data/useful_ckpt/BFM",
"hrnet_path": "/data/useful_ckpt/face_98lmks/HR18-WFLW.pth",
}
Download urls:
- Deep3DFaceRecon: https://github.com/sicxu/Deep3DFaceRecon_pytorch#04252023-update
- Arcface: https://1drv.ms/u/s!AswpsDO2toNKq0lWY69vN58GR6mw?e=p9Ov5d
- BFM: you can acquire the bfm data from https://faces.dmi.unibas.ch/bfm/main.php?nav=1-2&id=downloads
the dir structure of the BFM should be like:
BFM +-- [230M] 01_MorphableModel.mat +-- [ 90K] BFM_exp_idx.mat +-- [ 44K] BFM_front_idx.mat +-- [121M] BFM_model_front.mat +-- [ 49M] Exp_Pca.bin +-- [ 428] Exp_Pca.bin?Zone.Identifier +-- [722K] facemodel_info.mat +-- [ 61K] select_vertex_id.mat +-- [ 994] similarity_Lm3D_all.mat +-- [ 615] std_exp.txt
- scrfd: face detection model (for inference). Download from here.
- hrnet(optional): if you want to enable the gaze and mouth shape losses.
- PyTorch >= 2.0
- torchaudio with cuda support
- ffmpeg
- gradio
- etc.
You can also use the pre-built docker image: docker pull xuehy93/hififace
We use a mixed dataset from VGGFace2 HQ, Asia-celeb and some other data collected by ourselves. If you want to train your own model, you may download VGGFace2 HQ and Asia-celeb datasets.
- VGGFace2 HQ: https://github.com/NNNNAI/VGGFace2-HQ
- Asia-Celeb: http://trillionpairs.deepglint.com/data , download the
train_celebrity.tar.gz
We provide two pretrained models here , one is trained without gaze and mouth shape loss and the other is trained with attributes loss:
- standard_model
- with_gaze_and_mouth
a demo based on gradio is provided: app/app.py
You need to download all auxiliary models and pretrained models, then modify the model paths in app/app.py
and configs/train_config.py
see entry/train.py
We provide inference demo for image based face swap on huggingface with both models: HiFiFace-inference-demo
We find that the attributes of the faces such as eye gaze directions and mouth shapes cannot be well preserved after face swapping, and therefore we introduce some auxiliary losses including the eye heat map loss and the mouth heatmap loss.
You can enable them or disable them by setting
eye_hm_loss
and mouth_hm_loss
in configs/train_config.py
- The official repo https://github.com/johannwyh/HifiFace although they provide no codes, the disccusions in the issues are helpful.
Currently our implementation is not perfect:
- The attributes such as gaze and mouth shapes cannot be well preserved with the original model proposed in the paper. This problem is obvious in video face swapping results.
- With the eye and mouth heat map losses, the attributes are better preserved, however, the face swap similarity drops significantly.
Any dicussions and feedbacks are welcome!!!