DisCoHead: Audio-and-Video-Driven Talking Head Generation by Disentangled Control of Head Pose and Facial Expressions
Geumbyeol Hwang, Sunwon Hong, Seunghyun Lee, Sungwoo Park, and Gyeongsu Chae.
- CUDA 10.2
- PyTorch 1.10.0
- Python 3.7
You can install environements by 1. or 2.
- Copy created conda environment
git clone https://github.com/deepbrainai-research/discohead
cd discohead
conda env create -f discohead.yaml
conda activate discohead
- Install requirements yourself
git clone https://github.com/deepbrainai-research/discohead
cd discohead
conda env create -n discohead python=3.7
conda activate discohead
conda install pytorch==1.10.0 torchvision==0.11.1 torchaudio==0.10.0 cudatoolkit=10.2 -c pytorch
pip install -r requirements.txt
- Download the pre-trained checkpoints from google drive and put into
weight
folder - Create the
dataset
folder. - Unzip the
dataset_demo.zip
todataset
DiscoHead
directory should have the following structure
DiscoHead/
├── dataset/
│ ├── obama/
│ │ ├──demo1/
│ │ ├──demo2/
│ ├── grid/
│ │ ├──demo1/
│ │ ├──demo2/
│ ├── koeba/
│ │ ├──demo1/
│ │ ├──demo2/
├── weight/
│ ├── obama.pt
│ ├── grid.pt
│ ├── koeba.pt
‥‥
- The
--mode
argument is used for specifying which demo you want to generate - To reproduce demo1 of obama dataset, run command :
python test.py -m obama_demo1