TalkingHead-1KH is a talking-head dataset consisting of YouTube videos, originally created as a benchmark for face-vid2vid:
One-Shot Free-View Neural Talking-Head Synthesis for Video Conferencing
Ting-Chun Wang (NVIDIA), Arun Mallya (NVIDIA), Ming-Yu Liu (NVIDIA)
https://nvlabs.github.io/face-vid2vid/
https://arxiv.org/abs/2011.15126.pdf
The dataset consists of 500k video clips, of which about 80k are greater than 512x512 resolution. Only videos under permissive licenses are included. Note that the number of videos differ from that in the original paper because a more robust preprocessing script was used to split the videos. For business inquiries, please visit our website and submit the form: NVIDIA Research Licensing.
First, unzip the metadata and put it under the root directory:
unzip data_list.zip
This step downloads a small subset of the dataset to verify the scripts are working on your computer. You can also skip this step if you want to directly download the entire dataset.
bash videos_download_and_crop.sh small
The processed clips should appear in small/cropped_clips
.
Please run
bash videos_download_and_crop.sh train
The script will automatically download the YouTube videos, split them into short clips, and then crop and trim them to include only the face regions. The final processed clips should appear in train/cropped_clips
.
To download the evaluation set which consists of only 1080p videos, please run
bash videos_download_and_crop.sh val
The processed clips should appear in val/cropped_clips
.
We also provide the reconstruction results synthesized by our model here. For each video, we use only the first frame to reconstruct all the following frames.
Furthermore, for models trained using the VoxCeleb2 dataset, we also provide comparisons using another model trained on the VoxCeleb2 dataset. Please find the reconstruction results here.
The individual videos were published in YouTube by their respective authors under Creative Commons BY 3.0 license. The metadata file, the download script file, the processing script file, and the documentation file are made available under MIT license. You can use, redistribute, and adapt it, as long as you (a) give appropriate credit by citing our paper, (b) indicate any changes that you've made, and (c) distribute any derivative works under the same license.
When collecting the data, we were careful to only include videos that – to the best of our knowledge – were intended for free use and redistribution by their respective authors. That said, we are committed to protecting the privacy of individuals who do not wish their videos to be included.
If you would like to remove your video from the dataset, you can either
- Go to YouTube and change the license of your video, or remove your video entirely.
- Contact [email protected]. Please include your YouTube video link in the email.
This webpage borrows heavily from the FFHQ-dataset page.
If you use this dataset for your work, please cite
@inproceedings{wang2021facevid2vid,
title={One-Shot Free-View Neural Talking-Head Synthesis for Video Conferencing},
author={Ting-Chun Wang and Arun Mallya and Ming-Yu Liu},
booktitle={CVPR},
year={2021}
}