File size: 2,733 Bytes
2d43207
 
 
 
8866d42
2d43207
8866d42
2d43207
8866d42
c1a2fe4
8866d42
feab25f
8866d42
 
 
feab25f
468dced
 
 
 
 
 
 
 
 
 
 
 
 
 
 
577e386
 
 
 
 
 
 
8866d42
2d43207
8866d42
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
---
license: mit
---

# **FaceNormalSeg-ControlNet Dataset**  πŸš€ πŸš€

This is the training dataset for the ControlNet used in the **AnimPortrait3D** pipeline.  

For details about this ControlNet and access to the pretrained models, please visit:  
- **[Project Page](https://onethousandwu.com/animportrait3d.github.io/)**  
- **[Hugging Face Page](https://huggingface.co/onethousand/AnimPortrait3D_controlnet)**  

### **RGB Images**  
For facial RGB images, we use the **FFHQ dataset**. You can download it from **[here](https://github.com/NVlabs/ffhq-dataset)**.  
πŸ”Ή *Note: This dataset only provides annotated face normal maps and face segmentation maps; face RGB images are not included.*  


### Details

For **face** data, we utilize the [FFHQ](https://github.com/NVlabs/ffhq-dataset) and [LPFF](https://github.com/oneThousand1000/LPFF-dataset) (a large-pose variant of FFHQ) datasets. The text prompt for each image is extracted by [BLIP](https://huggingface.co/docs/transformers/main/en/model_doc/blip-2).
Using the **3D face reconstruction** method, we estimate normal maps as geometric conditional signals. 
We then apply [Face Parsing](https://github.com/hukenovs/easyportrait) to segment teeth and eye regions. Additionally, [MediaPipe](https://github.com/google-ai-edge/mediapipe) is used to track iris positions, providing further precision in gaze localization. 
 
 
For **eye** data, we first crop the eye regions from the face dataset. To augment the dataset with closed-eye variations, which are rare in in-the-wild portraits, we use [LivePortrait](https://github.com/KwaiVGI/LivePortrait), a portrait animation method, to generate closed-eye variations from the FFHQ dataset. These closed-eye face images are then processed using a similar methodology to extract conditions, and the eye regions are cropped and added to the eye dataset.

 
To construct the **mouth** dataset, we begin by cropping the mouth regions from the face dataset. To augment this dataset with a broader range of open-mouth variations, we incorporate additional images featuring open-mouth expressions sourced from the [NeRSemble](https://tobias-kirschstein.github.io/nersemble/) dataset. These open-mouth face images are processed using a similar methodology to extract conditions, after which their mouth regions are cropped and integrated into the mouth dataset.



### Download

```
huggingface-cli download onethousand/FaceNormalSeg-ControlNet-dataset   --local-dir ./FaceNormalSeg-ControlNet-dataset --repo-type dataset
```


## **πŸ“¦ Dataset Overview**  

| Region  | Image Count |
|---------|------------|
| Face    | 107,209    |
| Mouth   | 131,758    |
| Eye (left + right) | 214,418    |
| **Total**  | **453,385** |