Datasets:

Languages:
English
ArXiv:
License:
File size: 9,282 Bytes
58ff142
 
 
 
 
 
748df47
 
 
 
 
0dccf3b
748df47
6177c8d
4b75678
6177c8d
748df47
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ec40962
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
748df47
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16729f9
 
 
 
 
 
748df47
ec40962
2db117f
 
 
 
 
 
 
 
 
 
 
ec40962
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
---
license: apache-2.0
language:
- en
size_categories:
- 100M<n<1B
---


# Dataset Card for  Dataset Curation of 3DXTalker

### Dataset Description

- **Repository:** [https://github.com/EngineeringAI-LAB/3DXTalker/tree/main]
- **Paper:** [https://arxiv.org/abs/2602.10516]
- **Project :** [https://engineeringai-lab.github.io/3DXTalker.github.io/]

### Dataset Summary

This dataset is a large-scale, curated collection of talking head videos built for tasks such as high-fidelity 3D talking avatar generation, lip synchronization, and pose dynamics modeling. 

The dataset aggregates and standardizes data from six prominent sources (**GRID, RAVDESS, MEAD, VoxCeleb2, HDTF, Celebv-HQ**), processed through a rigorous data curation pipeline to ensure high quality in terms of face alignment, resolution, and audio-visual synchronization. It covers diverse environments (Lab vs. Wild) and a wide range of subjects.

### Supported Tasks and Leaderboards

- **3D Talking Head Generation:** Synthesizing realistic talking videos from driving speech.
- **Audio-Driven Lip Synchronization:** Aligning lip movements precisely with input speech.
- **Emotion Analysis & Synthesis:** Leveraging the emotional diversity in datasets like RAVDESS and MEAD.
- **Audio-Driven Head Pose Synthesis:** Modeling natural head movements and orientation directly driving speech.

## Dataset Structure
```
trainset/

├── V0-GRID/ # 6,570 sequences from GRID corpus

│ ├── V0-s1-00001/

│ │ ├── audio.wav # (N,) audio data

│ │ ├── cam.npy # (T, 3) camera parameters

│ │ ├── detailcode.npy # (T, 128) facial details

│ │ ├── envelope.npy # (N,) audio envelope

│ │ ├── expcode.npy # (T, 50) expression codes

│ │ ├── lightcode.npy # (T, 9, 3) lighting

│ │ ├── metadata.pkl # Sequence metadata

│ │ ├── posecode.npy # (T, 6) head pose

│ │ ├── refimg.npy # (C, H, W) reference image

│ │ ├── shapecode.npy # (T, 100) shape codes

│ │ └── texcode.npy # (T, 50) texture codes

│ ├── V0-s1-00002/

│ │ └── ... (same 11 files)

│ ├── V0-s1-00003/

│ └── ... (6,570 total sequences)



├── V1-RAVDESS/ # 583 sequences from RAVDESS dataset

│ ├── V1-Song-Actor_01-00001/

│ │ └── ... (same 11 files)

│ ├── V1-Song-Actor_01-00002/

│ ├── V1-Speech-Actor_01-00001/

│ ├── V1-Speech-Actor_02-00001/

│ └── ... (583 total sequences)



├── V2-MEAD/ # 1,939 sequences from MEAD dataset

│ ├── V2-M003-angry-00001/

│ │ └── ... (same 11 files)

│ ├── V2-M003-angry-00002/

│ ├── V2-M003-happy-00001/

│ ├── V2-W009-sad-00001/

│ └── ... (1,939 total sequences)



├── V3-VoxCeleb2/ # 1,296 sequences from VoxCeleb2

│ ├── {sequence_id}/

│ │ └── ... (same 11 files)

│ └── ... (1,296 total sequences)



├── V4-HDTF/ # 350 sequences from HDTF dataset

│ ├── {sequence_id}/

│ │ └── ... (same 11 files)

│ └── ... (350 total sequences)



└── V5-CelebV-HQ/ # 768 sequences from CelebV-HQ dataset

├── {sequence_id}/

│ └── ... (same 11 files)

└── ... (768 total sequences)

```


## Data Format Details

### File Overview
| File | Type | Shape | Description |
|------|------|-------|-------------|
| `audio.wav` | Audio | (N_samples,) | Original audio waveform|
| `cam.npy` | Parameters | (N_frames, 3) | Camera parameters (position/scale) |
| `detailcode.npy` | Parameters | (N_frames, 128) | Facial detail codes (wrinkles, fine features) |
| `envelope.npy` | Parameters | (N_audio_samples,) | Audio envelope/amplitude over time |
| `expcode.npy` | Parameters | (N_frames, 50) | FLAME expression parameters (50-dim) |
| `lightcode.npy` | Parameters | (N_frames, 9, 3) | Spherical harmonics lighting (9 bands × RGB) |
| `metadata.pkl` | Metadata | N/A | Sequence metadata (integer or dict) |
| `posecode.npy` | Parameters | (N_frames, 6) | 3 head pose + 3 jaw pose |
| `refimg.npy` | Image | (3, 224, 224) | Reference image (RGB, 224×224 pixels) |
| `shapecode.npy` | Parameters | (N_frames, 100) | FLAME shape parameters (100-dim) |
| `texcode.npy` | Parameters | (N_frames, 50) | Texture codes (50-dim) |

### Coordinate Systems and Conventions
- **FLAME model**: 3D Morphable Face Model with 5023 vertices
- **Expression space**: 50-dimensional linear basis
- **Shape space**: 100-dimensional PCA space
- **Pose representation**: 3 head pose + 3 jaw pose
- **Lighting**: 2nd-order spherical harmonics (9 bands)

### Temporal Synchronization
- **Video frames**: 25 FPS (frames per second)
- **Audio samples**: 16,000 samples per second
- All video parameters (`expcode`, `shapecode`, `detailcode`, `posecode`, `cam`, `lightcode`, `texcode`) share the same `N_frames` dimension
- Audio and video are temporally aligned (frame 0 corresponds to start of audio)



### Data Statistics

The dataset comprises **11,706** total video samples, spanning approximately **67.4 hours** of self-talking footage. The data is categorized by environment (Lab vs. Wild) and includes varying resolutions and subject diversity.

#### Detailed Statistics (from Curation Pipeline)

| Dataset     | ID | Environment | Year | Raw Resolution | Size (samples) | Subject | Total Duration (s) | Hours (h) | Avg. Duration (s/sample) |
|-------------|----|-------------|------|----------------|----------------|---------|--------------------|-----------|--------------------------|
| **GRID** | V0 | Lab         | 2006 | 720 × 576      | 6,600          | 34      | 99,257.81          | 27.57     | 15.04                    |
| **RAVDESS** | V1 | Lab         | 2018 | 1280 × 1024    | 613            | 24      | 10,071.88          | 2.80      | 16.43                    |
| **MEAD** | V2 | Lab         | 2020 | 1920 × 1080    | 1,969          | 60      | 42,868.77          | 11.91     | 21.77                    |
| **VoxCeleb2**| V3| Wild        | 2018 | 360P~720P      | 1,326          | 1k+     | 21,528.20          | 5.98      | 16.24                    |
| **HDTF** | V4 | Wild        | 2021 | 720P~1080P     | 400            | 300+    | 55,452.08          | 15.40     | 138.63                   |
| **Celebv-HQ**| V5| Wild        | 2022 | 512 × 512      | 798            | 700+    | 13,486.20          | 3.75      | 16.90                    |

### Data Splits

The dataset follows a strict training and testing split protocol to ensure fair evaluation. The testing set is composed of a balanced selection from each sub-dataset.

| Dataset       | ID  | Total Size | Training Set | Test Set |
| ------------- | --- | ---------- | ------------ | -------- |
| **GRID**      | V0  | 6,600      | 6,570        | 30       |
| **RAVDESS**   | V1  | 613        | 583          | 30       |
| **MEAD**      | V2  | 1,969      | 1,939        | 30       |
| **VoxCeleb2** | V3  | 1,326      | 1,296        | 30       |
| **HDTF**      | V4  | 400        | 350          | 50       |
| **Celebv-HQ** | V5  | 798        | 768          | 30       |
| **Summary**   |     | **11,706** | **11,506**   | **200**  |

## Dataset Creation

### Curation Rationale

Raw videos from the wild (e.g., VoxCeleb2, Celebv-HQ) often contain background noise, diverse languages, or varying resolutions. This dataset is the result of the following data curation pipeline  designed to ensure high-quality audio-visual consistency:

1.  **Duration Filtering:** To facilitate temporal modeling, short clips from lab datasets are concatenated to form 10–20s sequences, while wild samples shorter than 10s are filtered out.
2.  **Signal-to-Noise Ratio (SNR) Filtering:** Clips with strong background noise, music, or environmental interference are removed based on SNR thresholds to ensure clean audio features.
3.  **Language Filtering:** Linguistic consistency is enforced by using **Whisper** to discard non-English samples or those with low detection confidence.
4.  **Audio-Visual Sync Filtering:** **SyncNet** is used to eliminate clips with poor lip synchronization, abrupt scene cuts, or off-screen speakers (e.g., voice-overs).
5.  **Resolution Normalization:** All videos are resized and center-cropped to a unified **512×512** resolution and re-encoded at **25 FPS** with standardized RGB to harmonize data from diverse sources.

### Source Video Data

- **GRID:** https://zenodo.org/records/3625687
- **RAVDESS:** https://zenodo.org/records/1188976
- **MEAD:** https://wywu.github.io/projects/MEAD/MEAD.html
- **VoxCeleb2:** https://www.robots.ox.ac.uk/~vgg/data/voxceleb/vox2.html
- **HDTF:** https://huggingface.co/datasets/global-optima-research/HDTF
- **Celebv-HQ:** https://github.com/CelebV-HQ/CelebV-HQ/

## Citation
```bibtex
@misc{wang20263dxtalkerunifyingidentitylip,
      title={3DXTalker: Unifying Identity, Lip Sync, Emotion, and Spatial Dynamics in Expressive 3D Talking Avatars}, 
      author={Zhongju Wang and Zhenhong Sun and Beier Wang and Yifu Wang and Daoyi Dong and Huadong Mo and Hongdong Li},
      year={2026},
      eprint={2602.10516},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2602.10516}, 
}
```