File size: 9,282 Bytes
58ff142 748df47 0dccf3b 748df47 6177c8d 4b75678 6177c8d 748df47 ec40962 748df47 16729f9 748df47 ec40962 2db117f ec40962 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 | ---
license: apache-2.0
language:
- en
size_categories:
- 100M<n<1B
---
# Dataset Card for Dataset Curation of 3DXTalker
### Dataset Description
- **Repository:** [https://github.com/EngineeringAI-LAB/3DXTalker/tree/main]
- **Paper:** [https://arxiv.org/abs/2602.10516]
- **Project :** [https://engineeringai-lab.github.io/3DXTalker.github.io/]
### Dataset Summary
This dataset is a large-scale, curated collection of talking head videos built for tasks such as high-fidelity 3D talking avatar generation, lip synchronization, and pose dynamics modeling.
The dataset aggregates and standardizes data from six prominent sources (**GRID, RAVDESS, MEAD, VoxCeleb2, HDTF, Celebv-HQ**), processed through a rigorous data curation pipeline to ensure high quality in terms of face alignment, resolution, and audio-visual synchronization. It covers diverse environments (Lab vs. Wild) and a wide range of subjects.
### Supported Tasks and Leaderboards
- **3D Talking Head Generation:** Synthesizing realistic talking videos from driving speech.
- **Audio-Driven Lip Synchronization:** Aligning lip movements precisely with input speech.
- **Emotion Analysis & Synthesis:** Leveraging the emotional diversity in datasets like RAVDESS and MEAD.
- **Audio-Driven Head Pose Synthesis:** Modeling natural head movements and orientation directly driving speech.
## Dataset Structure
```
trainset/
├── V0-GRID/ # 6,570 sequences from GRID corpus
│ ├── V0-s1-00001/
│ │ ├── audio.wav # (N,) audio data
│ │ ├── cam.npy # (T, 3) camera parameters
│ │ ├── detailcode.npy # (T, 128) facial details
│ │ ├── envelope.npy # (N,) audio envelope
│ │ ├── expcode.npy # (T, 50) expression codes
│ │ ├── lightcode.npy # (T, 9, 3) lighting
│ │ ├── metadata.pkl # Sequence metadata
│ │ ├── posecode.npy # (T, 6) head pose
│ │ ├── refimg.npy # (C, H, W) reference image
│ │ ├── shapecode.npy # (T, 100) shape codes
│ │ └── texcode.npy # (T, 50) texture codes
│ ├── V0-s1-00002/
│ │ └── ... (same 11 files)
│ ├── V0-s1-00003/
│ └── ... (6,570 total sequences)
│
├── V1-RAVDESS/ # 583 sequences from RAVDESS dataset
│ ├── V1-Song-Actor_01-00001/
│ │ └── ... (same 11 files)
│ ├── V1-Song-Actor_01-00002/
│ ├── V1-Speech-Actor_01-00001/
│ ├── V1-Speech-Actor_02-00001/
│ └── ... (583 total sequences)
│
├── V2-MEAD/ # 1,939 sequences from MEAD dataset
│ ├── V2-M003-angry-00001/
│ │ └── ... (same 11 files)
│ ├── V2-M003-angry-00002/
│ ├── V2-M003-happy-00001/
│ ├── V2-W009-sad-00001/
│ └── ... (1,939 total sequences)
│
├── V3-VoxCeleb2/ # 1,296 sequences from VoxCeleb2
│ ├── {sequence_id}/
│ │ └── ... (same 11 files)
│ └── ... (1,296 total sequences)
│
├── V4-HDTF/ # 350 sequences from HDTF dataset
│ ├── {sequence_id}/
│ │ └── ... (same 11 files)
│ └── ... (350 total sequences)
│
└── V5-CelebV-HQ/ # 768 sequences from CelebV-HQ dataset
├── {sequence_id}/
│ └── ... (same 11 files)
└── ... (768 total sequences)
```
## Data Format Details
### File Overview
| File | Type | Shape | Description |
|------|------|-------|-------------|
| `audio.wav` | Audio | (N_samples,) | Original audio waveform|
| `cam.npy` | Parameters | (N_frames, 3) | Camera parameters (position/scale) |
| `detailcode.npy` | Parameters | (N_frames, 128) | Facial detail codes (wrinkles, fine features) |
| `envelope.npy` | Parameters | (N_audio_samples,) | Audio envelope/amplitude over time |
| `expcode.npy` | Parameters | (N_frames, 50) | FLAME expression parameters (50-dim) |
| `lightcode.npy` | Parameters | (N_frames, 9, 3) | Spherical harmonics lighting (9 bands × RGB) |
| `metadata.pkl` | Metadata | N/A | Sequence metadata (integer or dict) |
| `posecode.npy` | Parameters | (N_frames, 6) | 3 head pose + 3 jaw pose |
| `refimg.npy` | Image | (3, 224, 224) | Reference image (RGB, 224×224 pixels) |
| `shapecode.npy` | Parameters | (N_frames, 100) | FLAME shape parameters (100-dim) |
| `texcode.npy` | Parameters | (N_frames, 50) | Texture codes (50-dim) |
### Coordinate Systems and Conventions
- **FLAME model**: 3D Morphable Face Model with 5023 vertices
- **Expression space**: 50-dimensional linear basis
- **Shape space**: 100-dimensional PCA space
- **Pose representation**: 3 head pose + 3 jaw pose
- **Lighting**: 2nd-order spherical harmonics (9 bands)
### Temporal Synchronization
- **Video frames**: 25 FPS (frames per second)
- **Audio samples**: 16,000 samples per second
- All video parameters (`expcode`, `shapecode`, `detailcode`, `posecode`, `cam`, `lightcode`, `texcode`) share the same `N_frames` dimension
- Audio and video are temporally aligned (frame 0 corresponds to start of audio)
### Data Statistics
The dataset comprises **11,706** total video samples, spanning approximately **67.4 hours** of self-talking footage. The data is categorized by environment (Lab vs. Wild) and includes varying resolutions and subject diversity.
#### Detailed Statistics (from Curation Pipeline)
| Dataset | ID | Environment | Year | Raw Resolution | Size (samples) | Subject | Total Duration (s) | Hours (h) | Avg. Duration (s/sample) |
|-------------|----|-------------|------|----------------|----------------|---------|--------------------|-----------|--------------------------|
| **GRID** | V0 | Lab | 2006 | 720 × 576 | 6,600 | 34 | 99,257.81 | 27.57 | 15.04 |
| **RAVDESS** | V1 | Lab | 2018 | 1280 × 1024 | 613 | 24 | 10,071.88 | 2.80 | 16.43 |
| **MEAD** | V2 | Lab | 2020 | 1920 × 1080 | 1,969 | 60 | 42,868.77 | 11.91 | 21.77 |
| **VoxCeleb2**| V3| Wild | 2018 | 360P~720P | 1,326 | 1k+ | 21,528.20 | 5.98 | 16.24 |
| **HDTF** | V4 | Wild | 2021 | 720P~1080P | 400 | 300+ | 55,452.08 | 15.40 | 138.63 |
| **Celebv-HQ**| V5| Wild | 2022 | 512 × 512 | 798 | 700+ | 13,486.20 | 3.75 | 16.90 |
### Data Splits
The dataset follows a strict training and testing split protocol to ensure fair evaluation. The testing set is composed of a balanced selection from each sub-dataset.
| Dataset | ID | Total Size | Training Set | Test Set |
| ------------- | --- | ---------- | ------------ | -------- |
| **GRID** | V0 | 6,600 | 6,570 | 30 |
| **RAVDESS** | V1 | 613 | 583 | 30 |
| **MEAD** | V2 | 1,969 | 1,939 | 30 |
| **VoxCeleb2** | V3 | 1,326 | 1,296 | 30 |
| **HDTF** | V4 | 400 | 350 | 50 |
| **Celebv-HQ** | V5 | 798 | 768 | 30 |
| **Summary** | | **11,706** | **11,506** | **200** |
## Dataset Creation
### Curation Rationale
Raw videos from the wild (e.g., VoxCeleb2, Celebv-HQ) often contain background noise, diverse languages, or varying resolutions. This dataset is the result of the following data curation pipeline designed to ensure high-quality audio-visual consistency:
1. **Duration Filtering:** To facilitate temporal modeling, short clips from lab datasets are concatenated to form 10–20s sequences, while wild samples shorter than 10s are filtered out.
2. **Signal-to-Noise Ratio (SNR) Filtering:** Clips with strong background noise, music, or environmental interference are removed based on SNR thresholds to ensure clean audio features.
3. **Language Filtering:** Linguistic consistency is enforced by using **Whisper** to discard non-English samples or those with low detection confidence.
4. **Audio-Visual Sync Filtering:** **SyncNet** is used to eliminate clips with poor lip synchronization, abrupt scene cuts, or off-screen speakers (e.g., voice-overs).
5. **Resolution Normalization:** All videos are resized and center-cropped to a unified **512×512** resolution and re-encoded at **25 FPS** with standardized RGB to harmonize data from diverse sources.
### Source Video Data
- **GRID:** https://zenodo.org/records/3625687
- **RAVDESS:** https://zenodo.org/records/1188976
- **MEAD:** https://wywu.github.io/projects/MEAD/MEAD.html
- **VoxCeleb2:** https://www.robots.ox.ac.uk/~vgg/data/voxceleb/vox2.html
- **HDTF:** https://huggingface.co/datasets/global-optima-research/HDTF
- **Celebv-HQ:** https://github.com/CelebV-HQ/CelebV-HQ/
## Citation
```bibtex
@misc{wang20263dxtalkerunifyingidentitylip,
title={3DXTalker: Unifying Identity, Lip Sync, Emotion, and Spatial Dynamics in Expressive 3D Talking Avatars},
author={Zhongju Wang and Zhenhong Sun and Beier Wang and Yifu Wang and Daoyi Dong and Huadong Mo and Hongdong Li},
year={2026},
eprint={2602.10516},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2602.10516},
}
```
|