lzzzzy commited on
Commit
4f421f0
·
1 Parent(s): eae3b3b

fix:dataset viewer

Browse files
Files changed (3) hide show
  1. README.md +43 -37
  2. metadata.parquet +2 -2
  3. scripts/build_hf_metadata.py +1 -1
README.md CHANGED
@@ -1,18 +1,24 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
4
 
5
  # AVGen-Bench Generated Videos Data Card
6
 
7
  ## Overview
8
 
9
- This data card describes the generated audio-video outputs stored directly in the repository root by model directory.
10
 
11
  The collection is intended for **benchmarking and qualitative/quantitative evaluation** of text-to-audio-video (T2AV) systems. It is not a training dataset. Each item is a model-generated video produced from a prompt defined in `prompts/*.json`.
12
 
13
- Code repository: https://github.com/microsoft/AVGen-Bench
14
-
15
- For Hugging Face Hub compatibility, the repository includes a root-level `metadata.parquet` file so the Dataset Viewer can expose each video as a structured row with prompt metadata instead of treating the repo as an unindexed file dump.
 
16
 
17
  ## What This Dataset Contains
18
 
@@ -25,28 +31,28 @@ The dataset is organized by:
25
  A typical top-level structure is:
26
 
27
  ```text
28
- AVGen-Bench/
29
- ├── Kling_2.6/
30
- ├── LTX-2/
31
- ├── LTX-2.3/
32
- ├── MOVA_360p_Emu3.5/
33
- ├── MOVA_360p_NanoBanana_2/
34
- ├── Ovi_11/
35
- ├── Seedance_1.5_pro/
36
- ├── Sora_2/
37
- ├── Veo_3.1_fast/
38
- ├── Veo_3.1_quality/
39
- ├── Wan_2.2_HunyuanVideo-Foley/
40
- ├── Wan_2.6/
41
- ├── metadata.parquet
42
- ├── prompts/
43
- └── reference_image/ # optional, depending on generation pipeline
44
  ```
45
 
46
  Within each model directory, videos are grouped by category, for example:
47
 
48
  ```text
49
- Veo_3.1_fast/
50
  ├── ads/
51
  ├── animals/
52
  ├── asmr/
@@ -92,7 +98,7 @@ Each generated item is typically:
92
 
93
  - A single `.mp4` file
94
  - Containing model-generated video and, when supported by the model/pipeline, synthesized audio
95
- - Stored under `<model>/<category>/`
96
 
97
  The filename is usually derived from prompt content after sanitization. Exact naming may vary by generation script or provider wrapper.
98
  In the standard export pipeline, the filename is derived from the prompt's `content` field using the following logic:
@@ -111,18 +117,18 @@ def safe_filename(name: str, max_len: int = 180) -> str:
111
 
112
  So the expected output path pattern is:
113
 
114
- ```text
115
- <model>/<category>/<safe_filename(content)>.mp4
116
- ```
117
-
118
- For Dataset Viewer indexing, `metadata.parquet` stores one row per exported video with:
119
-
120
- - `file_name`: relative path to the `.mp4`
121
- - `model`: model directory name
122
- - `category`: benchmark category
123
- - `content`: prompt short name
124
- - `prompt`: full generation prompt
125
- - `prompt_id`: index inside `prompts/<category>.json`
126
 
127
  ## How The Data Was Produced
128
 
 
1
+ ---
2
+ license: mit
3
+ configs:
4
+ - config_name: default
5
+ data_files:
6
+ - split: train
7
+ path: metadata.parquet
8
+ ---
9
 
10
  # AVGen-Bench Generated Videos Data Card
11
 
12
  ## Overview
13
 
14
+ This data card describes the generated audio-video outputs stored directly in the repository root by model directory.
15
 
16
  The collection is intended for **benchmarking and qualitative/quantitative evaluation** of text-to-audio-video (T2AV) systems. It is not a training dataset. Each item is a model-generated video produced from a prompt defined in `prompts/*.json`.
17
 
18
+ Code repository: https://github.com/microsoft/AVGen-Bench
19
+
20
+ For Hugging Face Hub compatibility, the repository includes a root-level `metadata.parquet` file so the Dataset Viewer can expose each video as a structured row with prompt metadata instead of treating the repo as an unindexed file dump.
21
+ The relative video path is stored as a plain string column (`video_path`) rather than a media-typed `file_name` column, which avoids current Dataset Viewer post-processing failures on video rows.
22
 
23
  ## What This Dataset Contains
24
 
 
31
  A typical top-level structure is:
32
 
33
  ```text
34
+ AVGen-Bench/
35
+ ├── Kling_2.6/
36
+ ├── LTX-2/
37
+ ├── LTX-2.3/
38
+ ├── MOVA_360p_Emu3.5/
39
+ ├── MOVA_360p_NanoBanana_2/
40
+ ├── Ovi_11/
41
+ ├── Seedance_1.5_pro/
42
+ ├── Sora_2/
43
+ ├── Veo_3.1_fast/
44
+ ├── Veo_3.1_quality/
45
+ ├── Wan_2.2_HunyuanVideo-Foley/
46
+ ├── Wan_2.6/
47
+ ├── metadata.parquet
48
+ ├── prompts/
49
+ └── reference_image/ # optional, depending on generation pipeline
50
  ```
51
 
52
  Within each model directory, videos are grouped by category, for example:
53
 
54
  ```text
55
+ Veo_3.1_fast/
56
  ├── ads/
57
  ├── animals/
58
  ├── asmr/
 
98
 
99
  - A single `.mp4` file
100
  - Containing model-generated video and, when supported by the model/pipeline, synthesized audio
101
+ - Stored under `<model>/<category>/`
102
 
103
  The filename is usually derived from prompt content after sanitization. Exact naming may vary by generation script or provider wrapper.
104
  In the standard export pipeline, the filename is derived from the prompt's `content` field using the following logic:
 
117
 
118
  So the expected output path pattern is:
119
 
120
+ ```text
121
+ <model>/<category>/<safe_filename(content)>.mp4
122
+ ```
123
+
124
+ For Dataset Viewer indexing, `metadata.parquet` stores one row per exported video with:
125
+
126
+ - `video_path`: relative path to the `.mp4` stored as a plain string
127
+ - `model`: model directory name
128
+ - `category`: benchmark category
129
+ - `content`: prompt short name
130
+ - `prompt`: full generation prompt
131
+ - `prompt_id`: index inside `prompts/<category>.json`
132
 
133
  ## How The Data Was Produced
134
 
metadata.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:236f73fb879f176c4999a60838f53d1dae7a8904b070344d60802f58ecb449c9
3
- size 112572
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e7d3f7dafd4392ae31e2ec656c4ccfb435a1f6ef234c1252e1db30fde3b57a2
3
+ size 112574
scripts/build_hf_metadata.py CHANGED
@@ -69,7 +69,7 @@ def build_rows(repo_root: Path) -> list[dict[str, object]]:
69
 
70
  rows.append(
71
  {
72
- "file_name": rel_path,
73
  "model": model,
74
  "category": category,
75
  "content": prompt_record["content"] if prompt_record else filename,
 
69
 
70
  rows.append(
71
  {
72
+ "video_path": rel_path,
73
  "model": model,
74
  "category": category,
75
  "content": prompt_record["content"] if prompt_record else filename,