Mark7121983123 commited on
Commit
3235399
·
verified ·
1 Parent(s): 75d9124

Add dataset card

Browse files
Files changed (1) hide show
  1. README.md +130 -0
README.md ADDED
@@ -0,0 +1,130 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-4.0
3
+ task_categories:
4
+ - video-text-to-text
5
+ - video-classification
6
+ language:
7
+ - en
8
+ size_categories:
9
+ - n<1K
10
+ pretty_name: VBVR-MultiStep-Bench
11
+ tags:
12
+ - video-reasoning
13
+ - multi-step
14
+ - long-horizon
15
+ - image-to-video
16
+ - evaluation
17
+ - benchmark
18
+ configs:
19
+ - config_name: default
20
+ data_files:
21
+ - split: test
22
+ path: "Multi-*/Multi-*_task/Multi-*_*/question_metadata.json"
23
+ ---
24
+
25
+ # VBVR-MultiStep-Bench
26
+
27
+ The frozen **180-instance public evaluation split** released alongside the [VBVR-MultiStep](https://huggingface.co/datasets/Video-Reason/VBVR-MultiStep) training corpus. Designed for long-horizon multi-step image-to-video (I2V) reasoning evaluation.
28
+
29
+ This dataset is part of the **VBVR (Very Big Video Reasoning Suite)** project. See the parent suite at <https://video-reason.com> and the suite paper [VBVR: A Very Big Video Reasoning Suite (Wang et al., ICML 2026)](https://icml.cc/virtual/2026/poster/65709).
30
+
31
+ ## At a glance
32
+
33
+ | Property | Value |
34
+ |---|---|
35
+ | Tasks | **36** parameterized tasks (`Multi-01` … `Multi-36`) |
36
+ | Reasoning families | Navigation, Planning, CSP, Execution, Geometry, Physics |
37
+ | Instances | **180** (5 per task × 36) |
38
+ | Per-instance artifacts | 5 (see below) |
39
+ | License | CC-BY-4.0 |
40
+
41
+ ## Five-artifact data contract
42
+
43
+ Every instance lives at:
44
+
45
+ ```
46
+ Multi-XX_<name>_data-generator/Multi-XX_<name>_data-generator_task/Multi-XX_<name>_data-generator_<id>/
47
+ ```
48
+
49
+ and contains exactly:
50
+
51
+ | File | Role |
52
+ |---|---|
53
+ | `first_frame.png` | Model conditioning image (the only visual input the model receives at inference) |
54
+ | `prompt.txt` | Natural-language task contract |
55
+ | `final_frame.png` | Target endpoint (held out from the model) |
56
+ | `ground_truth.mp4` | Reference rollout demonstrating the correct trajectory |
57
+ | `question_metadata.json` | Seed, version, tolerances, task-specific fields |
58
+
59
+ A top-level `metadata.parquet` indexes every instance with the task id, family, seed, and per-instance metadata for fast filtering.
60
+
61
+ ## Reasoning families
62
+
63
+ | Family | Characteristic | Released tasks |
64
+ |---|---|---|
65
+ | Navigation | Discrete motion under adjacency / obstacle constraints | 6 |
66
+ | Planning | Operator-based state transformation | 6 |
67
+ | CSP | Incremental labeling under global consistency | 6 (3 used for human judging) |
68
+ | Execution | Clocked deterministic update rules | 6 |
69
+ | Geometry | Ordered constructive geometry | 6 |
70
+ | Physics | Continuous dynamics with contact / conservation | 6 |
71
+
72
+ Tasks `Multi-13`, `Multi-14`, `Multi-15` (CSP) are excluded from the human-judging pool described in the paper but are included in this release for completeness.
73
+
74
+ ## Intended use
75
+
76
+ - **Primary use**: trajectory-level evaluation of I2V systems under a fixed five-artifact contract.
77
+ - **Comparison protocol**: blind human pairwise judging on three independent axes — process correctness, reference fidelity, render quality.
78
+ - **Companion training corpus**: [Video-Reason/VBVR-MultiStep](https://huggingface.co/datasets/Video-Reason/VBVR-MultiStep) (~360k samples).
79
+
80
+ ## Loading
81
+
82
+ ```python
83
+ import pandas as pd
84
+ meta = pd.read_parquet("hf://datasets/Video-Reason/VBVR-MultiStep-Bench/metadata.parquet")
85
+ ```
86
+
87
+ Or pull a single instance:
88
+
89
+ ```python
90
+ from huggingface_hub import hf_hub_download
91
+ prompt_path = hf_hub_download(
92
+ "Video-Reason/VBVR-MultiStep-Bench",
93
+ "Multi-01_maze_shortest_path_data-generator/Multi-01_maze_shortest_path_data-generator_task/Multi-01_maze_shortest_path_data-generator_00000000/prompt.txt",
94
+ repo_type="dataset",
95
+ )
96
+ ```
97
+
98
+ ## License
99
+
100
+ Released under **CC-BY-4.0**. The reference rollouts are produced from generators that consume only released task definitions; no third-party copyrighted content is embedded.
101
+
102
+ Wan2.2-I2V-A14B (Apache-2.0) is referenced as a baseline model and a fine-tuning ancestor for `VBVR-Wan2.2`; this dataset does not redistribute Wan2.2 weights.
103
+
104
+ ## Citation
105
+
106
+ ```bibtex
107
+ @inproceedings{vbvr_multistep_2026,
108
+ title={Could Video Generation Models Solve Long-Horizon Multi-Step Reasoning Tasks?},
109
+ author={Anonymous},
110
+ booktitle={NeurIPS Datasets and Benchmarks},
111
+ year={2026},
112
+ note={Under review.}
113
+ }
114
+ ```
115
+
116
+ The parent VBVR suite:
117
+
118
+ ```bibtex
119
+ @inproceedings{wang2026vbvr,
120
+ title={A Very Big Video Reasoning Suite},
121
+ author={Wang, Maijunxian and others},
122
+ booktitle={ICML},
123
+ year={2026},
124
+ url={https://icml.cc/virtual/2026/poster/65709}
125
+ }
126
+ ```
127
+
128
+ ## Responsible AI
129
+
130
+ This dataset is fully synthetic — generators produce every instance from controlled parameters. There are no human subjects, no scraped media, and no personal information. See the [Croissant file](./croissant.json) for the complete RAI metadata.