DoomFrameDataset / README.md
sanatsingh's picture
Improve dataset card and add preview config
e65d35c verified
---
pretty_name: Doom Frame Dataset
tags:
- doom
- vizdoom
- reinforcement-learning
- imitation-learning
- webdataset
configs:
- config_name: preview
data_files:
- split: train
path: data/train-000000.tar
- config_name: full
data_files:
- split: train
path: data/train-*.tar
---
# DoomFrameDataset
DoomFrameDataset is a ViZDoom frame-action dataset generated from policy rollouts. It is packaged as WebDataset tar shards for streaming training, imitation learning, behavior cloning, and offline reinforcement-learning experiments.
The dataset contains RGB game frames paired with the action selected by the rollout policy and per-step metadata such as reward, episode id, step id, terminal flag, and value estimate.
## Dataset Size
| Config | Files | Samples | Intended use |
| --- | ---: | ---: | --- |
| `preview` | 1 shard | ~79k | Hugging Face preview and quick sanity checks |
| `full` | 31 shards | 2,398,745 | Training and full streaming reads |
The packaged dataset is about 68 GB.
## Files
```text
data/
train-000000.tar
train-000001.tar
...
train-000030.tar
action_map.json
README.md
```
Each tar shard contains paired files with the same numeric key:
```text
000000000000.png
000000000000.json
000000000001.png
000000000001.json
...
```
The PNG is the game frame. The JSON is the metadata for that frame.
## Sample Metadata
```json
{
"action_id": 1,
"action_name": "TURN_RIGHT",
"action_vector": [0.0, 0.0, 0.0, 0.0, 1.0, 0.0],
"curriculum_level": 1,
"done": false,
"episode": 1,
"frame_path": "frames/episode_001/step_000000.png",
"global_step": 0,
"reward": 0.0,
"source_frame_path": "frames/episode_001/step_000000.png",
"step": 0,
"value": 1.7968196868896484,
"webdataset_key": "000000000000"
}
```
See `action_map.json` for the full action id, action name, and action vector mapping.
## Load The Preview Config
Use `preview` when you only want to verify the dataset or inspect examples in the Hugging Face Dataset Viewer.
```python
from datasets import load_dataset
ds = load_dataset(
"brahmandam/DoomFrameDataset",
"preview",
split="train",
streaming=True,
)
sample = next(iter(ds))
print(sample.keys())
print(sample["json"])
image = sample["png"]
```
## Stream The Full Dataset
Use `full` for training.
```python
from datasets import load_dataset
ds = load_dataset(
"brahmandam/DoomFrameDataset",
"full",
split="train",
streaming=True,
)
for sample in ds:
image = sample["png"]
metadata = sample["json"]
action_id = metadata["action_id"]
break
```
You can also read the shards directly with WebDataset:
```python
import webdataset as wds
urls = "https://huggingface.co/datasets/brahmandam/DoomFrameDataset/resolve/main/data/train-{000000..000030}.tar"
dataset = (
wds.WebDataset(urls)
.decode("pil")
.to_tuple("png", "json")
)
image, metadata = next(iter(dataset))
```
## Notes
The `preview` config intentionally points to a single shard so the Hub can inspect a small part of the dataset without processing the full 68 GB. For training, use the `full` config.
This dataset was generated from automated ViZDoom policy rollouts. It should be treated as gameplay observation/action data, not human demonstrations.