| | --- |
| | license: mit |
| | task_categories: |
| | - robotics |
| | - video-classification |
| | tags: |
| | - minecraft |
| | - vla |
| | - vision-language-action |
| | - gaming |
| | - behavioral-cloning |
| | size_categories: |
| | - 1M<n<10M |
| | --- |
| | |
| | # Minecraft VLA Stage 1: Action Pretraining Data |
| |
|
| | Vision-Language-Action training data for Minecraft, processed from OpenAI's VPT contractor dataset. |
| |
|
| | ## Dataset Description |
| |
|
| | This dataset contains frame-action pairs from Minecraft gameplay, designed for training VLA models following the [Lumine](https://www.lumine-ai.org/) methodology. |
| |
|
| | ### Source |
| | - **Original**: [OpenAI VPT Contractor Data](https://github.com/openai/Video-Pre-Training) (7.x subset) |
| | - **Videos**: ~17,886 videos (~330 hours of early-game gameplay) |
| | - **Task**: "Play Minecraft" with focus on first 30 minutes of new worlds |
| |
|
| | ### Format |
| |
|
| | Each sample contains: |
| | | Field | Type | Description | |
| | |-------|------|-------------| |
| | | `image` | bytes | 640x360 JPEG frame | |
| | | `video_id` | string | Source video identifier | |
| | | `frame_idx` | int | Frame number at 5Hz | |
| | | `action` | string | Lumine-format action string | |
| |
|
| | ### Action Format |
| |
|
| | ``` |
| | <|action_start|> mouse_x mouse_y scroll ; K1 ; K2 ; K3 ; K4 <|action_end|> |
| | ``` |
| |
|
| | - `mouse_x`, `mouse_y`: Mouse delta (-1000 to 1000) |
| | - `scroll`: Hotbar scroll (always 0 - VPT uses number keys) |
| | - `K1` to `K4`: Key combinations per 50ms chunk |
| |
|
| | **Example:** |
| | ``` |
| | <|action_start|> 45 -12 0 ; W ; W Space ; W LMB ; W LMB <|action_end|> |
| | ``` |
| |
|
| | ### Processing Details |
| |
|
| | - **Frame rate**: 5 FPS (downsampled from VPT's 20 FPS) |
| | - **Action chunks**: 4 per frame (each 50ms = 200ms total) |
| | - **Filtering**: Idle frames removed, loading screens filtered |
| |
|
| | ## Usage |
| |
|
| | ```python |
| | from datasets import load_dataset |
| | |
| | # Streaming (recommended - no download required) |
| | ds = load_dataset("TESS-Computer/minecraft-vla-stage1", split="train", streaming=True) |
| | |
| | for sample in ds: |
| | image = sample["image"] # PIL Image or bytes |
| | action = sample["action"] |
| | # Process... |
| | ``` |
| |
|
| | ## Training Pipeline |
| |
|
| | This is Stage 1 of a 3-stage training pipeline: |
| | 1. **Stage 1** (this dataset): Action pretraining - learn observation→action mapping |
| | 2. **Stage 2**: Instruction following - add task instructions from JARVIS-VLA |
| | 3. **Stage 3**: Reasoning - add chain-of-thought before complex actions |
| |
|
| | ## Citation |
| |
|
| | If you use this dataset, please cite: |
| | - [OpenAI VPT](https://arxiv.org/abs/2206.11795) - Original contractor data |
| | - [JARVIS-VLA](https://craftjarvis.github.io/JarvisVLA/) - Instruction annotations |
| | - [Lumine](https://www.lumine-ai.org/) - Training methodology |
| |
|
| | ## License |
| |
|
| | MIT License. Original VPT data is released under MIT by OpenAI. |
| |
|