|
--- |
|
license: apache-2.0 |
|
pretty_name: 1X World Model Challenge Dataset |
|
size_categories: |
|
- 10M<n<100M |
|
viewer: false |
|
--- |
|
|
|
# 1X World Model Compression Challenge Dataset |
|
This repository hosts the dataset for the [1X World Model Compression Challenge](https://huggingface.co/spaces/1x-technologies/1X_World_Model_Challenge_Compression). |
|
|
|
```bash |
|
huggingface-cli download 1x-technologies/worldmodel --repo-type dataset --local-dir data |
|
``` |
|
|
|
## Updates Since v1.1 |
|
|
|
- **Train/Val v2.0 (~100 hours)**, replacing v1.1 |
|
- **Test v2.0 dataset for the Compression Challenge** |
|
- **Faces blurred** for privacy |
|
- **New raw video dataset** (CC-BY-NC-SA 4.0) at [worldmodel_raw_data](https://huggingface.co/datasets/1x-technologies/worldmodel_raw_data) |
|
- **Example scripts** now split into: |
|
- `cosmos_video_decoder.py` — for decoding Cosmos Tokenized bins |
|
- `unpack_data_test.py` — for reading the new test set |
|
- `unpack_data_train_val.py` — for reading the train/val sets |
|
|
|
--- |
|
|
|
## Train & Val v2.0 |
|
|
|
### Format |
|
|
|
Each split is sharded: |
|
- `video_{shard}.bin` — [NVIDIA Cosmos Tokenizer](https://github.com/NVIDIA/Cosmos-Tokenizer) discrete DV8×8×8 tokens at 30 Hz |
|
- `segment_idx_{shard}.bin` — segment boundaries |
|
- `states_{shard}.bin` — `np.float32` states (see below) |
|
- `metadata.json` / `metadata_{shard}.json` — overall vs. per‐shard metadata |
|
|
|
|
|
--- |
|
|
|
## Test v2.0 |
|
|
|
We provide a 450 sample **test_v2.0** dataset for the [World Model Compression Challenge](https://huggingface.co/spaces/1x-technologies/1X_World_Model_Challenge_Compression) with a similar structure (`video_{shard}.bin`, `states_{shard}.bin`). Use: |
|
- `unpack_data_test.py` to read the test set |
|
- `unpack_data_train_val.py` to read train/val |
|
--- |
|
|
|
### State Index Definition (New) |
|
``` |
|
0: HIP_YAW |
|
1: HIP_ROLL |
|
2: HIP_PITCH |
|
3: KNEE_PITCH |
|
4: ANKLE_ROLL |
|
5: ANKLE_PITCH |
|
6: LEFT_SHOULDER_PITCH |
|
7: LEFT_SHOULDER_ROLL |
|
8: LEFT_SHOULDER_YAW |
|
9: LEFT_ELBOW_PITCH |
|
10: LEFT_ELBOW_YAW |
|
11: LEFT_WRIST_PITCH |
|
12: LEFT_WRIST_ROLL |
|
13: RIGHT_SHOULDER_PITCH |
|
14: RIGHT_SHOULDER_ROLL |
|
15: RIGHT_SHOULDER_YAW |
|
16: RIGHT_ELBOW_PITCH |
|
17: RIGHT_ELBOW_YAW |
|
18: RIGHT_WRIST_PITCH |
|
19: RIGHT_WRIST_ROLL |
|
20: NECK_PITCH |
|
21: Left hand closure (0= open, 1= closed) |
|
22: Right hand closure (0= open, 1= closed) |
|
23: Linear Velocity |
|
24: Angular Velocity |
|
``` |
|
|
|
## Previous v1.1 |
|
|
|
- `video.bin` — 16×16 patches at 30Hz, quantized |
|
- `segment_ids.bin` — segment boundaries |
|
- `actions/` folder storing multiple `.bin`s for states, closures, etc. |
|
|
|
### v1.1 Joint Index |
|
``` |
|
{ |
|
0: HIP_YAW |
|
1: HIP_ROLL |
|
2: HIP_PITCH |
|
3: KNEE_PITCH |
|
4: ANKLE_ROLL |
|
5: ANKLE_PITCH |
|
6: LEFT_SHOULDER_PITCH |
|
7: LEFT_SHOULDER_ROLL |
|
8: LEFT_SHOULDER_YAW |
|
9: LEFT_ELBOW_PITCH |
|
10: LEFT_ELBOW_YAW |
|
11: LEFT_WRIST_PITCH |
|
12: LEFT_WRIST_ROLL |
|
13: RIGHT_SHOULDER_PITCH |
|
14: RIGHT_SHOULDER_ROLL |
|
15: RIGHT_SHOULDER_YAW |
|
16: RIGHT_ELBOW_PITCH |
|
17: RIGHT_ELBOW_YAW |
|
18: RIGHT_WRIST_PITCH |
|
19: RIGHT_WRIST_ROLL |
|
20: NECK_PITCH |
|
} |
|
|
|
A separate `val_v1.1` set is available. |
|
|
|
--- |
|
|
|
## Provided Checkpoints |
|
|
|
- `magvit2.ckpt` from [MAGVIT2](https://github.com/TencentARC/Open-MAGVIT2) used in v1.1 |
|
- For v2.0, see [NVIDIA Cosmos Tokenizer](https://github.com/NVIDIA/Cosmos-Tokenizer); we supply `cosmos_video_decoder.py`. |
|
|
|
--- |
|
|
|
## Directory Structure Example |
|
|
|
``` |
|
train_v1.1/ |
|
val_v1.1/ |
|
train_v2.0/ |
|
val_v2.0/ |
|
test_v2.0/ |
|
├── video_{shard}.bin |
|
├── states_{shard}.bin |
|
├── ... |
|
├── metadata_{shard}.json |
|
cosmos_video_decoder.py |
|
unpack_data_test.py |
|
unpack_data_train_val.py |
|
``` |
|
|
|
**License**: [Apache-2.0](./LICENSE) |
|
**Author**: 1X Technologies |
|
``` |