stonet2000 commited on
Commit
44f4016
1 Parent(s): c568082

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +82 -0
README.md CHANGED
@@ -1,3 +1,85 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ language:
4
+ - en
5
+ tags:
6
+ - robotics
7
+ - reinforcement learning
8
+ - embodied ai
9
+ - computer vision
10
+ - simulation
11
+ size_categories:
12
+ - 1M<n<10M
13
  ---
14
+
15
+ # ManiSkill2 Data
16
+
17
+ ![teaser](figures/teaser_v2.jpg)
18
+
19
+ [![PyPI version](https://badge.fury.io/py/mani-skill2.svg)](https://badge.fury.io/py/mani-skill2) [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/haosulab/ManiSkill2/blob/main/examples/tutorials/1_quickstart.ipynb)
20
+ [![Docs status](https://img.shields.io/badge/docs-passing-brightgreen.svg)](https://haosulab.github.io/ManiSkill2)
21
+ [![Discord](https://img.shields.io/discord/996566046414753822?logo=discord)](https://discord.gg/x8yUZe5AdN)
22
+ <!-- [![Docs](https://github.com/haosulab/ManiSkill2/actions/workflows/gh-pages.yml/badge.svg)](https://haosulab.github.io/ManiSkill2) -->
23
+
24
+ ManiSkill2 is a unified benchmark for learning generalizable robotic manipulation skills powered by [SAPIEN](https://sapien.ucsd.edu/). **It features 20 out-of-box task families with 2000+ diverse object models and 4M+ demonstration frames**. Moreover, it empowers fast visual input learning algorithms so that **a CNN-based policy can collect samples at about 2000 FPS with 1 GPU and 16 processes on a workstation**. The benchmark can be used to study a wide range of algorithms: 2D & 3D vision-based reinforcement learning, imitation learning, sense-plan-act, etc.
25
+ This is the huggingface datasets page for all data related to [ManiSkill2](https://github.com/haosulab/ManiSkill2),
26
+ including **assets, robot demonstrations, and pretrained models**
27
+
28
+ For detailed information about ManiSkill2, head over to our [GitHub repository](https://github.com/haosulab/ManiSkill2), [website](https://maniskill2.github.io/), or [ICLR 2023 paper](https://arxiv.org/abs/2302.04659)
29
+ [documentation](https://haosulab.github.io/ManiSkill2/index.html)
30
+
31
+ ## Assets
32
+
33
+ Some environments require you to download additional assets, which are stored here.
34
+
35
+ You can download all the assets by running
36
+
37
+ ```
38
+ python -m mani_skill2.utils.download_asset all
39
+ ```
40
+
41
+ or download task-specific assets by running
42
+
43
+ ```
44
+ python -m mani_skill2.utils.download_asset ${ENV_ID}
45
+ ```
46
+
47
+ ## Demonstration Data
48
+
49
+ The robot demonstrations consist of 4 million+ frames across 20+ robot manipulation tasks.
50
+
51
+ We provide a command line tool (mani_skill2.utils.download_demo) to download demonstrations from here.
52
+
53
+ ```
54
+ # Download all the demonstration datasets
55
+ python -m mani_skill2.utils.download_demo all
56
+ # Download the demonstration dataset for a specific task
57
+ python -m mani_skill2.utils.download_demo ${ENV_ID}
58
+ # Download the demonstration datasets for all rigid-body tasks to "./demos"
59
+ python -m mani_skill2.utils.download_demo rigid_body -o ./demos
60
+ # Download the demonstration datasets for all soft-body tasks
61
+ python -m mani_skill2.utils.download_demo soft_body
62
+ ```
63
+
64
+ To learn how to use the demonstrations and what environments are available, go to the demonstrations documentation page: https://haosulab.github.io/ManiSkill2/concepts/demonstrations.html
65
+
66
+
67
+ ## License
68
+
69
+ All rigid body environments in ManiSkill are licensed under fully permissive licenses (e.g., Apache-2.0).
70
+
71
+ However, the soft body environments will follow Warp's license. Currently, they are licensed under
72
+ [NVIDIA Source Code License for Warp](https://github.com/NVIDIA/warp/blob/main/LICENSE.md).
73
+
74
+ The assets are licensed under [CC BY-NC 4.0](https://creativecommons.org/licenses/by-nc/4.0/legalcode).
75
+
76
+ ## Citation
77
+
78
+ If you use ManiSkill2 or its assets, models, and demonstrations, please cite using the following BibTeX entry:
79
+
80
+ @inproceedings{gu2023maniskill2,
81
+ title={ManiSkill2: A Unified Benchmark for Generalizable Manipulation Skills},
82
+ author={Gu, Jiayuan and Xiang, Fanbo and Li, Xuanlin and Ling, Zhan and Liu, Xiqiaing and Mu, Tongzhou and Tang, Yihe and Tao, Stone and Wei, Xinyue and Yao, Yunchao and Yuan, Xiaodi and Xie, Pengwei and Huang, Zhiao and Chen, Rui and Su, Hao},
83
+ booktitle={International Conference on Learning Representations},
84
+ year={2023}
85
+ }