Datasets:

ArXiv:
License:
xiaoxiaoxh commited on
Commit
80d1daf
·
1 Parent(s): d1ca6fc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -0
README.md CHANGED
@@ -20,6 +20,7 @@ pretty_name: garment-tracking
20
  - **Point of Contact:**
21
 
22
  ## Dataset Summary
 
23
  This is the **VR-Folding** dataset created by the CVPR 2023 paper [GarmentTracking: Category-Level Garment Pose Tracking](https://garment-tracking.robotflow.ai).
24
  This dataset is recorded with a system called **VR-Garment**, which is a garment-hand interaction environment based on Unity.
25
 
@@ -64,5 +65,10 @@ Here is the detailed stucture of a data example of one frame:
64
  ├── rgb (30000, 3) uint8
65
  └── sizes (4,) int64
66
  ```
 
 
67
  # Dataset Example
68
  Please see [example](data/data_examples/README.md) for example data and visualization scripts.
 
 
 
 
20
  - **Point of Contact:**
21
 
22
  ## Dataset Summary
23
+ ![VR-Garment](assets/vr_garment.png)
24
  This is the **VR-Folding** dataset created by the CVPR 2023 paper [GarmentTracking: Category-Level Garment Pose Tracking](https://garment-tracking.robotflow.ai).
25
  This dataset is recorded with a system called **VR-Garment**, which is a garment-hand interaction environment based on Unity.
26
 
 
65
  ├── rgb (30000, 3) uint8
66
  └── sizes (4,) int64
67
  ```
68
+ Specifically, we render 4-view RGB-D images with Unity and generate concated point clouds for each frame. Here `grip_vertex_id` is the vertex index list of the grasped points of the mesh.
69
+
70
  # Dataset Example
71
  Please see [example](data/data_examples/README.md) for example data and visualization scripts.
72
+ Here are two video examples for flattening and folding task.
73
+ ![flattening](assets/flattening_example.png)
74
+ ![folding](assets/folding_example.png)