nielsr HF Staff commited on
Commit
d09e1ba
·
verified ·
1 Parent(s): 03bb359

Add paper link, project page, code, and task category

Browse files

Hi, I'm Niels from the community science team at Hugging Face. This PR improves the dataset card for the AniGen Sample Data by:
- Adding the `image-to-3d` task category to the metadata.
- Linking the dataset to its associated research paper ([AniGen: Unified S3 Fields for Animatable 3D Asset Generation](https://huggingface.co/papers/2604.08746)).
- Providing links to the official project page and GitHub repository.
- Adding training commands and a BibTeX citation for better usability.

Files changed (1) hide show
  1. README.md +51 -11
README.md CHANGED
@@ -1,22 +1,28 @@
1
  ---
2
  license: mit
3
- pretty_name: AniGen Sample Data
4
  size_categories:
5
- - n<1K
 
6
  tags:
7
- - 3d
8
- - image
 
 
9
  configs:
10
- - config_name: default
11
- default: true
12
- data_files:
13
- - split: train
14
- path: samples.csv
15
  ---
16
 
17
  # AniGen Sample Data
18
 
19
- This directory is a compact example subset of the AniGen training dataset.
 
 
 
 
20
 
21
  ## What Is Included
22
 
@@ -54,4 +60,38 @@ For a row with sample key `<file_identifier>`:
54
  - voxel files: `voxels/<file_identifier>.ply` and `voxels/<file_identifier>_skeleton.ply`
55
  - image feature: `features/dinov2_vitl14_reg/<file_identifier>.npz`
56
  - mesh latents: files under `latents/*/<file_identifier>.npz`
57
- - structure latents: files under `ss_latents/*/<file_identifier>.npz`
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
 
3
  size_categories:
4
+ - n<1K
5
+ pretty_name: AniGen Sample Data
6
  tags:
7
+ - 3d
8
+ - image
9
+ task_categories:
10
+ - image-to-3d
11
  configs:
12
+ - config_name: default
13
+ default: true
14
+ data_files:
15
+ - split: train
16
+ path: samples.csv
17
  ---
18
 
19
  # AniGen Sample Data
20
 
21
+ [Paper](https://huggingface.co/papers/2604.08746) | [Project Page](https://yihua7.github.io/AniGen_web/) | [GitHub](https://github.com/VAST-AI-Research/AniGen)
22
+
23
+ This directory is a compact example subset of the AniGen training dataset, as presented in the paper [AniGen: Unified $S^3$ Fields for Animatable 3D Asset Generation](https://huggingface.co/papers/2604.08746).
24
+
25
+ AniGen is a unified framework that directly generates animate-ready 3D assets conditioned on a single image by representing shape, skeleton, and skinning as mutually consistent $S^3$ Fields.
26
 
27
  ## What Is Included
28
 
 
60
  - voxel files: `voxels/<file_identifier>.ply` and `voxels/<file_identifier>_skeleton.ply`
61
  - image feature: `features/dinov2_vitl14_reg/<file_identifier>.npz`
62
  - mesh latents: files under `latents/*/<file_identifier>.npz`
63
+ - structure latents: files under `ss_latents/*/<file_identifier>.npz`
64
+
65
+ ## Sample Usage (Training)
66
+
67
+ According to the [official repository](https://github.com/VAST-AI-Research/AniGen), you can use this data for training by following these stages:
68
+
69
+ ```bash
70
+ # Stage 1: Skin AutoEncoder
71
+ python train.py --config configs/anigen_skin_ae.json --output_dir outputs/anigen_skin_ae
72
+
73
+ # Stage 2: Sparse Structure DAE
74
+ python train.py --config configs/ss_dae.json --output_dir outputs/ss_dae
75
+
76
+ # Stage 3: Structured Latent DAE
77
+ python train.py --config configs/slat_dae.json --output_dir outputs/slat_dae
78
+
79
+ # Stage 4: SS Flow Matching (image-conditioned generation)
80
+ python train.py --config configs/ss_flow_duet.json --output_dir outputs/ss_flow_duet
81
+
82
+ # Stage 5: SLAT Flow Matching (image-conditioned generation)
83
+ python train.py --config configs/slat_flow_auto.json --output_dir outputs/slat_flow_auto
84
+ ```
85
+
86
+ ## Citation
87
+
88
+ ```bibtex
89
+ @article{huang2026anigen,
90
+ title = {AniGen: Unified $S^3$ Fields for Animatable 3D Asset Generation},
91
+ author = {Huang, Yi-Hua and Zhou, Zi-Xin and He, Yuting and Chang, Chirui
92
+ and Pu, Cheng-Feng and Yang, Ziyi and Guo, Yuan-Chen
93
+ and Cao, Yan-Pei and Qi, Xiaojuan},
94
+ journal = {ACM SIGGRAPH},
95
+ year = {2026}
96
+ }
97
+ ```