Image-to-3D
3d
hwjiang commited on
Commit
cdf2934
1 Parent(s): c91a29c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -3
README.md CHANGED
@@ -1,3 +1,28 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ ---
4
+ <b>Real3D</b>
5
+
6
+
7
+ **Model Details**:
8
+
9
+ **Model Description**:
10
+
11
+ We use the model architecture provided by [TripoSR](https://github.com/VAST-AI-Research/TripoSR), which is a Transformer model for 2D-to-3D mapping built on [LRM](https://arxiv.org/abs/2311.04400).
12
+
13
+ We scale it further on in-the-wild image collections by enabling unsupervised self-training and automatric data curation.
14
+
15
+ * Developed by: [Hanwen Jiang](https://hwjiang1510.github.io/)
16
+ * License: MIT
17
+ * Hardware: We train Real3D on 1 node (8GPU) with equivalent batch size of 80 for 5-6 days.
18
+
19
+ **Model Sources**:
20
+ * Paper: https://arxiv.org/abs/2406.08479
21
+ * Project: https://hwjiang1510.github.io/Real3D/
22
+ * Code for training and evaluation: https://github.com/hwjiang1510/Real3D
23
+
24
+ **Training Data**:
25
+ Real3D is jointly trained on synthetic data (Objaverse) and in-the-wild image collections. The former prevents training divergence, the latter introduces new knowldege from a broader distribution of real images. We use Objaverse renderings from [Zero-1-to-3](https://github.com/cvlab-columbia/zero123) and [GObjaverse](https://aigc3d.github.io/gobjaverse/). The in the wild images are from [ImageNet](https://www.image-net.org/), [OpenImages](https://storage.googleapis.com/openimages/web/index.html), etc.
26
+
27
+ **Misuse, Malicious Use, and Out-of-Scope Use**:
28
+ The model should not be used to intentionally create or disseminate 3D models that people would foreseeably find disturbing, distressing, or offensive; or content that propagates historical or current stereotypes.