basilevh commited on
Commit
65beccc
1 Parent(s): 857c308

Update instructions.md

Browse files
Files changed (1) hide show
  1. instructions.md +10 -8
instructions.md CHANGED
@@ -1,11 +1,13 @@
1
- ## Tips if not getting expected results :detective:
2
- 1. If you are trying out images of humans, especially faces, note that it's unfortunately not the intended use cases. We would encourage to try out images of everyday objects or even artworks.
3
- 2. If some part of the object is missing, check the interactive angle visualization pane (top right) where you can find a panel of the actual input image to the model after preprocessing steps and see if the segmented image contains the entire object you are trying to visualize.
4
- 3. The model is probabilistic, therefore, if the number of samples is selected to be bigger than 1 and results look different, that's expected as the model tries to predict a diverse set of possibilities given the input image and the specified camera viewpoint.
5
- 4. Under "advanced options", you can tune two parameters as you can typically find in other stable diffusion demos as well:
6
- - Diffusion Guidance Scale defines how much you want the model to respect the input information (image + angles). Higher scale typically leads to less diversity and higher image distortion.
7
- - Number of diffusion inference steps controls the number of diffusion steps is applied to generate each image. Usually the higher the better with a diminishing return.
8
 
9
- Have fun!
 
 
 
 
 
 
 
 
10
 
11
  A model card can be found here: [uses.md](https://github.com/cvlab-columbia/zero123/blob/main/zero123/uses.md)
1
+ ## Tips if you are not getting expected results:
 
 
 
 
 
 
2
 
3
+ 1. If you are getting an error, try an image with a smaller resolution.
4
+ 2. If you are trying out images of humans, especially faces, note that it is unfortunately not the intended use case. We would encourage trying out images of everyday objects instead, or even artworks.
5
+ 3. If some part of the object is missing, check the 3D interactive camera angle visualization pane (top right) where you can find the input image actually used for the model after the automatic preprocessing steps, and verify whether the segmented image contains the entire object you are trying to rotate.
6
+ 4. The model is probabilistic, therefore if the number of samples is selected to be bigger than 1 and results look different, that is expected behavior as the model tries to predict a diverse set of possibilities given the input image and the specified camera viewpoint.
7
+ 5. Under "advanced options", you can tune two parameters as you can typically find in other diffusion model demos as well:
8
+ - Diffusion Guidance Scale defines how much you want the model to respect the input information (image + angles). Higher scale typically leads to less diversity and higher image distortion.
9
+ - Number of diffusion inference steps controls the number of denoising iterations that are applied in order to generate each image. Usually the higher the better (with potentially diminishing returns).
10
+
11
+ Have fun! :smiley:
12
 
13
  A model card can be found here: [uses.md](https://github.com/cvlab-columbia/zero123/blob/main/zero123/uses.md)