ZeroCool94 commited on
Commit
882b6cd
1 Parent(s): 70faf98

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -6
README.md CHANGED
@@ -32,13 +32,13 @@ metrics:
32
 
33
  # About the model
34
  -----------------
35
- This model is a fine-tune of Stable Diffusion v1.5, trained on the [Imaginary Network Expanded Dataset](https://github.com/Sygil-Dev/INE-dataset), with the big advantage of allowing the use of multiple namespaces (labeled tags) to control various parts of the final generation.
36
  While current models usually are prone to “context errors” and need substantial negative prompting to set them on the right track, the use of namespaces in this model (eg. “species:seal” or “studio:dc”) stop the model from misinterpreting a seal as the singer Seal, or DC Comics as Washington DC.
37
  This model is also able to understand other languages besides English, currently it can partially understand prompts in Chinese, Japanese and Spanish. More training is already being done in order to have the model completely understand those languages and have it work just like how it works with English prompts.
38
 
39
  As the model is fine-tuned on a wide variety of content, it’s able to generate many types of images and compositions, and easily outperforms the original model when it comes to portraits, architecture, reflections, fantasy, concept art, anime, landscapes and a lot more without being hyper-specialized like other community fine-tunes that are currently available.
40
 
41
- **Note: The prompt engineering techniques needed are slightly different from other fine-tunes and the original SD 1.5, so while you can still use your favorite prompts, for best results you might need to tweak them to make use of namespaces. A more detailed guide will be available shortly, but the examples here and this [Dataset Explorer](https://huggingface.co/spaces/Sygil/INE-dataset-explorer) should be able to start you off on the right track.
42
 
43
  If you find our work useful, please consider supporting us on [OpenCollective](https://opencollective.com/sygil_dev)!
44
 
@@ -81,8 +81,8 @@ image.save("fantasy_forest_illustration.png")
81
  - If you have low GPU RAM available, make sure to add a `pipe.enable_attention_slicing()` after sending it to `cuda` for less VRAM usage (to the cost of speed).
82
 
83
  ## Available Checkpoints:
84
- - [Sygil Diffusion v0.1](https://huggingface.co/Sygil/Sygil-Diffusion/blob/main/sygil-diffusion-v0.1.ckpt): Trained for 800,000 steps
85
- - [sygil-diffusion-v0.2_1708350_lora.ckpt](https://huggingface.co/Sygil/Sygil-Diffusion/blob/main/sygil-diffusion-v0.2_1708350_lora.ckpt): Resumed from Sygil Diffusion v0.1 and now up to 1.70 million steps.
86
 
87
  ## Training
88
 
@@ -92,7 +92,7 @@ The model was trained on the following dataset:
92
 
93
  **Hardware and others**
94
  - **Hardware:** 1 x Nvidia RTX 3050 8GB GPU
95
- - **Hours Trained:** 610 hours approximately.
96
  - **Optimizer:** AdamW
97
  - **Adam Beta 1**: 0.9
98
  - **Adam Beta 2**: 0.999
@@ -105,7 +105,7 @@ The model was trained on the following dataset:
105
  - **Lora unet Learning Rate**: 1e-7
106
  - **Lora Text Encoder Learning Rate**: 1e-7
107
  - **Resolution**: 512 pixels
108
- - **Total Training Steps:** 1,489,983
109
 
110
  Developed by: [ZeroCool94](https://github.com/ZeroCool940711) at [Sygil-Dev](https://github.com/Sygil-Dev/)
111
 
 
32
 
33
  # About the model
34
  -----------------
35
+ This model is a fine-tune of Stable Diffusion, trained on the [Imaginary Network Expanded Dataset](https://github.com/Sygil-Dev/INE-dataset), with the big advantage of allowing the use of multiple namespaces (labeled tags) to control various parts of the final generation.
36
  While current models usually are prone to “context errors” and need substantial negative prompting to set them on the right track, the use of namespaces in this model (eg. “species:seal” or “studio:dc”) stop the model from misinterpreting a seal as the singer Seal, or DC Comics as Washington DC.
37
  This model is also able to understand other languages besides English, currently it can partially understand prompts in Chinese, Japanese and Spanish. More training is already being done in order to have the model completely understand those languages and have it work just like how it works with English prompts.
38
 
39
  As the model is fine-tuned on a wide variety of content, it’s able to generate many types of images and compositions, and easily outperforms the original model when it comes to portraits, architecture, reflections, fantasy, concept art, anime, landscapes and a lot more without being hyper-specialized like other community fine-tunes that are currently available.
40
 
41
+ **Note: The prompt engineering techniques needed are slightly different from other fine-tunes and the original Stable Diffusion model, so while you can still use your favorite prompts, for best results you might need to tweak them to make use of namespaces. A more detailed guide will be available later on, but you can use the tags and namespaces found here [Dataset Explorer](https://huggingface.co/spaces/Sygil/INE-dataset-explorer) should be able to start you off on the right track.
42
 
43
  If you find our work useful, please consider supporting us on [OpenCollective](https://opencollective.com/sygil_dev)!
44
 
 
81
  - If you have low GPU RAM available, make sure to add a `pipe.enable_attention_slicing()` after sending it to `cuda` for less VRAM usage (to the cost of speed).
82
 
83
  ## Available Checkpoints:
84
+ - [Sygil Diffusion v0.1](https://huggingface.co/Sygil/Sygil-Diffusion/blob/main/sygil-diffusion-v0.1.ckpt): Trained on Stable Diffusion 1.5 for 800,000 steps.
85
+ - [sygil-diffusion-v0.2](https://huggingface.co/Sygil/Sygil-Diffusion/blob/main/sygil-diffusion-v0.2.ckpt): Resumed from Sygil Diffusion v0.1 and trained for a total of 1.77 million steps.
86
 
87
  ## Training
88
 
 
92
 
93
  **Hardware and others**
94
  - **Hardware:** 1 x Nvidia RTX 3050 8GB GPU
95
+ - **Hours Trained:** 630 hours approximately.
96
  - **Optimizer:** AdamW
97
  - **Adam Beta 1**: 0.9
98
  - **Adam Beta 2**: 0.999
 
105
  - **Lora unet Learning Rate**: 1e-7
106
  - **Lora Text Encoder Learning Rate**: 1e-7
107
  - **Resolution**: 512 pixels
108
+ - **Total Training Steps:** 1,770,717
109
 
110
  Developed by: [ZeroCool94](https://github.com/ZeroCool940711) at [Sygil-Dev](https://github.com/Sygil-Dev/)
111