hollowstrawberry commited on
Commit
0bf793f
1 Parent(s): f392a14

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +38 -6
README.md CHANGED
@@ -26,7 +26,11 @@ language:
26
  1. [Generation parameters](#gen)
27
  * [Extensions](#extensions)
28
  * [Loras](#lora)
29
- * [Upscalers](#upscale)
 
 
 
 
30
  * [ControlNet](#controlnet)
31
  * [Lora Training](#train)
32
  * [Tips for training character Loras](#trainchars)
@@ -176,7 +180,7 @@ Here you can select your model and VAE. We will go over what these are and how y
176
  Others:
177
  * **Restore faces:** May improve realistic faces. I never need it with the models and prompts listed in this guide as well as hires fix.
178
  * **Tiling:** Used to produce repeating textures to put on a grid. Not very useful.
179
- * **Script:** Lets you access useful features and extensions, such as `X/Y/Z Plot` which lets you compare images with varying parameters on a grid. Very powerful.
180
 
181
   
182
 
@@ -189,7 +193,7 @@ Here are some useful extensions. Most of these come installed in the collab, and
189
  * [Image Browser (fixed fork)](https://github.com/aka7774/sd_images_browser) - This will let you browse your past generated images very efficiently, as well as directly sending their prompts and parameters back to txt2img, img2img, etc.
190
  * [TagComplete](https://github.com/DominikDoom/a1111-sd-webui-tagcomplete) - Absolutely essential for anime art. It will show you the matching booru tags as you type. Anime models work via booru tags, and rarely work at all if you go outside them, so knowing them is godmode. Not all tags will work well in all models though, specially if they're rare.
191
  * [ControlNet](https://github.com/Mikubill/sd-webui-controlnet) - A huge extension deserving of [its own guide ▼](#controlnet). It lets you take AI data from any image and use it as an input for your image. Practically speaking, it can create any pose or environment you want. Very powerful if used with external tools succh as Blender.
192
- * [Ultimate Upscale](https://github.com/Coyote-A/ultimate-upscale-for-automatic1111) - A semi-advanced script usable from the img2img section to make really large images, where normally you can only go as high as your VRAM allows.
193
  * [Two-shot](https://github.com/opparco/stable-diffusion-webui-two-shot) - Normally you can't create more than one distinct character in the same image without them blending together. This extension lets you divide the image into parts; full, left side, right side; allowing you to make nice 2-character images. It is an optional launch setting in the collab.
194
  * [Dynamic Prompts](https://github.com/adieyal/sd-dynamic-prompts) - A script to let you generate randomly chosen elements in your image, among other things.
195
  * [Model Converter](https://github.com/Akegarasu/sd-webui-model-converter) - Lets you convert most 7GB/4GB models down to 2GB, by choosing `safetensors`, `fp16`, and `no-ema`. These pruned models work "almost the same" as the full models, which is to say, there is no appreciable difference due to math reasons. Most models come in 2 GB form nowadays regardless.
@@ -212,7 +216,9 @@ An example of a Lora is [Thicker Lines Anime Style](https://civitai.com/models/1
212
 
213
  # Upscaling <a name="upscale"></a>[▲](#index)
214
 
215
- You can download additional upscalers and put them in your `stable-diffusion-webui/models/ESRGAN` folder. They will then be available in Hires fix, SD Upscale, Ultimate SD Upscaler, and Extras.
 
 
216
 
217
  The collab comes with several of them, including Remacri, which is one of the best for all sorts of images.
218
 
@@ -220,7 +226,26 @@ The collab comes with several of them, including Remacri, which is one of the be
220
  * LDSR is an advanced yet slow upscaler, its model and config can be [found here](https://huggingface.co/hollowstrawberry/upscalers-backup/tree/main/LDSR) and both must be placed in `stable-diffusion-webui/models/LDSR`.
221
  * The [Upscale Wiki](https://upscale.wiki/wiki/Model_Database) contains dozens of historical choices.
222
 
223
- Coming soon: How to use ultimate upscaler.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
224
 
225
  &nbsp;
226
 
@@ -271,7 +296,14 @@ First, you must scroll down in the txt2img page and click on ControlNet to open
271
 
272
  * **Scribble**
273
 
274
- Lets you make a crude sketch and convert it into a finished piece with the help of your prompt. Example soon.
 
 
 
 
 
 
 
275
 
276
  You will notice that there are 2 results for each method. The first is an intermediate step called the *preprocessed image*, which is then used to produce the final image. You can supply the preprocessed image yourself, in which case you should set the preprocessor to *None*. This is extremely powerful with external tools such as Blender.
277
 
 
26
  1. [Generation parameters](#gen)
27
  * [Extensions](#extensions)
28
  * [Loras](#lora)
29
+ * [Upscaling](#upscale)
30
+ * [Scripts](#scripts)
31
+ * [X/Y/Z Plot](#plot)
32
+ * [Prompt Matrix](#matrix)
33
+ * [Ultimate Upscaler](#ultimate)
34
  * [ControlNet](#controlnet)
35
  * [Lora Training](#train)
36
  * [Tips for training character Loras](#trainchars)
 
180
  Others:
181
  * **Restore faces:** May improve realistic faces. I never need it with the models and prompts listed in this guide as well as hires fix.
182
  * **Tiling:** Used to produce repeating textures to put on a grid. Not very useful.
183
+ * **Script:** Lets you access useful features and extensions, such as [X/Y/Z Plot ▼](#plot) which lets you compare images with varying parameters on a grid. Very powerful.
184
 
185
  &nbsp;
186
 
 
193
  * [Image Browser (fixed fork)](https://github.com/aka7774/sd_images_browser) - This will let you browse your past generated images very efficiently, as well as directly sending their prompts and parameters back to txt2img, img2img, etc.
194
  * [TagComplete](https://github.com/DominikDoom/a1111-sd-webui-tagcomplete) - Absolutely essential for anime art. It will show you the matching booru tags as you type. Anime models work via booru tags, and rarely work at all if you go outside them, so knowing them is godmode. Not all tags will work well in all models though, specially if they're rare.
195
  * [ControlNet](https://github.com/Mikubill/sd-webui-controlnet) - A huge extension deserving of [its own guide ▼](#controlnet). It lets you take AI data from any image and use it as an input for your image. Practically speaking, it can create any pose or environment you want. Very powerful if used with external tools succh as Blender.
196
+ * [Ultimate Upscale](https://github.com/Coyote-A/ultimate-upscale-for-automatic1111) - A semi-advanced script usable from the img2img section to make really large images, where normally you can only go as high as your VRAM allows. See [Ultimate Upscaler ▼](#ultimate).
197
  * [Two-shot](https://github.com/opparco/stable-diffusion-webui-two-shot) - Normally you can't create more than one distinct character in the same image without them blending together. This extension lets you divide the image into parts; full, left side, right side; allowing you to make nice 2-character images. It is an optional launch setting in the collab.
198
  * [Dynamic Prompts](https://github.com/adieyal/sd-dynamic-prompts) - A script to let you generate randomly chosen elements in your image, among other things.
199
  * [Model Converter](https://github.com/Akegarasu/sd-webui-model-converter) - Lets you convert most 7GB/4GB models down to 2GB, by choosing `safetensors`, `fp16`, and `no-ema`. These pruned models work "almost the same" as the full models, which is to say, there is no appreciable difference due to math reasons. Most models come in 2 GB form nowadays regardless.
 
216
 
217
  # Upscaling <a name="upscale"></a>[▲](#index)
218
 
219
+ As mentioned in [Generation Parameters ▲](#parameters), normally you shouldn't go above 768 width or height when generating an image. Instead you should use `Hires. fix` with your choice of upscaler and an appropiate denoising level. Hires fix is limited by your VRAM however, so you may be interested in [Ultimate Upscaler ▼](#ultimate) to go even larger.
220
+
221
+ You can download additional upscalers and put them in your `stable-diffusion-webui/models/ESRGAN` folder. They will then be available in Hires fix, Ultimate Upscaler, and Extras.
222
 
223
  The collab comes with several of them, including Remacri, which is one of the best for all sorts of images.
224
 
 
226
  * LDSR is an advanced yet slow upscaler, its model and config can be [found here](https://huggingface.co/hollowstrawberry/upscalers-backup/tree/main/LDSR) and both must be placed in `stable-diffusion-webui/models/LDSR`.
227
  * The [Upscale Wiki](https://upscale.wiki/wiki/Model_Database) contains dozens of historical choices.
228
 
229
+ In the future I may present examples for each upscaler.
230
+
231
+ &nbsp;
232
+
233
+ # Scripts <a name="scripts"></a>[▲](#index)
234
+
235
+ Scripts can be found at the bottom of your generation parameters in txt2img or img2img.
236
+
237
+ * **X/Y/Z Plot** <a name="plot"></a>[▲](#index)
238
+
239
+ Capable of generating a series of images, usually with the exact same seed, but varying parameters of your choice. Can compare almost anything you want, including different models, parts of your prompt, sampler, upscaler and much more. You can have 1, 2, or 3 variable parameters, hence the X, Y and Z.
240
+
241
+ Your parameters in X/Y/Z Plot are separated by commas, but anything else can go inbetween. The most common parameter to compare is **S/R Prompt**, where the first term is a phrase in your prompt and each term afterwards will replace the original. Knowing this, you can compare, say, Lora intensity, like this:
242
+
243
+ `<lora:my lora:0.4>, <lora:my lora:0.6>, <lora:my lora:0.8>, <lora:my lora:1>`
244
+
245
+ Here I made a comparison between different **models** (columns) and faces of different ethnicities via **S/R Prompt** (rows):
246
+
247
+ ![X Y Z plot of models and ethnicities](#XYZplot)
248
+
249
 
250
  &nbsp;
251
 
 
296
 
297
  * **Scribble**
298
 
299
+ Lets you make a simple sketch and convert it into a finished piece with the help of your prompt. For this one I asked a friend:
300
+
301
+ <details>
302
+ <summary>Scribble example, click to open</summary>
303
+
304
+ ![Scribble sample image](images/scribble1.png)
305
+ ![Scribble output image](images/scribble2.png)
306
+ </details>
307
 
308
  You will notice that there are 2 results for each method. The first is an intermediate step called the *preprocessed image*, which is then used to produce the final image. You can supply the preprocessed image yourself, in which case you should set the preprocessor to *None*. This is extremely powerful with external tools such as Blender.
309