hollowstrawberry commited on
Commit
6a642df
•
1 Parent(s): 421cc73

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +91 -77
README.md CHANGED
@@ -8,30 +8,29 @@ tags:
8
  - stable-diffusion-webui
9
  ---
10
 
11
- **[CLICK HERE TO OPEN THIS DOCUMENT IN FULL WIDTH](https://huggingface.co/hollowstrawberry/stable-diffusion-guide/blob/main/README.md#sdguide-index)**
12
 
13
   
14
 
15
- # Index <a name="sdguide-index"></a>
16
-
17
- * [Introduction](#sdguide-intro)
18
- * [Installation](#sdguide-install)
19
- * [Getting Started](#sdguide-start)
20
- 1. [Edit your starting parameters](#sdguide-params)
21
- 1. [Getting a model](#sdguide-model)
22
- 1. [Getting a VAE](#sdguide-vae)
23
- 1. [Launching and settings](#sdguide-launch)
24
- 1. [Prompts](#sdguide-prompt)
25
- 1. [Generation parameters](#sdguide-gen)
26
- * [Extensions](#sdguide-extensions)
27
- * [Loras](#sdguide-lora)
28
- * [Upscalers](#sdguide-upscale)
29
  * ControlNet
30
  * Tips for training character Loras
31
 
32
  &nbsp;
33
 
34
- # Introduction <a name="sdguide-intro"></a>[â–²](#sdguide-index)
35
 
36
  Stable Diffusion is a very powerful AI image generation software you can run on your own home computer. It uses "models", which function like the brain of the AI, and can make almost anything given that someone has trained it to do it. The biggest uses are anime art, photorealism, and NSFW content.
37
 
@@ -39,70 +38,83 @@ The images you create may be used for any purpose, depending on the used model's
39
 
40
  &nbsp;
41
 
42
- # Installation <a name="sdguide-install"></a>[â–²](#sdguide-index)
43
 
44
- * __**Requirements:**__ To run Stable Diffusion on your own computer you'll need at least 16 GB of RAM and 4 GB of VRAM. I will only cover the case where you are running Windows 10/11 and using an NVIDIA graphics card series 16XX, 20XX or 30XX (though 10XX also work). AMD users are out of luck, as it's very inconsistent to get it working. Same with Linux and Mac, though it's possible to do it.
45
 
46
- * __**Installer:**__ The easiest way is to download the latest release [HERE](https://github.com/EmpireMediaScience/A1111-Web-UI-Installer/releases).
47
 
48
- * __**Alternative:**__ If you don't meet the hardware requirements, don't worry, you can still use Stable Diffusion for free to its full extent through Google Collab. It borrows Google's computers to use AI, with variable time limitations, usually a few hours every day. To get started, [go here](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb#scrollTo=PjzwxTkPSPHf) and follow the steps.
 
 
 
 
 
 
 
 
49
 
50
  &nbsp;
51
 
52
- # Getting Started <a name="sdguide-start"></a>[â–²](#sdguide-index)
 
 
53
 
54
- Before generating some images, here are some useful steps you can follow to improve your experience.
55
 
56
- 1. **Edit your starting parameters** <a name="sdguide-params"></a>[â–²](#sdguide-index)
57
 
58
- If you're using the collab, skip this step.
59
 
60
- If you're using the launcher, turn on **medvram** and **xformers**. Then, set your *Additional Launch Options* to: `--opt-channelslast --no-half-vae`. All of these should offer minor but significant improvements to performance.
61
- * If your graphics card has more than 8 GB of VRAM, you may turn off medvram to make generations faster. However, medvram still allows you to generate larger images and more images at the same time.
62
- * If your graphics card has 4 or 6 GB of VRAM, add `--opt-split-attention-v1` as it may lower vram usage even further.
63
  * If you want to run the program from your computer but want to use it in another device, such as your phone, add `--listen`. Then, use your computer's local IP in the same WiFi network to access the interface.
64
- * If you're using the original stable-diffusion-webui, you can add these parameters by editing your webui-user.bat, right next to `set COMMANDLINE_ARGS=`
65
  * Full list of possible parameters [here](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Command-Line-Arguments-and-Settings)
66
 
67
- 1. **Getting a model** <a name="sdguide-model"></a>[â–²](#sdguide-index)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
68
 
69
- The model is the brain of your AI, designed for the purpose of producing certain types of images. There are many options, most of which are on [civitai](https://civitai.com). But which to choose? These are my recommendations:
70
- * For anime, [7th Heaven Mix](https://civitai.com/models/4669/corneos-7th-heaven-mix) has a nice aesthetic similar to anime movies, while [Abyss Orange Mix 3](https://civitai.com/models/9942/abyssorangemix3-aom3) *(__Note:__ scroll down and choose the AOM3 option)* offers more realism in the form of advanced lighting and softer shading, as well as more lewdness. I remixed the two options above into [Heaven Orange Mix](https://civitai.com/models/14305/heavenorangemix). While AOM3 is extremely capable for NSFW, the popular [Grapefruit](https://civitai.com/models/2583/grapefruit-hentai-model) hentai model may also fit your needs.
 
71
  * For general art go with [DreamShaper](https://civitai.com/models/4384/dreamshaper), there are few options quite like it in terms of raw creativity. An honorable mention goes to [Pastel Mix](https://civitai.com/models/5414/pastel-mix-stylized-anime-model), which has a beautiful and unique aesthetic with the addition of anime.
72
  * For photorealism go with [Deliberate](https://civitai.com/models/4823/deliberate). It can do almost anything, but specially photographs. Very intricate results.
73
  * The [Uber Realistic Porn Merge](https://civitai.com/models/2661/uber-realistic-porn-merge-urpm) is self-explanatory.
74
 
75
  *Launcher:* It will let you choose the path to your models folder. Otherwise the models normally go into `stable-diffusion-webui/models/Stable-diffusion`.
76
 
77
- *Collab:* copy the **direct download link to the file** and put it in `MODEL_LINK:`. Turn on `safetensors`, and `Use_temp_storage` if you don't want to save it to your google drive. After the first time you use the collab, you may place more models manually into your Google Drive folder at: `MyDrive/sd/stable-diffusion-webui/models/Stable-diffusion`
78
 
79
- Please note that models in the format `.safetensors` are safe to use while `.ckpt` **may** contain viruses. Be careful.
80
 
81
- 1. **Getting a VAE** <a name="sdguide-vae"></a>[â–²](#sdguide-index)
82
 
83
  Most models don't come with a VAE built in. The VAE is a small separate model, which "converts your image from AI format into human format". Without it, you'll get faded colors and ugly eyes, among other things.
84
 
 
 
85
  There are practically only 3 different VAEs out there worth talking about:
86
- * [anime vae](https://huggingface.co/WarriorMama777/OrangeMixs/resolve/main/VAEs/orangemix.vae.pt), also known as the AnythingV3 vae, also known as the orangemix vae. All anime models use this.
87
  * [vae-ft-mse](https://huggingface.co/stabilityai/sd-vae-ft-mse-original/blob/main/vae-ft-mse-840000-ema-pruned.safetensors), the latest from Stable Diffusion itself. Used by photorealism models and such.
88
- * [kl-f8-anime2](https://huggingface.co/hakurei/waifu-diffusion-v1-4/resolve/main/vae/kl-f8-anime2.ckpt), also known as the waifu diffusion VAE, it is older and produces much brighter results. Used by Pastel Mix.
89
 
90
  *Launcher:* It lets you choose the default VAE, otherwise put them in the `stable-diffusion-webui/models/VAE` folder.
91
-
92
- *Collab:* You will have to place it in your Google Drive, in `MyDrive/sd/stable-diffusion-webui/models/VAE`.
93
 
94
- 1. **Launching and settings** <a name="sdguide-launch"></a>[â–²](#sdguide-index)
95
-
96
- It is finally time to launch the WebUI.
97
- *Launcher:* Press the button on the launcher and wait patiently for it to start. Then, it will open the interface in your browser. It's like a website, but on your computer.
98
- *Collab:* Press the play buttons, **in order, one at a time**. Wait for each one to finish before pressing the next one. **You may skip the ControlNet section this time**. When the final step is finished, it will produce a link you can use to access the interface as a website. This will be open as long as the page stays open.
99
-
100
- The starting page is where you can make your images. But first, we'll go to the Settings tab. There will be sections on the left.
101
- * In the *Stable Diffusion* section, scroll down and increase **Clip Skip** from 1 to 2. This is said to produce better images, specially for anime. You can also set your VAE from here, but I have a better idea:
102
- * In the *User Interface* section, scroll down to **Quicksettings list** and change it to `sd_model_checkpoint, sd_vae`.
103
- * Scroll back up, click the big orange **Apply settings** button, then **Reload UI** next to it. You can now change your model as well as your VAE from the top of the page at any time.
104
-
105
- 1. **Prompts** <a name="sdguide-prompt"></a>[â–²](#sdguide-index)
106
 
107
  On the first tab, **txt2img**, you'll be making most of your images. This is where you'll find your *prompt* and *negative prompt*.
108
  Stable Diffusion is not like Midjourney or other popular image generation software, you can't just ask it what you want and get a good image. You have to be specific. *Very* specific.
@@ -117,62 +129,64 @@ Before generating some images, here are some useful steps you can follow to impr
117
  * `EasyNegative, worst quality, low quality, normal quality, child, painting, drawing, sketch, cartoon, anime, render, 3d, blurry, deformed, disfigured, morbid, mutated, bad anatomy, bad art`
118
 
119
  * **EasyNegative:** The negative prompts above use EasyNegative, which is a *textual inversion embedding* or "magic word" that codifies many bad things to make your images better. Typically one would write a very long, very specific, very redundant, and sometimes silly negative prompt. EasyNegative is as of March 2023 the best choice if you want to avoid that.
120
- * [Get EasyNegative here](https://huggingface.co/datasets/gsdf/EasyNegative/resolve/main/EasyNegative.safetensors) and put it in your `stable-diffusion-webui/embeddings` folder. Then, go to the bottom of your WebUI page and click *Reload UI*. It will now work when you type the word.
 
 
 
121
 
122
- After a "base prompt" like the above, you may then start typing what you want. For example `young woman in a bikini in the beach, full body shot`. Feel free to add other terms you don't like to your negatives such as `old, ugly, futanari, furry`, etc.
123
- You can also save your prompts to reuse later with the buttons below Generate. Click the small 💾 *Save style* and give it a name. Later, you can open your *Styles* dropdown to choose, then click 📋 *Apply selected styles to the current prompt*.
124
 
125
- Note that when you surround something in `(parentheses)`, it will have emphasis or more **weight** in your resulting image, equal to `1.1`. The normal weight is 1, and each parentheses will multiply by an additional 1.1. You can also specify the weight yourself, like this: `(full body:1.4)`. You can also go below 1 to de-emphasize a word: `[brackets]` will multiply by 0.9, but you still use normal parentheses to go lower, like `(this:0.5)`.
126
 
127
- 1. **Generation parameters** <a name="sdguide-gen"></a>[â–²](#sdguide-index)
128
 
129
- * *Sampling method:* These dictate how your image is formulated, and each produce different results. The default of `Euler a` is almost always the best. There are also very good results for `DPM++ 2M Karras` and `DPM++ SDE Karras`.
130
- * *Sampling steps:* These are "calculated" beforehand, and so more steps doesn't always mean more detail. I always go with 30, you may go from 20-50 and find good results.
131
- * *Width and Height:* 512x512 is the default, you should almost never go above 768 in either direction as it may distort and deform your image. To produce bigger images see `Hires. fix`
132
- * *Batch Count and Batch Size:* Batch *size* is how many images your graphics card will create at the same time, which is limited by your graphics card. Batch count is how many repeats of those to produce. Batches have sequential seeds, more on seeds below.
133
- * *CFG Scale:* "Lower values produce more creative results". You should almost always stick to 7, but 4 to 10 is an acceptable range. It gets strange outside that.
134
- * *Seed:* A number that guides the creation of your image. The same seed with the same prompt and parameters produces almost exacly the same image every time.
135
 
136
- *Hires. fix:* Lets you create larger images without distortion. Usually used at 2x scale. When selected, more options appear:
137
- * *Upscaler:* The algorithm to upscale with. `Latent` and its variations produce creative results, and you may also like `R-ESRGAN 4x+` and its anime version. Also see [Upscalers](#sdguide-upscale).
138
- * *Hires steps:* I recommend at least half as many as your sampling steps. Higher values aren't always better.
139
- * *Denoising strength:* The most important parameter. Near 0, no detail will be added to the image. Near 1, the image will be changed completely. I recommend something between 0.2 and 0.6 depending on the image.
140
 
141
  Others:
142
- * *Restore faces:* May improve realistic faces. I never need it with the models and prompts listed in this guid as well as hires fix.
143
- * *Tiling:* Used to produce repeating textures to put on a grid. Not very useful.
144
- * *Script:* Lets you access useful features and extensions such as `X/Y/Z Plot` which lets you compare images with varying parameters on a grid.
145
 
146
  &nbsp;
147
 
148
- # Extensions <a name="sdguide-extensions"></a>[â–²](#sdguide-index)
149
 
150
  *Stable Diffusion WebUI* supports extensions to add additional functionality and quality of life. These can be added by going into the **Extensions** tab, then **Install from URL**, and pasting the links found here or elsewhere. Then, click *Install* and wait for it to finish. Then, go to **Installed** and click *Apply and restart UI*.
151
 
152
- Here are some useful extensions, I hugely recommend the first 2:
153
  * [Image Browser (fixed fork)](https://github.com/aka7774/sd_images_browser) - This will let you browse your past generated images very efficiently, as well as directly sending their prompts and parameters back to txt2img, img2img, etc.
154
  * [TagComplete](https://github.com/DominikDoom/a1111-sd-webui-tagcomplete) - Absolutely essential for anime art. It will show you the matching booru tags as you type. Anime models work via booru tags, and rarely work at all if you go outside them, so knowing them is godmode. Not all tags will work well in all models though, specially if they're rare.
155
- * [ControlNet](https://github.com/Mikubill/sd-webui-controlnet) - A huge extension deserving of its own guide. It lets you take AI data from any image and use it as an input for your image. Practically speaking, it can create any pose or environment you want. Very powerful if used with external tools succh as Blender.
156
  * [Ultimate Upscale](https://github.com/Coyote-A/ultimate-upscale-for-automatic1111) - A semi-advanced script usable from the img2img section to make really large images, where normally you can only go as high as your VRAM allows.
157
- * [Two-shot](https://github.com/opparco/stable-diffusion-webui-two-shot) - Normally you can't create more than one distinct character in the same image without them blending together. This extension lets you divide the image into parts; full, left side, right side; allowing you to make nice 2-character images.
158
  * [Dynamic Prompts](https://github.com/adieyal/sd-dynamic-prompts) - A script to let you generate randomly chosen elements in your image, among other things.
159
  * [Model Converter](https://github.com/Akegarasu/sd-webui-model-converter) - Lets you convert most 7GB/4GB models down to 2GB, by choosing `safetensors`, `fp16`, and `no-ema`. These pruned models work "almost the same" as the full models, which is to say, there is no appreciable difference due to math reasons. Most models come in 2 GB form nowadays regardless.
160
 
161
  &nbsp;
162
 
163
- # Loras <a name="sdguide-lora"></a>[â–²](#sdguide-index)
164
 
165
- LoRA or *Low-Rank Adaptation* is a form of **Extra Network** and the latest technology that lets you append a smaller model to any of your full models. They are similar to embeddings, one of which you might've seen [earlier](#sdguide-prompt), but Loras are larger and often more capable. Technical details omitted.
166
 
167
- Loras can represent a character, an artstyle, poses, clothes, or even a human face (though I do not endorse this). Models are usually capable enough for general work, but when it comes to specific details with little existing examples, they fall short. That's where Loras come in. They can be downloaded from [civitai](https://civitai.com) and are 144 MB by default, but they can go as low as 1 MB and sometimes several hundreds of MB. Bigger Loras are not necessarily better. They come in `.safetensor` format, same as models.
168
 
169
- Place your lora files in the `stable-diffusion-webui/models/Lora` folder. Then, look for the 🎴 *Show extra networks* button below the big orange Generate button. It will open a new section. Click on the Lora tab and press the **Refresh** button, and your loras should appear. When you click a Lora in that menu it will get added to your prompt. It will look like this: `<lora:filename:1` The start is always the same. The filename will be the exact filename in your system without the `.safetensors` extension. Finally, the number is the weight, like we saw in [Prompts](#sdguide-prompt). Most Loras work between 0.5 and 1 weight, and too high values might "fry" your image.
170
 
171
  An example of a Lora is [Thicker Lines Anime Style](https://civitai.com/models/13910/thicker-lines-anime-style-lora-mix), which is perfect if you want your images to look more like traditional anime.
172
 
173
  &nbsp;
174
 
175
- # Upscalers <a name="sdguide-upscale"></a>[â–²](#sdguide-index)
176
 
177
  You can download additional upscalers and put them in your `stable-diffusion-webui/models/ESRGAN` folder.
178
 
 
8
  - stable-diffusion-webui
9
  ---
10
 
11
+ **[CLICK HERE TO OPEN THIS DOCUMENT IN FULL WIDTH](https://huggingface.co/hollowstrawberry/stable-diffusion-guide/blob/main/README.md#index)**
12
 
13
  &nbsp;
14
 
15
+ # Index <a name="index"></a>
16
+
17
+ * [Introduction](#intro)
18
+ * [Google Collab](#collab)
19
+ * [Local Installation (Windows + Nvidia)](#install)
20
+ * [Getting Started](#start)
21
+ 1. [Models](#model)
22
+ 1. [VAEs](#vae)
23
+ 1. [Prompts](#prompt)
24
+ 1. [Generation parameters](#gen)
25
+ * [Extensions](#extensions)
26
+ * [Loras](#lora)
27
+ * [Upscalers](#upscale)
 
28
  * ControlNet
29
  * Tips for training character Loras
30
 
31
  &nbsp;
32
 
33
+ # Introduction <a name="intro"></a>[â–²](#index)
34
 
35
  Stable Diffusion is a very powerful AI image generation software you can run on your own home computer. It uses "models", which function like the brain of the AI, and can make almost anything given that someone has trained it to do it. The biggest uses are anime art, photorealism, and NSFW content.
36
 
 
38
 
39
  &nbsp;
40
 
41
+ # Google Collab <a name="collab"></a>[â–²](#index)
42
 
43
+ The easiest way to use Stable Diffusion is through Google Collab. It borrows Google's computers to use AI, with variable time limitations, usually a few hours every day. You will need a Google account (or several, wink wink) and we will be using Google Drive for ease of access.
44
 
45
+ If you instead want to run it on your own computer, [scroll down](#install).
46
 
47
+ 1. Enter [this page](https://colab.research.google.com/drive/1wEa-tS10h4LlDykd87TF5zzpXIIQoCmq).
48
+
49
+ 2. Near the top, click **Copy to Drive**. Wait for the new window to open and close the old one.
50
+
51
+ 3. Turn on the following options under **Configurations**: `output_to_drive, configs_in_drive, no_custom_theme`. Then, turn on the following options under **Models, etc**: `anything_vae`, `wd_vae`, `sd_vae`.
52
+
53
+ 4. Press the play button to the left, anywhere in the first section of the page labeled **Start 🚀**. Wait a few minutes for it to finish, then, a public gradio link will appear, which you can open in a new tab to start using Stable Diffusion. **Keep both of these tabs open!**
54
+
55
+ 5. You can now make some decent anime images thanks to the default **Anything 4.5** model. But we can do better, as well as things other than anime. Also, what are all of these options? [Scroll down](#start) to get started.
56
 
57
  &nbsp;
58
 
59
+ # Local Installation (Windows + Nvidia) <a name="install"></a>[â–²](#index)
60
+
61
+ To run Stable Diffusion on your own computer you'll need at least 16 GB of RAM and 4 GB of VRAM (preferably 8). I will only cover the case where you are running Windows 10/11 and using an NVIDIA graphics card series 16XX, 20XX or 30XX (though 10XX also work). My apologies to AMD, Linux, and Mac users, but their cases are harder to cover. If you don't meet the hardware requirements, you can just proceed with the Google Collab method [above](#collab).
62
 
63
+ 1. Get the latest release from [this page](https://github.com/EmpireMediaScience/A1111-Web-UI-Installer/releases).
64
 
65
+ 2. Run the installer, choose a simple location to install to, and wait for it to finish.
66
 
67
+ 3. Run the program. You will see a few options. First, turn on **medvram** and **xformers**. You may skip medvram if you have 12 or more GB of VRAM.
68
 
69
+ 4. Set your *Additional Launch Options* to: `--opt-channelslast --no-half-vae`
70
+ * If your graphics card has less than 8 GB of VRAM, add `--opt-split-attention-v1` as it may lower vram usage even further.
 
71
  * If you want to run the program from your computer but want to use it in another device, such as your phone, add `--listen`. Then, use your computer's local IP in the same WiFi network to access the interface.
 
72
  * Full list of possible parameters [here](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Command-Line-Arguments-and-Settings)
73
 
74
+ 5. Click **Launch** and wait for a browser window to open with the interface. It may take a while the first time.
75
+
76
+ 6. The page is now open. It's your own private website. The starting page is where you can make your images. But first, we'll go to the **Settings** tab. There will be sections of settings on the left.
77
+ * In the *Stable Diffusion* section, scroll down and increase **Clip Skip** from 1 to 2. This is said to produce better images, specially for anime.
78
+ * In the *User Interface* section, scroll down to **Quicksettings list** and change it to `sd_model_checkpoint, sd_vae`.
79
+ * Scroll back up, click the big orange **Apply settings** button, then **Reload UI** next to it.
80
+
81
+ 7. You can now generate some images, but you only have the basic model available. It's not great, at most it can make some paintings. Also, what are all of these options? See [below](#start) to get started.
82
+
83
+ &nbsp;
84
+
85
+ # Getting Started <a name="start"></a>[â–²](#index)
86
+
87
+ Before or after generating your first few images, here are several useful steps you can follow to improve your experience and results.
88
+
89
+ 1. **Models** <a name="model"></a>[â–²](#index)
90
 
91
+ The **model**, also called **checkpoint**, is the brain of your AI, designed for the purpose of producing certain types of images. There are many options, most of which are on [civitai](https://civitai.com). But which to choose? These are my recommendations:
92
+ * For anime, [7th Heaven Mix](https://civitai.com/models/4669/corneos-7th-heaven-mix) has a nice aesthetic similar to anime movies, while [Abyss Orange Mix 3](https://civitai.com/models/9942/abyssorangemix3-aom3) *(__Note:__ scroll down and choose the AOM3 option)* offers more realism in the form of advanced lighting and softer shading, as well as more lewdness. I remixed the two options above into [Heaven Orange Mix](https://civitai.com/models/14305/heavenorangemix).
93
+ * While AOM3 is extremely capable for NSFW, the popular [Grapefruit](https://civitai.com/models/2583/grapefruit-hentai-model) hentai model may also fit your needs.
94
  * For general art go with [DreamShaper](https://civitai.com/models/4384/dreamshaper), there are few options quite like it in terms of raw creativity. An honorable mention goes to [Pastel Mix](https://civitai.com/models/5414/pastel-mix-stylized-anime-model), which has a beautiful and unique aesthetic with the addition of anime.
95
  * For photorealism go with [Deliberate](https://civitai.com/models/4823/deliberate). It can do almost anything, but specially photographs. Very intricate results.
96
  * The [Uber Realistic Porn Merge](https://civitai.com/models/2661/uber-realistic-porn-merge-urpm) is self-explanatory.
97
 
98
  *Launcher:* It will let you choose the path to your models folder. Otherwise the models normally go into `stable-diffusion-webui/models/Stable-diffusion`.
99
 
100
+ *Collab:* Copy the **direct download link to the file** and paste it in the text box labeled `custom_urls`. Multiple links are separated by commas.
101
 
102
+ Please note that checkpoints in the format `.safetensors` are safe to use while `.ckpt` **may** contain viruses. Be careful.
103
 
104
+ 1. **VAEs** <a name="vae"></a>[â–²](#index)
105
 
106
  Most models don't come with a VAE built in. The VAE is a small separate model, which "converts your image from AI format into human format". Without it, you'll get faded colors and ugly eyes, among other things.
107
 
108
+ If you're using the collab, you should already have the below VAEs, which you can select at the top of the page, next to your models.
109
+
110
  There are practically only 3 different VAEs out there worth talking about:
111
+ * [anything vae](https://huggingface.co/WarriorMama777/OrangeMixs/resolve/main/VAEs/orangemix.vae.pt), also known as the orangemix vae. All anime models use this.
112
  * [vae-ft-mse](https://huggingface.co/stabilityai/sd-vae-ft-mse-original/blob/main/vae-ft-mse-840000-ema-pruned.safetensors), the latest from Stable Diffusion itself. Used by photorealism models and such.
113
+ * [kl-f8-anime2](https://huggingface.co/hakurei/waifu-diffusion-v1-4/resolve/main/vae/kl-f8-anime2.ckpt), also known as the waifu diffusion VAE, it is older and produces more saturated results. Used by Pastel Mix.
114
 
115
  *Launcher:* It lets you choose the default VAE, otherwise put them in the `stable-diffusion-webui/models/VAE` folder.
 
 
116
 
117
+ 1. **Prompts** <a name="prompt"></a>[â–²](#index)
 
 
 
 
 
 
 
 
 
 
 
118
 
119
  On the first tab, **txt2img**, you'll be making most of your images. This is where you'll find your *prompt* and *negative prompt*.
120
  Stable Diffusion is not like Midjourney or other popular image generation software, you can't just ask it what you want and get a good image. You have to be specific. *Very* specific.
 
129
  * `EasyNegative, worst quality, low quality, normal quality, child, painting, drawing, sketch, cartoon, anime, render, 3d, blurry, deformed, disfigured, morbid, mutated, bad anatomy, bad art`
130
 
131
  * **EasyNegative:** The negative prompts above use EasyNegative, which is a *textual inversion embedding* or "magic word" that codifies many bad things to make your images better. Typically one would write a very long, very specific, very redundant, and sometimes silly negative prompt. EasyNegative is as of March 2023 the best choice if you want to avoid that.
132
+ * [Get EasyNegative here](https://huggingface.co/datasets/gsdf/EasyNegative/resolve/main/EasyNegative.pt). For collab, paste the link into the `custom_urls` text box. For Windows, put it in your `stable-diffusion-webui/embeddings` folder. Then, go to the bottom of your WebUI page and click *Reload UI*. It will now work when you type the word.
133
+
134
+ After a "base prompt" like the above, you may then start typing what you want. For example `young woman in a bikini in the beach, full body shot`. Feel free to add other terms you don't like to your negatives such as `old, ugly, futanari, furry`, etc.
135
+ You can also save your prompts to reuse later with the buttons below Generate. Click the small 💾 *Save style* button and give it a name. Later, you can open your *Styles* dropdown to choose, then click 📋 *Apply selected styles to the current prompt*.
136
 
137
+ Note that when you surround something in `(parentheses)`, it will have more emphasis or **weight** in your resulting image, equal to `1.1`. The normal weight is 1, and each parentheses will multiply by an additional 1.1. You can also specify the weight yourself, like this: `(full body:1.4)`. You can also go below 1 to de-emphasize a word: `[brackets]` will multiply by 0.9, but you must still use normal parentheses to go lower, like `(this:0.5)`.
 
138
 
139
+ 1. **Generation parameters** <a name="gen"></a>[â–²](#index)
140
 
141
+ At the top of the page you'll be able to choose your checkpoint and VAE, and we've already covered the prompt. Here are the rest of the options:
142
 
143
+ * **Sampling method:** These dictate how your image is formulated, and each produce different results. The default of `Euler a` is almost always the best. There are also very good results for `DPM++ 2M Karras` and `DPM++ SDE Karras`.
144
+ * **Sampling steps:** These are "calculated" beforehand, and so more steps doesn't always mean more detail. I always go with 30, you may go from 20-50 and find good results.
145
+ * **Width and Height:** 512x512 is the default, and you should almost never go above 768 in either direction as it may distort and deform your image. To produce bigger images see `Hires. fix`
146
+ * **Batch Count and Batch Size:** Batch *size* is how many images your graphics card will create at the same time, which is limited by your graphics card. Batch count is how many repeats of those to produce. Batches have sequential seeds, more on seeds below.
147
+ * **CFG Scale:** "Lower values produce more creative results". You should almost always stick to 7, but 4 to 10 is an acceptable range. It gets strange outside that.
148
+ * **Seed:** A number that guides the creation of your image. The same seed with the same prompt and parameters produces almost exacly the same image every time.
149
 
150
+ **Hires. fix:** Lets you create larger images without distortion. Often used at 2x scale. When selected, more options appear:
151
+ * **Upscaler:** The algorithm to upscale with. `Latent` and its variations produce creative results, and you may also like `R-ESRGAN 4x+` and its anime version. Also see [Upscalers](#upscale).
152
+ * **Hires steps:** I recommend at least half as many as your sampling steps. Higher values aren't always better, and they take a long time, so be conservative here.
153
+ * **Denoising strength:** The most important parameter. Near 0.0, no detail will be added to the image. Near 1.0, the image will be changed completely. I recommend something between 0.2 and 0.6 depending on the image, to add enough detail as the image gets larger, without *destroying* any original details you like.
154
 
155
  Others:
156
+ * **Restore faces:** May improve realistic faces. I never need it with the models and prompts listed in this guide as well as hires fix.
157
+ * **Tiling:** Used to produce repeating textures to put on a grid. Not very useful.
158
+ * **Script:** Lets you access useful features and extensions, such as `X/Y/Z Plot` which lets you compare images with varying parameters on a grid. Very powerful.
159
 
160
  &nbsp;
161
 
162
+ # Extensions <a name="extensions"></a>[â–²](#index)
163
 
164
  *Stable Diffusion WebUI* supports extensions to add additional functionality and quality of life. These can be added by going into the **Extensions** tab, then **Install from URL**, and pasting the links found here or elsewhere. Then, click *Install* and wait for it to finish. Then, go to **Installed** and click *Apply and restart UI*.
165
 
166
+ Here are some useful extensions, most of these come installed in the collab, and I hugely recommend the first 2 if you're running locally:
167
  * [Image Browser (fixed fork)](https://github.com/aka7774/sd_images_browser) - This will let you browse your past generated images very efficiently, as well as directly sending their prompts and parameters back to txt2img, img2img, etc.
168
  * [TagComplete](https://github.com/DominikDoom/a1111-sd-webui-tagcomplete) - Absolutely essential for anime art. It will show you the matching booru tags as you type. Anime models work via booru tags, and rarely work at all if you go outside them, so knowing them is godmode. Not all tags will work well in all models though, specially if they're rare.
169
+ * [ControlNet](https://github.com/Mikubill/sd-webui-controlnet) - A huge extension deserving of its own guide (coming soon). It lets you take AI data from any image and use it as an input for your image. Practically speaking, it can create any pose or environment you want. Very powerful if used with external tools succh as Blender.
170
  * [Ultimate Upscale](https://github.com/Coyote-A/ultimate-upscale-for-automatic1111) - A semi-advanced script usable from the img2img section to make really large images, where normally you can only go as high as your VRAM allows.
171
+ * [Two-shot](https://github.com/opparco/stable-diffusion-webui-two-shot) - Normally you can't create more than one distinct character in the same image without them blending together. This extension lets you divide the image into parts; full, left side, right side; allowing you to make nice 2-character images. It is an optional launch setting in the collab.
172
  * [Dynamic Prompts](https://github.com/adieyal/sd-dynamic-prompts) - A script to let you generate randomly chosen elements in your image, among other things.
173
  * [Model Converter](https://github.com/Akegarasu/sd-webui-model-converter) - Lets you convert most 7GB/4GB models down to 2GB, by choosing `safetensors`, `fp16`, and `no-ema`. These pruned models work "almost the same" as the full models, which is to say, there is no appreciable difference due to math reasons. Most models come in 2 GB form nowadays regardless.
174
 
175
  &nbsp;
176
 
177
+ # Loras <a name="lora"></a>[â–²](#index)
178
 
179
+ LoRA or *Low-Rank Adaptation* is a form of **Extra Network** and the latest technology that lets you append a smaller model to any of your full models. They are similar to embeddings, one of which you might've seen [earlier](#prompt), but Loras are larger and often more capable. Technical details omitted.
180
 
181
+ Loras can represent a character, an artstyle, poses, clothes, or even a human face (though I do not endorse this). Checkpoints are usually capable enough for general work, but when it comes to specific details with little existing examples, they fall short. That's where Loras come in. They can be downloaded from [civitai](https://civitai.com) or [elsewhere (NSFW)](https://gitgud.io/gayshit/makesomefuckingporn#lora-list) and are 144 MB by default, but they can go as low as 1 MB and sometimes several hundreds of MB. Bigger Loras are not necessarily better. They come in `.safetensors` format, same as most checkpoints.
182
 
183
+ Place your lora files in the `stable-diffusion-webui/models/Lora` folder, or for collab, paste the direct download link to their file into the `custom_urls` text box. Then, look for the 🎴 *Show extra networks* button below the big orange Generate button. It will open a new section. Click on the Lora tab and press the **Refresh** button, and your loras should appear. When you click a Lora in that menu it will get added to your prompt, looking like this: `<lora:filename:1>`. The start is always the same. The filename will be the exact filename in your system without the `.safetensors` extension. Finally, the number is the weight, like we saw in [Prompts](#prompt). Most Loras work between 0.5 and 1 weight, and too high values might "fry" your image, specially if using multiple Loras at the same time.
184
 
185
  An example of a Lora is [Thicker Lines Anime Style](https://civitai.com/models/13910/thicker-lines-anime-style-lora-mix), which is perfect if you want your images to look more like traditional anime.
186
 
187
  &nbsp;
188
 
189
+ # Upscalers <a name="upscale"></a>[â–²](#index)
190
 
191
  You can download additional upscalers and put them in your `stable-diffusion-webui/models/ESRGAN` folder.
192