initial commit for spop model
Browse files- .gitattributes +2 -0
- README.md +210 -1
- dwspop12_13440.safetensors +3 -0
- dwspop12_13440.yaml +70 -0
- feature_extractor/preprocessor_config.json +28 -0
- images/00787-1219898403-a photo of a woman with long purple hairwith a happy expression floating in space dwspop stylea4e527532c9db223a920409540ef1349024e35a3.png +3 -0
- images/00796-2257850020-a photo of a woman with long purple hair with a happy expression floating in outerspace dwspop style2ec894da3f237b6ddb0e77fa2dee23abf4062a00.png +3 -0
- images/00995-198300932-Perfectly-centered portrait-photograph of emma stone laughing at a purple mushroom near a cottage in a forest dwspop styleb33c4927b579e69a09f325eaa8d8222cde6967cd.png +3 -0
- images/01016-585905465-Perfectly-centered portrait-photograph of superman with big blue eyes standing near a small village dwspop stylecfe4536946b21beefd84252956f60c06ac3e6825.png +3 -0
- images/01105-1864722104-aaliyah with detailed glittering eyes floating through outer space with asteroids and rocks dwspop style79cddf6b8a84f5fd5e6ef422866f53299f1b201c.png +3 -0
- images/01109-920945318-a lizard with big green eyes and long dark hair and wearing a helmet on their head near a huge volcano dwspop styleff53f0a833597d635acf37b008c39cf3ad9c4d25.png +3 -0
- images/01153-1557785744-aaliyah with detailed glittering eyes floating through outer space with asteroids and rocks dwspop style4f0cca6a94e53f1d2cd466491a7e990ac9d065b9.png +3 -0
- images/01155-2736036669-princess zelda with detailed glittering eyes looking out a window in a castle dwspop styled7f84d42d4d597e2a5c95dc84c9efa84484ea7e5.png +3 -0
- images/01166-4165139102-Perfectly-centered portrait-photograph of snoop dog with long braids glittering red big eyes standing in a forest of maraje55b96d91297b7ba414c2bce500323a928f701ca.png +3 -0
- images/01900-2286177263-a photo of a path in a forest with glowing lights and rocks and trees on either side of the path dynamic lighting bright light6c6b818658e2ff962e137c397916f6b1cdee97de.png +3 -0
- images/01910-1243702611-a scene of a tattered cottage in the middle of a huge and thick forest dynamic lighting bright lights dwspop forestac777ecd6310af3d72750c11fc5330f191b1a7fb.png +3 -0
- images/02079-2474176896-a scene of a cracked mountain with a castle in the middle of it and a sky filled with clouds above dwspop landscapee20aed695c14f3968a4b92717bcbea0591aab578.png +3 -0
- images/02150-2325504057-red sun in the background dwspop landscapeee51fe7fe16dba7ca38b928c72f9a7b5068aa250.png +3 -0
- images/02218-86584250-red sun in the background dwspop landscape7770a975793e1abd170babd1d0fa07cd3143a7e1.png +3 -0
- images/02247-3780749798-a scene of a village built into a massive mountain with stairs dwspop landscapefcba91a63d08ef4964b41887ea9fffb5760ba1a7.png +3 -0
- images/02258-2177805898-a scene of a a mountain with fire burning in the sky and a cracked ground surface in the foreground dwspop landscape5b965a056213b89e710f9067c0e42335bf59a5ec.png +3 -0
- images/02277-1634402072-a scene of a planet in space with stars in the background dwspop spacee89f3c716238c27939cd68ba0d3190747abfa54d.png +3 -0
- images/02280-432062230-a beautiful photo of a forked path in a forest with glowing lights and rocks and trees on either side of the path dwspop forestb650addb2bf906942dffe2ae9385b0f2cf82cd28.png +3 -0
- images/02319-2148816834-a beautiful photo of a woman showing her palms to the camera dwspop style86ef8692b4a9300da030d44954a8a8aab9128b8f.png +3 -0
- images/02371-1707325028-Perfectly-centered portrait-photograph of jason momoa in a forest dwspop stylee5393d28293aa9b89f10a24dbc23b34032aacc03.png +3 -0
- images/cfg_grid.png +3 -0
- images/dwspop_forest_grid.png +3 -0
- images/dwspop_landscape_grid.png +3 -0
- images/dwspop_space_grid.png +3 -0
- images/img2img_denoise_samplers.png +3 -0
- images/img2img_sampler_cfg.png +3 -0
- images/img2img_steps_denoising.png +3 -0
- images/sampler_grid.png +3 -0
- model_index.json +33 -0
- safety_checker/config.json +181 -0
- safety_checker/pytorch_model.bin +3 -0
- scheduler/scheduler_config.json +13 -0
- text_encoder/config.json +25 -0
- text_encoder/pytorch_model.bin +3 -0
- tokenizer/merges.txt +0 -0
- tokenizer/special_tokens_map.json +24 -0
- tokenizer/tokenizer_config.json +34 -0
- tokenizer/vocab.json +0 -0
- unet/config.json +42 -0
- unet/diffusion_pytorch_model.bin +3 -0
- vae/config.json +31 -0
- vae/diffusion_pytorch_model.bin +3 -0
.gitattributes
CHANGED
@@ -32,3 +32,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
35 |
+
*.png filter=lfs diff=lfs merge=lfs -text
|
36 |
+
*.jpg filter=lfs diff=lfs merge=lfs -text
|
README.md
CHANGED
@@ -1,3 +1,212 @@
|
|
1 |
---
|
2 |
-
license:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
license: creativeml-openrail-m
|
3 |
+
tags:
|
4 |
+
- stable-diffusion
|
5 |
+
- stable-diffusion-diffusers
|
6 |
+
- diffusers
|
7 |
+
- text-to-image
|
8 |
+
- image-to-image
|
9 |
+
- art
|
10 |
+
- artistic
|
11 |
+
- dreambooth
|
12 |
---
|
13 |
+
|
14 |
+
# spop style
|
15 |
+
|
16 |
+
This model features four different concepts: humans, outer space, forests, and landscapes in the specific style of SPOP: She-Ra and the Princesses of Power, the Dreamworks version.
|
17 |
+
|
18 |
+
This is a fine-tuned Stable Diffusion model, based on ```SD 1.5```.
|
19 |
+
|
20 |
+
The goal of this model is to capture the _style_ - not the individual characters featured in the series.
|
21 |
+
|
22 |
+
> 💖 **Disclaimer**: This is my favorite show. I won't go into that here but a lot of love went into this model.
|
23 |
+
|
24 |
+
![Detailed Samples](https://huggingface.co/zuleo/spop/resolve/main/images/00787-1219898403-a photo of a woman with long purple hairwith a happy expression floating in space dwspop stylea4e527532c9db223a920409540ef1349024e35a3.png)
|
25 |
+
|
26 |
+
![Detailed Samples](https://huggingface.co/zuleo/spop/resolve/main/images/01153-1557785744-aaliyah with detailed glittering eyes floating through outer space with asteroids and rocks dwspop style4f0cca6a94e53f1d2cd466491a7e990ac9d065b9.png)
|
27 |
+
|
28 |
+
## Model Usage
|
29 |
+
|
30 |
+
This model was trained on multiple concepts. Use the tokens below:
|
31 |
+
|
32 |
+
| Token | Description |
|
33 |
+
|-----------------------|--------------------------------------|
|
34 |
+
| 👤 `dwspop style` | Uses concepts trained on people |
|
35 |
+
| 🌌 `dwspop space` | Uses concepts trained on outer space |
|
36 |
+
| 🌲 `dwspop forest` | Uses concepts trained on forests |
|
37 |
+
| 🌄 `dwspop landscape` | Uses concepts trained on landscapes |
|
38 |
+
|
39 |
+
|
40 |
+
### 👤 `dwspop style` style examples:
|
41 |
+
|
42 |
+
This token is capable of handling multiple genders and uses `person` which can be then used for `woman`, `man`,
|
43 |
+
or `cat-like woman`, or even `lizard`, `dog`, `snoop dog`... it's awesome:
|
44 |
+
|
45 |
+
- ```a photo of a person in a forest, dwspop style```
|
46 |
+
|
47 |
+
- ```a photo of a woman floating in space, dwspop style```
|
48 |
+
|
49 |
+
- ```a photo of a man inside of a palace standing near a window, dwspop style```
|
50 |
+
|
51 |
+
⛔ Negative prompt: ```((out of focus body)), ((out of focus face)), ((((ugly)))), (((duplicate))), ((morbid)), ((mutilated)), [out of frame], extra fingers, mutated hands, ((poorly drawn hands)), ((poorly drawn face)), (((mutation))), (((deformed))), ((ugly)), blurry, ((bad anatomy)), (((bad proportions))), ((extra limbs)), cloned face, (((disfigured))), out of frame, ugly, extra limbs, (bad anatomy), gross proportions, (malformed limbs), ((missing arms)), ((missing legs)), (((extra arms))), (((extra legs))), mutated hands, (fused fingers), (too many fingers), (((long neck)))```
|
52 |
+
|
53 |
+
### 🌌 `dwspop space` outer space examples:
|
54 |
+
|
55 |
+
The space token is versatile when prompting, especially when generating galaxies and solar systems. This token is capable of handling different camera angles by desribing in your prompts as a `scene`.
|
56 |
+
|
57 |
+
- ```a scene of outer space with asteroids and rocks floating in space getting melted by a bright light, dwspop space```
|
58 |
+
|
59 |
+
- ```a scene of an outer space solar system with planets, stars and galaxies in the background, dwspop space```
|
60 |
+
|
61 |
+
- ```a scene of a planet in space with stars in the background, dwspop space```
|
62 |
+
|
63 |
+
⛔ Negative prompt: ```((out of focus face)), (((duplicate))), [out of frame], blurry, out of frame, ugly, blur, motion blur```
|
64 |
+
|
65 |
+
|
66 |
+
### 🌲 `dwspop forest` forest examples:
|
67 |
+
|
68 |
+
The forest token is able to generate random forest scenes due to the regularization images that were used. When prompting, additional enviromental objects are supported, such as `crystals`, `rocks`, `flowers`, `cottage`. Finally, mix in time of day: `sunrise`, `dawn`, `sunset`, `evening`.
|
69 |
+
|
70 |
+
- ```a beautiful photo of a path in a forest with glowing lights and rocks and trees on either side of the path, dwspop forest```
|
71 |
+
|
72 |
+
- ```a forest during night time with a full moon in the sky, dynamic lighting, bright lights, dwspop forest```
|
73 |
+
|
74 |
+
- ```a scene of an entrance to a huge forest with pink flowers, dynamic lighting, bright lights, dwspop forest```
|
75 |
+
|
76 |
+
⛔ Negative prompt: ```((out of focus face)), (((duplicate))), [out of frame], blurry, out of frame, ugly, blur, motion blur```
|
77 |
+
|
78 |
+
### 🌄 `dwspop landscape` landscape examples:
|
79 |
+
|
80 |
+
The landscape token is primarly for landscapes but also supports a small percentage of architecture. Blending your prompts to have both an establishing shot of a landscape with architecture woven in and out is where this token shines.
|
81 |
+
|
82 |
+
- ```a scene of a weapon shop that has many different swords hanging on the wall and arrows and staffs inside of barrels, a small shop with a tent in the background, dwspop landscape```
|
83 |
+
|
84 |
+
- ```a scene of a village with a waterfall, wooden stairs leading to the top of trees, dynamic lighting, dwspop landscape```
|
85 |
+
|
86 |
+
- ```a beautiful scene of a palace with wide doors and a fountain and flowers near a window, sunset, dynamic lighting, dwspop landscape```
|
87 |
+
|
88 |
+
⛔ Negative prompt: ```((out of focus face)), (((duplicate))), [out of frame], blurry, out of frame, ugly, blur, motion blur```
|
89 |
+
|
90 |
+
---
|
91 |
+
|
92 |
+
## 🧨 Diffusers
|
93 |
+
|
94 |
+
This model can be used just like any other Stable Diffusion model. For more information,
|
95 |
+
see [Stable Diffusion](https://huggingface.co/docs/diffusers/api/pipelines/stable_diffusion).
|
96 |
+
|
97 |
+
Export the model:
|
98 |
+
- [ONNX](https://huggingface.co/docs/diffusers/optimization/onnx)
|
99 |
+
- [MPS](https://huggingface.co/docs/diffusers/optimization/mps)
|
100 |
+
- [FLAX/JAX](https://huggingface.co/blog/stable_diffusion_jax)
|
101 |
+
|
102 |
+
|
103 |
+
```python
|
104 |
+
from diffusers import StableDiffusionPipeline
|
105 |
+
import torch
|
106 |
+
|
107 |
+
model_id = "zuleo/spop"
|
108 |
+
pipe = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float16)
|
109 |
+
pipe = pipe.to("cuda")
|
110 |
+
|
111 |
+
prompt = "Perfectly-centered close up portrait-photograph of a person, marketplace in the background, sunrise, dwspop style"
|
112 |
+
image = pipe(prompt).images[0]
|
113 |
+
|
114 |
+
image.save("./spop_person.png")
|
115 |
+
```
|
116 |
+
|
117 |
+
---
|
118 |
+
|
119 |
+
## 📅 text2img Range Grids
|
120 |
+
|
121 |
+
It's always great to get a visual of what's going on with sampler, CFG scale, and other settings. See the examples below and tune them to your liking.
|
122 |
+
|
123 |
+
### Sampler
|
124 |
+
|
125 |
+
Using different samplers can produce different results. My favorites are using `DPM++ 2S a Karras`, `DPM++ SDE Karras`, `DPM adaptive` for cartoon scenes.
|
126 |
+
|
127 |
+
> 🔥 **DPM Adaptive**: DPM Adaptive does not use steps. This sampler is fixed depending on the CFG scale and additional configurations.
|
128 |
+
|
129 |
+
View the XY grids below for details:
|
130 |
+
|
131 |
+
- Space: https://huggingface.co/zuleo/spop/resolve/main/images/dwspop_space_grid.png
|
132 |
+
|
133 |
+
- Forest: https://huggingface.co/zuleo/spop/resolve/main/images/dwspop_forest_grid.png
|
134 |
+
|
135 |
+
- Landscape: https://huggingface.co/zuleo/spop/resolve/main/images/dwspop_landscape_grid.png
|
136 |
+
|
137 |
+
|
138 |
+
### Sampling Steps for person
|
139 |
+
|
140 |
+
Values between `25 - 38` is a good range for _most_ samplers but not all. See the Sampling Steps grid with each sampler below:
|
141 |
+
|
142 |
+
[Sampling Steps Grid](https://huggingface.co/zuleo/spop/resolve/main/images/sampler_grid.png)
|
143 |
+
|
144 |
+
|
145 |
+
### CFG Scale
|
146 |
+
|
147 |
+
Values between `7 - 11` is a good range. See the CFG Scale grid:
|
148 |
+
|
149 |
+
[CFG Scale Grid](https://huggingface.co/zuleo/spop/resolve/main/images/cfg_grid.png)
|
150 |
+
|
151 |
+
---
|
152 |
+
|
153 |
+
## 📅 img2img Grids
|
154 |
+
|
155 |
+
This model works with img2img with a balanced configuration between `CFG scale`, `denoising`, and adding more detail with `sampling steps`.
|
156 |
+
|
157 |
+
### Denoising & Steps
|
158 |
+
|
159 |
+
Steps: `39 - 46`, Denoising: `0.49 - 0.6`:
|
160 |
+
|
161 |
+
- [Denoising & Steps Grid](https://huggingface.co/zuleo/spop/resolve/main/images/img2img_steps_denoising.png)
|
162 |
+
|
163 |
+
|
164 |
+
### Samplers & Denoising
|
165 |
+
|
166 |
+
Samplers: `all`, Denoising: `0.6 - 0.7`:
|
167 |
+
|
168 |
+
- [Samplers & Denoising Grid](https://huggingface.co/zuleo/spop/resolve/main/images/img2img_denoise_samplers.png)
|
169 |
+
|
170 |
+
### Samplers & CFG Scale
|
171 |
+
|
172 |
+
Samplers: `all`, CFG Scale: `7.0 - 11.0`:
|
173 |
+
|
174 |
+
- [Samplers & CFG Scale Grid](https://huggingface.co/zuleo/spop/resolve/main/images/img2img_sampler_cfg.png)
|
175 |
+
|
176 |
+
---
|
177 |
+
|
178 |
+
## 🌐 Regularization images
|
179 |
+
|
180 |
+
If you would like to use the regularization images from this training, see the datasets below:
|
181 |
+
|
182 |
+
- `space`: https://huggingface.co/datasets/3ee/regularization-space
|
183 |
+
|
184 |
+
- `forest`: https://huggingface.co/datasets/3ee/regularization-forest
|
185 |
+
|
186 |
+
- `landscape`: https://huggingface.co/datasets/3ee/regularization-landscape
|
187 |
+
|
188 |
+
---
|
189 |
+
|
190 |
+
## 📊 Training Stats
|
191 |
+
|
192 |
+
🔋 GPU: `Nvidia 3080 16GB`
|
193 |
+
|
194 |
+
🔬 Learning Rate: `1e-6` & Training steps per image: `96` (`13400` total)
|
195 |
+
|
196 |
+
💬 Text Encoder Training: `0.2`
|
197 |
+
|
198 |
+
🎨 Total input images per concept: `14` (total = `14 * 4 = 56`)
|
199 |
+
|
200 |
+
☕ cups of coffee consumed: `21`
|
201 |
+
|
202 |
+
- I spent a lot of time experimenting and fine tuning. I started with `25` steps per image at first and went on from there.
|
203 |
+
|
204 |
+
- I went as high as `176` steps per image and attempted many different `batch sizes` with `Gradient Accumulation`. The results were different (but really good!) than what I had planned for this model which is: capture the style of SPOP. I plan to use the knowledge I gained from those "failed" models to great use on future ones!
|
205 |
+
|
206 |
+
- In conclusion, the `13400` version came out to be the version that captured the Dreamworks style.
|
207 |
+
|
208 |
+
---
|
209 |
+
|
210 |
+
☕ If you enjoy this model, buy me a coffee [![Buy a coffee](https://badgen.net/badge/icon/kofi?icon=kofi&label=buy%20us%20a%20coffee)](https://ko-fi.com/3eegames)
|
211 |
+
|
212 |
+
---
|
dwspop12_13440.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cc878ede7ed0d75576a6ef11f66eac2bd97718dd4ad690c18e25bf5e172565cb
|
3 |
+
size 4097789433
|
dwspop12_13440.yaml
ADDED
@@ -0,0 +1,70 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
model:
|
2 |
+
base_learning_rate: 1.0e-04
|
3 |
+
target: ldm.models.diffusion.ddpm.LatentDiffusion
|
4 |
+
params:
|
5 |
+
linear_start: 0.00085
|
6 |
+
linear_end: 0.0120
|
7 |
+
num_timesteps_cond: 1
|
8 |
+
log_every_t: 200
|
9 |
+
timesteps: 1000
|
10 |
+
first_stage_key: "image"
|
11 |
+
cond_stage_key: "caption"
|
12 |
+
image_size: 64
|
13 |
+
channels: 4
|
14 |
+
cond_stage_trainable: true # Note: different from the one we trained before
|
15 |
+
conditioning_key: crossattn
|
16 |
+
monitor: val/loss_simple_ema
|
17 |
+
scale_factor: 0.18215
|
18 |
+
use_ema: False
|
19 |
+
|
20 |
+
scheduler_config: # 10000 warmup steps
|
21 |
+
target: ldm.lr_scheduler.LambdaLinearScheduler
|
22 |
+
params:
|
23 |
+
warm_up_steps: [ 10000 ]
|
24 |
+
cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
|
25 |
+
f_start: [ 1.e-6 ]
|
26 |
+
f_max: [ 1. ]
|
27 |
+
f_min: [ 1. ]
|
28 |
+
|
29 |
+
unet_config:
|
30 |
+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
|
31 |
+
params:
|
32 |
+
image_size: 32 # unused
|
33 |
+
in_channels: 4
|
34 |
+
out_channels: 4
|
35 |
+
model_channels: 320
|
36 |
+
attention_resolutions: [ 4, 2, 1 ]
|
37 |
+
num_res_blocks: 2
|
38 |
+
channel_mult: [ 1, 2, 4, 4 ]
|
39 |
+
num_heads: 8
|
40 |
+
use_spatial_transformer: True
|
41 |
+
transformer_depth: 1
|
42 |
+
context_dim: 768
|
43 |
+
use_checkpoint: True
|
44 |
+
legacy: False
|
45 |
+
|
46 |
+
first_stage_config:
|
47 |
+
target: ldm.models.autoencoder.AutoencoderKL
|
48 |
+
params:
|
49 |
+
embed_dim: 4
|
50 |
+
monitor: val/rec_loss
|
51 |
+
ddconfig:
|
52 |
+
double_z: true
|
53 |
+
z_channels: 4
|
54 |
+
resolution: 256
|
55 |
+
in_channels: 3
|
56 |
+
out_ch: 3
|
57 |
+
ch: 128
|
58 |
+
ch_mult:
|
59 |
+
- 1
|
60 |
+
- 2
|
61 |
+
- 4
|
62 |
+
- 4
|
63 |
+
num_res_blocks: 2
|
64 |
+
attn_resolutions: [ ]
|
65 |
+
dropout: 0.0
|
66 |
+
lossconfig:
|
67 |
+
target: torch.nn.Identity
|
68 |
+
|
69 |
+
cond_stage_config:
|
70 |
+
target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
|
feature_extractor/preprocessor_config.json
ADDED
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"crop_size": {
|
3 |
+
"height": 224,
|
4 |
+
"width": 224
|
5 |
+
},
|
6 |
+
"do_center_crop": true,
|
7 |
+
"do_convert_rgb": true,
|
8 |
+
"do_normalize": true,
|
9 |
+
"do_rescale": true,
|
10 |
+
"do_resize": true,
|
11 |
+
"feature_extractor_type": "CLIPFeatureExtractor",
|
12 |
+
"image_mean": [
|
13 |
+
0.48145466,
|
14 |
+
0.4578275,
|
15 |
+
0.40821073
|
16 |
+
],
|
17 |
+
"image_processor_type": "CLIPImageProcessor",
|
18 |
+
"image_std": [
|
19 |
+
0.26862954,
|
20 |
+
0.26130258,
|
21 |
+
0.27577711
|
22 |
+
],
|
23 |
+
"resample": 3,
|
24 |
+
"rescale_factor": 0.00392156862745098,
|
25 |
+
"size": {
|
26 |
+
"shortest_edge": 224
|
27 |
+
}
|
28 |
+
}
|
images/00787-1219898403-a photo of a woman with long purple hairwith a happy expression floating in space dwspop stylea4e527532c9db223a920409540ef1349024e35a3.png
ADDED
Git LFS Details
|
images/00796-2257850020-a photo of a woman with long purple hair with a happy expression floating in outerspace dwspop style2ec894da3f237b6ddb0e77fa2dee23abf4062a00.png
ADDED
Git LFS Details
|
images/00995-198300932-Perfectly-centered portrait-photograph of emma stone laughing at a purple mushroom near a cottage in a forest dwspop styleb33c4927b579e69a09f325eaa8d8222cde6967cd.png
ADDED
Git LFS Details
|
images/01016-585905465-Perfectly-centered portrait-photograph of superman with big blue eyes standing near a small village dwspop stylecfe4536946b21beefd84252956f60c06ac3e6825.png
ADDED
Git LFS Details
|
images/01105-1864722104-aaliyah with detailed glittering eyes floating through outer space with asteroids and rocks dwspop style79cddf6b8a84f5fd5e6ef422866f53299f1b201c.png
ADDED
Git LFS Details
|
images/01109-920945318-a lizard with big green eyes and long dark hair and wearing a helmet on their head near a huge volcano dwspop styleff53f0a833597d635acf37b008c39cf3ad9c4d25.png
ADDED
Git LFS Details
|
images/01153-1557785744-aaliyah with detailed glittering eyes floating through outer space with asteroids and rocks dwspop style4f0cca6a94e53f1d2cd466491a7e990ac9d065b9.png
ADDED
Git LFS Details
|
images/01155-2736036669-princess zelda with detailed glittering eyes looking out a window in a castle dwspop styled7f84d42d4d597e2a5c95dc84c9efa84484ea7e5.png
ADDED
Git LFS Details
|
images/01166-4165139102-Perfectly-centered portrait-photograph of snoop dog with long braids glittering red big eyes standing in a forest of maraje55b96d91297b7ba414c2bce500323a928f701ca.png
ADDED
Git LFS Details
|
images/01900-2286177263-a photo of a path in a forest with glowing lights and rocks and trees on either side of the path dynamic lighting bright light6c6b818658e2ff962e137c397916f6b1cdee97de.png
ADDED
Git LFS Details
|
images/01910-1243702611-a scene of a tattered cottage in the middle of a huge and thick forest dynamic lighting bright lights dwspop forestac777ecd6310af3d72750c11fc5330f191b1a7fb.png
ADDED
Git LFS Details
|
images/02079-2474176896-a scene of a cracked mountain with a castle in the middle of it and a sky filled with clouds above dwspop landscapee20aed695c14f3968a4b92717bcbea0591aab578.png
ADDED
Git LFS Details
|
images/02150-2325504057-red sun in the background dwspop landscapeee51fe7fe16dba7ca38b928c72f9a7b5068aa250.png
ADDED
Git LFS Details
|
images/02218-86584250-red sun in the background dwspop landscape7770a975793e1abd170babd1d0fa07cd3143a7e1.png
ADDED
Git LFS Details
|
images/02247-3780749798-a scene of a village built into a massive mountain with stairs dwspop landscapefcba91a63d08ef4964b41887ea9fffb5760ba1a7.png
ADDED
Git LFS Details
|
images/02258-2177805898-a scene of a a mountain with fire burning in the sky and a cracked ground surface in the foreground dwspop landscape5b965a056213b89e710f9067c0e42335bf59a5ec.png
ADDED
Git LFS Details
|
images/02277-1634402072-a scene of a planet in space with stars in the background dwspop spacee89f3c716238c27939cd68ba0d3190747abfa54d.png
ADDED
Git LFS Details
|
images/02280-432062230-a beautiful photo of a forked path in a forest with glowing lights and rocks and trees on either side of the path dwspop forestb650addb2bf906942dffe2ae9385b0f2cf82cd28.png
ADDED
Git LFS Details
|
images/02319-2148816834-a beautiful photo of a woman showing her palms to the camera dwspop style86ef8692b4a9300da030d44954a8a8aab9128b8f.png
ADDED
Git LFS Details
|
images/02371-1707325028-Perfectly-centered portrait-photograph of jason momoa in a forest dwspop stylee5393d28293aa9b89f10a24dbc23b34032aacc03.png
ADDED
Git LFS Details
|
images/cfg_grid.png
ADDED
Git LFS Details
|
images/dwspop_forest_grid.png
ADDED
Git LFS Details
|
images/dwspop_landscape_grid.png
ADDED
Git LFS Details
|
images/dwspop_space_grid.png
ADDED
Git LFS Details
|
images/img2img_denoise_samplers.png
ADDED
Git LFS Details
|
images/img2img_sampler_cfg.png
ADDED
Git LFS Details
|
images/img2img_steps_denoising.png
ADDED
Git LFS Details
|
images/sampler_grid.png
ADDED
Git LFS Details
|
model_index.json
ADDED
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "StableDiffusionPipeline",
|
3 |
+
"_diffusers_version": "0.10.2",
|
4 |
+
"feature_extractor": [
|
5 |
+
"transformers",
|
6 |
+
"CLIPImageProcessor"
|
7 |
+
],
|
8 |
+
"requires_safety_checker": null,
|
9 |
+
"safety_checker": [
|
10 |
+
null,
|
11 |
+
null
|
12 |
+
],
|
13 |
+
"scheduler": [
|
14 |
+
"diffusers",
|
15 |
+
"DDIMScheduler"
|
16 |
+
],
|
17 |
+
"text_encoder": [
|
18 |
+
"transformers",
|
19 |
+
"CLIPTextModel"
|
20 |
+
],
|
21 |
+
"tokenizer": [
|
22 |
+
"transformers",
|
23 |
+
"CLIPTokenizer"
|
24 |
+
],
|
25 |
+
"unet": [
|
26 |
+
"diffusers",
|
27 |
+
"UNet2DConditionModel"
|
28 |
+
],
|
29 |
+
"vae": [
|
30 |
+
"diffusers",
|
31 |
+
"AutoencoderKL"
|
32 |
+
]
|
33 |
+
}
|
safety_checker/config.json
ADDED
@@ -0,0 +1,181 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_commit_hash": "cb41f3a270d63d454d385fc2e4f571c487c253c5",
|
3 |
+
"_name_or_path": "CompVis/stable-diffusion-safety-checker",
|
4 |
+
"architectures": [
|
5 |
+
"StableDiffusionSafetyChecker"
|
6 |
+
],
|
7 |
+
"initializer_factor": 1.0,
|
8 |
+
"logit_scale_init_value": 2.6592,
|
9 |
+
"model_type": "clip",
|
10 |
+
"projection_dim": 768,
|
11 |
+
"text_config": {
|
12 |
+
"_name_or_path": "",
|
13 |
+
"add_cross_attention": false,
|
14 |
+
"architectures": null,
|
15 |
+
"attention_dropout": 0.0,
|
16 |
+
"bad_words_ids": null,
|
17 |
+
"begin_suppress_tokens": null,
|
18 |
+
"bos_token_id": 0,
|
19 |
+
"chunk_size_feed_forward": 0,
|
20 |
+
"cross_attention_hidden_size": null,
|
21 |
+
"decoder_start_token_id": null,
|
22 |
+
"diversity_penalty": 0.0,
|
23 |
+
"do_sample": false,
|
24 |
+
"dropout": 0.0,
|
25 |
+
"early_stopping": false,
|
26 |
+
"encoder_no_repeat_ngram_size": 0,
|
27 |
+
"eos_token_id": 2,
|
28 |
+
"exponential_decay_length_penalty": null,
|
29 |
+
"finetuning_task": null,
|
30 |
+
"forced_bos_token_id": null,
|
31 |
+
"forced_eos_token_id": null,
|
32 |
+
"hidden_act": "quick_gelu",
|
33 |
+
"hidden_size": 768,
|
34 |
+
"id2label": {
|
35 |
+
"0": "LABEL_0",
|
36 |
+
"1": "LABEL_1"
|
37 |
+
},
|
38 |
+
"initializer_factor": 1.0,
|
39 |
+
"initializer_range": 0.02,
|
40 |
+
"intermediate_size": 3072,
|
41 |
+
"is_decoder": false,
|
42 |
+
"is_encoder_decoder": false,
|
43 |
+
"label2id": {
|
44 |
+
"LABEL_0": 0,
|
45 |
+
"LABEL_1": 1
|
46 |
+
},
|
47 |
+
"layer_norm_eps": 1e-05,
|
48 |
+
"length_penalty": 1.0,
|
49 |
+
"max_length": 20,
|
50 |
+
"max_position_embeddings": 77,
|
51 |
+
"min_length": 0,
|
52 |
+
"model_type": "clip_text_model",
|
53 |
+
"no_repeat_ngram_size": 0,
|
54 |
+
"num_attention_heads": 12,
|
55 |
+
"num_beam_groups": 1,
|
56 |
+
"num_beams": 1,
|
57 |
+
"num_hidden_layers": 12,
|
58 |
+
"num_return_sequences": 1,
|
59 |
+
"output_attentions": false,
|
60 |
+
"output_hidden_states": false,
|
61 |
+
"output_scores": false,
|
62 |
+
"pad_token_id": 1,
|
63 |
+
"prefix": null,
|
64 |
+
"problem_type": null,
|
65 |
+
"projection_dim": 512,
|
66 |
+
"pruned_heads": {},
|
67 |
+
"remove_invalid_values": false,
|
68 |
+
"repetition_penalty": 1.0,
|
69 |
+
"return_dict": true,
|
70 |
+
"return_dict_in_generate": false,
|
71 |
+
"sep_token_id": null,
|
72 |
+
"suppress_tokens": null,
|
73 |
+
"task_specific_params": null,
|
74 |
+
"temperature": 1.0,
|
75 |
+
"tf_legacy_loss": false,
|
76 |
+
"tie_encoder_decoder": false,
|
77 |
+
"tie_word_embeddings": true,
|
78 |
+
"tokenizer_class": null,
|
79 |
+
"top_k": 50,
|
80 |
+
"top_p": 1.0,
|
81 |
+
"torch_dtype": null,
|
82 |
+
"torchscript": false,
|
83 |
+
"transformers_version": "4.25.1",
|
84 |
+
"typical_p": 1.0,
|
85 |
+
"use_bfloat16": false,
|
86 |
+
"vocab_size": 49408
|
87 |
+
},
|
88 |
+
"text_config_dict": {
|
89 |
+
"hidden_size": 768,
|
90 |
+
"intermediate_size": 3072,
|
91 |
+
"num_attention_heads": 12,
|
92 |
+
"num_hidden_layers": 12
|
93 |
+
},
|
94 |
+
"torch_dtype": "float32",
|
95 |
+
"transformers_version": null,
|
96 |
+
"vision_config": {
|
97 |
+
"_name_or_path": "",
|
98 |
+
"add_cross_attention": false,
|
99 |
+
"architectures": null,
|
100 |
+
"attention_dropout": 0.0,
|
101 |
+
"bad_words_ids": null,
|
102 |
+
"begin_suppress_tokens": null,
|
103 |
+
"bos_token_id": null,
|
104 |
+
"chunk_size_feed_forward": 0,
|
105 |
+
"cross_attention_hidden_size": null,
|
106 |
+
"decoder_start_token_id": null,
|
107 |
+
"diversity_penalty": 0.0,
|
108 |
+
"do_sample": false,
|
109 |
+
"dropout": 0.0,
|
110 |
+
"early_stopping": false,
|
111 |
+
"encoder_no_repeat_ngram_size": 0,
|
112 |
+
"eos_token_id": null,
|
113 |
+
"exponential_decay_length_penalty": null,
|
114 |
+
"finetuning_task": null,
|
115 |
+
"forced_bos_token_id": null,
|
116 |
+
"forced_eos_token_id": null,
|
117 |
+
"hidden_act": "quick_gelu",
|
118 |
+
"hidden_size": 1024,
|
119 |
+
"id2label": {
|
120 |
+
"0": "LABEL_0",
|
121 |
+
"1": "LABEL_1"
|
122 |
+
},
|
123 |
+
"image_size": 224,
|
124 |
+
"initializer_factor": 1.0,
|
125 |
+
"initializer_range": 0.02,
|
126 |
+
"intermediate_size": 4096,
|
127 |
+
"is_decoder": false,
|
128 |
+
"is_encoder_decoder": false,
|
129 |
+
"label2id": {
|
130 |
+
"LABEL_0": 0,
|
131 |
+
"LABEL_1": 1
|
132 |
+
},
|
133 |
+
"layer_norm_eps": 1e-05,
|
134 |
+
"length_penalty": 1.0,
|
135 |
+
"max_length": 20,
|
136 |
+
"min_length": 0,
|
137 |
+
"model_type": "clip_vision_model",
|
138 |
+
"no_repeat_ngram_size": 0,
|
139 |
+
"num_attention_heads": 16,
|
140 |
+
"num_beam_groups": 1,
|
141 |
+
"num_beams": 1,
|
142 |
+
"num_channels": 3,
|
143 |
+
"num_hidden_layers": 24,
|
144 |
+
"num_return_sequences": 1,
|
145 |
+
"output_attentions": false,
|
146 |
+
"output_hidden_states": false,
|
147 |
+
"output_scores": false,
|
148 |
+
"pad_token_id": null,
|
149 |
+
"patch_size": 14,
|
150 |
+
"prefix": null,
|
151 |
+
"problem_type": null,
|
152 |
+
"projection_dim": 512,
|
153 |
+
"pruned_heads": {},
|
154 |
+
"remove_invalid_values": false,
|
155 |
+
"repetition_penalty": 1.0,
|
156 |
+
"return_dict": true,
|
157 |
+
"return_dict_in_generate": false,
|
158 |
+
"sep_token_id": null,
|
159 |
+
"suppress_tokens": null,
|
160 |
+
"task_specific_params": null,
|
161 |
+
"temperature": 1.0,
|
162 |
+
"tf_legacy_loss": false,
|
163 |
+
"tie_encoder_decoder": false,
|
164 |
+
"tie_word_embeddings": true,
|
165 |
+
"tokenizer_class": null,
|
166 |
+
"top_k": 50,
|
167 |
+
"top_p": 1.0,
|
168 |
+
"torch_dtype": null,
|
169 |
+
"torchscript": false,
|
170 |
+
"transformers_version": "4.25.1",
|
171 |
+
"typical_p": 1.0,
|
172 |
+
"use_bfloat16": false
|
173 |
+
},
|
174 |
+
"vision_config_dict": {
|
175 |
+
"hidden_size": 1024,
|
176 |
+
"intermediate_size": 4096,
|
177 |
+
"num_attention_heads": 16,
|
178 |
+
"num_hidden_layers": 24,
|
179 |
+
"patch_size": 14
|
180 |
+
}
|
181 |
+
}
|
safety_checker/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:193490b58ef62739077262e833bf091c66c29488058681ac25cf7df3d8190974
|
3 |
+
size 1216061799
|
scheduler/scheduler_config.json
ADDED
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "DDIMScheduler",
|
3 |
+
"_diffusers_version": "0.10.2",
|
4 |
+
"beta_end": 0.012,
|
5 |
+
"beta_schedule": "scaled_linear",
|
6 |
+
"beta_start": 0.00085,
|
7 |
+
"clip_sample": false,
|
8 |
+
"num_train_timesteps": 1000,
|
9 |
+
"prediction_type": "epsilon",
|
10 |
+
"set_alpha_to_one": false,
|
11 |
+
"steps_offset": 1,
|
12 |
+
"trained_betas": null
|
13 |
+
}
|
text_encoder/config.json
ADDED
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "C:\\dev\\api\\stable-diffusion-webui\\models\\dreambooth\\dwspop12\\working",
|
3 |
+
"architectures": [
|
4 |
+
"CLIPTextModel"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 0,
|
8 |
+
"dropout": 0.0,
|
9 |
+
"eos_token_id": 2,
|
10 |
+
"hidden_act": "quick_gelu",
|
11 |
+
"hidden_size": 768,
|
12 |
+
"initializer_factor": 1.0,
|
13 |
+
"initializer_range": 0.02,
|
14 |
+
"intermediate_size": 3072,
|
15 |
+
"layer_norm_eps": 1e-05,
|
16 |
+
"max_position_embeddings": 77,
|
17 |
+
"model_type": "clip_text_model",
|
18 |
+
"num_attention_heads": 12,
|
19 |
+
"num_hidden_layers": 12,
|
20 |
+
"pad_token_id": 1,
|
21 |
+
"projection_dim": 768,
|
22 |
+
"torch_dtype": "float32",
|
23 |
+
"transformers_version": "4.25.1",
|
24 |
+
"vocab_size": 49408
|
25 |
+
}
|
text_encoder/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:120f3d1a0b7fe5a86b913dd7e7c97d0e729bd71c94314dae174866c9479b5675
|
3 |
+
size 492308087
|
tokenizer/merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer/special_tokens_map.json
ADDED
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<|startoftext|>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": true,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "<|endoftext|>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": true,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": "<|endoftext|>",
|
17 |
+
"unk_token": {
|
18 |
+
"content": "<|endoftext|>",
|
19 |
+
"lstrip": false,
|
20 |
+
"normalized": true,
|
21 |
+
"rstrip": false,
|
22 |
+
"single_word": false
|
23 |
+
}
|
24 |
+
}
|
tokenizer/tokenizer_config.json
ADDED
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"bos_token": {
|
4 |
+
"__type": "AddedToken",
|
5 |
+
"content": "<|startoftext|>",
|
6 |
+
"lstrip": false,
|
7 |
+
"normalized": true,
|
8 |
+
"rstrip": false,
|
9 |
+
"single_word": false
|
10 |
+
},
|
11 |
+
"do_lower_case": true,
|
12 |
+
"eos_token": {
|
13 |
+
"__type": "AddedToken",
|
14 |
+
"content": "<|endoftext|>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": true,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false
|
19 |
+
},
|
20 |
+
"errors": "replace",
|
21 |
+
"model_max_length": 77,
|
22 |
+
"name_or_path": "C:\\dev\\api\\stable-diffusion-webui\\models\\dreambooth\\dwspop12\\working\\tokenizer",
|
23 |
+
"pad_token": "<|endoftext|>",
|
24 |
+
"special_tokens_map_file": "./special_tokens_map.json",
|
25 |
+
"tokenizer_class": "CLIPTokenizer",
|
26 |
+
"unk_token": {
|
27 |
+
"__type": "AddedToken",
|
28 |
+
"content": "<|endoftext|>",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": true,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false
|
33 |
+
}
|
34 |
+
}
|
tokenizer/vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
unet/config.json
ADDED
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "UNet2DConditionModel",
|
3 |
+
"_diffusers_version": "0.10.2",
|
4 |
+
"_name_or_path": "C:\\dev\\api\\stable-diffusion-webui\\models\\dreambooth\\dwspop12\\working",
|
5 |
+
"act_fn": "silu",
|
6 |
+
"attention_head_dim": 8,
|
7 |
+
"block_out_channels": [
|
8 |
+
320,
|
9 |
+
640,
|
10 |
+
1280,
|
11 |
+
1280
|
12 |
+
],
|
13 |
+
"center_input_sample": false,
|
14 |
+
"cross_attention_dim": 768,
|
15 |
+
"down_block_types": [
|
16 |
+
"CrossAttnDownBlock2D",
|
17 |
+
"CrossAttnDownBlock2D",
|
18 |
+
"CrossAttnDownBlock2D",
|
19 |
+
"DownBlock2D"
|
20 |
+
],
|
21 |
+
"downsample_padding": 1,
|
22 |
+
"dual_cross_attention": false,
|
23 |
+
"flip_sin_to_cos": true,
|
24 |
+
"freq_shift": 0,
|
25 |
+
"in_channels": 4,
|
26 |
+
"layers_per_block": 2,
|
27 |
+
"mid_block_scale_factor": 1,
|
28 |
+
"norm_eps": 1e-05,
|
29 |
+
"norm_num_groups": 32,
|
30 |
+
"num_class_embeds": null,
|
31 |
+
"only_cross_attention": false,
|
32 |
+
"out_channels": 4,
|
33 |
+
"sample_size": 64,
|
34 |
+
"up_block_types": [
|
35 |
+
"UpBlock2D",
|
36 |
+
"CrossAttnUpBlock2D",
|
37 |
+
"CrossAttnUpBlock2D",
|
38 |
+
"CrossAttnUpBlock2D"
|
39 |
+
],
|
40 |
+
"upcast_attention": false,
|
41 |
+
"use_linear_projection": false
|
42 |
+
}
|
unet/diffusion_pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cc79beb9129170135c182e03bdd10865ae2769d17b3f2d7a7803a60805f21038
|
3 |
+
size 3438364325
|
vae/config.json
ADDED
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "AutoencoderKL",
|
3 |
+
"_diffusers_version": "0.10.2",
|
4 |
+
"_name_or_path": "C:\\dev\\api\\stable-diffusion-webui\\models\\dreambooth\\dwspop12\\working",
|
5 |
+
"act_fn": "silu",
|
6 |
+
"block_out_channels": [
|
7 |
+
128,
|
8 |
+
256,
|
9 |
+
512,
|
10 |
+
512
|
11 |
+
],
|
12 |
+
"down_block_types": [
|
13 |
+
"DownEncoderBlock2D",
|
14 |
+
"DownEncoderBlock2D",
|
15 |
+
"DownEncoderBlock2D",
|
16 |
+
"DownEncoderBlock2D"
|
17 |
+
],
|
18 |
+
"in_channels": 3,
|
19 |
+
"latent_channels": 4,
|
20 |
+
"layers_per_block": 2,
|
21 |
+
"norm_num_groups": 32,
|
22 |
+
"out_channels": 3,
|
23 |
+
"sample_size": 512,
|
24 |
+
"scaling_factor": 0.18215,
|
25 |
+
"up_block_types": [
|
26 |
+
"UpDecoderBlock2D",
|
27 |
+
"UpDecoderBlock2D",
|
28 |
+
"UpDecoderBlock2D",
|
29 |
+
"UpDecoderBlock2D"
|
30 |
+
]
|
31 |
+
}
|
vae/diffusion_pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1449178401951a158f9a4326edcefc518d2e9d6ac491e0986e5ddf4ebeb2bf72
|
3 |
+
size 167402961
|