Kardbord commited on
Commit
dc85ca0
1 Parent(s): fbaa942

Upload folder using huggingface_hub (#1)

Browse files

- 97d68c49ee781b9ee8669f4cc50c79bd2c15ce0d5dcae7e28bc76d46a28559f9 (b32ef921a83cd3451c9eecffaa0a67c785deb089)
- 4b9aea23b42c1971a95dd51cb2975abf11d9f31c37edba35fd79525f7f85645c (dd24eb13f1f4a76ae007b0d963b89e738559828c)
- 46d2187ce299b2a916ca2e0c36715a7739d2b6e272396f19b47127dd7c8ebd78 (c163d6359eccc89da60df375e3d2e775a3322254)
- 5a9879665e7c5baae2f24679deec50f5d95b11d1df79dd85cfacc21260abbc58 (d58f48f49de661c4716fbe59f67dea6b6ecb3906)
- 0fd121886c7477d5bfb932d60e3a251c871b8c63b591fb57040bf0eaef4072c2 (04d153f8ea0de4f75d9aef4459c0d1966d8eeba1)
- 404aac17a0336138e7d27f0e8f443d82a4254559d77e0dfc7ca1e54209c1b4d0 (6c9f99a9b5e90ce50a68a26080682baf143b0197)

Model Weights.png ADDED
ProtoGen_X3.4-pruned-fp16.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5346d7de1f448e6953a12e9c186f3996ac07b6e1ea6076fc242bc484b48b7c95
3
+ size 1886665781
ProtoGen_X3.4-pruned-fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef8629e2c89e19a993146302418cf1ff3ba0384dd98523eab6b88ac33ead9d39
3
+ size 1886474920
ProtoGen_X3.4.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61a37adf761fbbf4cb3d88da480216341113d0fbcf9f0f77ea71863866a9b6fc
3
+ size 5984615834
ProtoGen_X3.4.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44f90a09727ca8b62ea304e140546a0af96ba6edcb229c20c677aa4460449c21
3
+ size 5984232961
Protogen_x3.4-512.png ADDED
README.md ADDED
@@ -0,0 +1,365 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: creativeml-openrail-m
5
+ tags:
6
+ - stable-diffusion
7
+ - stable-diffusion-diffusers
8
+ - text-to-image
9
+ - art
10
+ - artistic
11
+ - diffusers
12
+ - protogen
13
+ inference: true
14
+ ---
15
+ # Overview
16
+
17
+ This is simply darkstorm2150/Protogen_x3.4_Official_Release with the safety checker disabled.
18
+
19
+ **DO NOT** attempt to use this model to generate harmful or illegal content.
20
+
21
+ <center><img src="https://huggingface.co/darkstorm2150/Protogen_x3.4_Official_Release/resolve/main/Protogen_x3.4-512.png" style="height:690px; border-radius: 8%; border: 10px solid #663380; padding-top:0px;" span title="Protogen x3.4 Raw Output"></center>
22
+
23
+
24
+
25
+ <center><h1>Protogen x3.4 (Photorealism) Official Release</h1></center>
26
+ <center><p><em>Research Model by <a href="https://instagram.com/officialvictorespinoza">darkstorm2150</a></em></p></center>
27
+ </div>
28
+
29
+ ## Table of contents
30
+ * [General info](#general-info)
31
+ * [Granular Adaptive Learning](#granular-adaptive-learning)
32
+ * [Trigger Words](#trigger-words)
33
+ * [Setup](#setup)
34
+ * [Space](#space)
35
+ * [CompVis](#compvis)
36
+ * [Diffusers](#🧨-diffusers)
37
+ * [Checkpoint Merging Data Reference](#checkpoint-merging-data-reference)
38
+ * [License](#license)
39
+
40
+ ## General info
41
+ Protogen x3.4
42
+
43
+ Protogen was warm-started with [Stable Diffusion v1-5](https://huggingface.co/runwayml/stable-diffusion-v1-5) and fine-tuned on various high quality image datasets.
44
+ Version 3.4 continued training from [ProtoGen v2.2](https://huggingface.co/darkstorm2150/Protogen_v2.2_Official_Release) with added photorealism.
45
+
46
+ ## Granular Adaptive Learning
47
+
48
+ Granular adaptive learning is a machine learning technique that focuses on adjusting the learning process at a fine-grained level, rather than making global adjustments to the model. This approach allows the model to adapt to specific patterns or features in the data, rather than making assumptions based on general trends.
49
+
50
+ Granular adaptive learning can be achieved through techniques such as active learning, which allows the model to select the data it wants to learn from, or through the use of reinforcement learning, where the model receives feedback on its performance and adapts based on that feedback. It can also be achieved through techniques such as online learning where the model adjust itself as it receives more data.
51
+
52
+ Granular adaptive learning is often used in situations where the data is highly diverse or non-stationary and where the model needs to adapt quickly to changing patterns. This is often the case in dynamic environments such as robotics, financial markets, and natural language processing.
53
+
54
+ ## Trigger Words
55
+
56
+ modelshoot style, analog style, mdjrny-v4 style, nousr robot
57
+
58
+ Trigger words are available for the hassan1.4 and f222, might have to google them :)
59
+
60
+ ## Setup
61
+ To run this model, download the model.ckpt or model.safetensor and install it in your "stable-diffusion-webui\models\Stable-diffusion" directory
62
+
63
+ ## Space
64
+
65
+ We support a [Gradio](https://github.com/gradio-app/gradio) Web UI:
66
+ [![Open In Spaces](https://camo.githubusercontent.com/00380c35e60d6b04be65d3d94a58332be5cc93779f630bcdfc18ab9a3a7d3388/68747470733a2f2f696d672e736869656c64732e696f2f62616467652f25463025394625413425393725323048756767696e67253230466163652d5370616365732d626c7565)](https://huggingface.co/spaces/darkstorm2150/Stable-Diffusion-Protogen-webui)
67
+
68
+ ### CompVis
69
+
70
+ ## CKPT
71
+ [Download ProtoGen x3.4.ckpt (5.98GB)](https://huggingface.co/darkstorm2150/Protogen_x3.4_Official_Release/resolve/main/ProtoGen_X3.4.ckpt)
72
+
73
+ [Download ProtoGen X3.4-pruned-fp16.ckpt (1.89 GB)](https://huggingface.co/darkstorm2150/Protogen_x3.4_Official_Release/resolve/main/ProtoGen_X3.4-pruned-fp16.ckpt)
74
+
75
+ ## Safetensors
76
+ [Download ProtoGen x3.4.safetensors (5.98GB)](https://huggingface.co/darkstorm2150/Protogen_x3.4_Official_Release/resolve/main/ProtoGen_X3.4.safetensors)
77
+
78
+ [Download ProtoGen x3.4-pruned-fp16.safetensors (1.89GB)](https://huggingface.co/darkstorm2150/Protogen_x3.4_Official_Release/resolve/main/ProtoGen_X3.4-pruned-fp16.safetensors)
79
+
80
+
81
+ ### 🧨 Diffusers
82
+
83
+ This model can be used just like any other Stable Diffusion model. For more information,
84
+ please have a look at the [Stable Diffusion Pipeline](https://huggingface.co/docs/diffusers/api/pipelines/stable_diffusion).
85
+
86
+ ```python
87
+ from diffusers import StableDiffusionPipeline, DPMSolverMultistepScheduler
88
+ import torch
89
+
90
+ prompt = (
91
+ "modelshoot style, (extremely detailed CG unity 8k wallpaper), full shot body photo of the most beautiful artwork in the world, "
92
+ "english medieval witch, black silk vale, pale skin, black silk robe, black cat, necromancy magic, medieval era, "
93
+ "photorealistic painting by Ed Blinkey, Atey Ghailan, Studio Ghibli, by Jeremy Mann, Greg Manchess, Antonio Moro, trending on ArtStation, "
94
+ "trending on CGSociety, Intricate, High Detail, Sharp focus, dramatic, photorealistic painting art by midjourney and greg rutkowski"
95
+ )
96
+
97
+ model_id = "darkstorm2150/Protogen_x3.4_Official_Release"
98
+ pipe = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float16)
99
+ pipe.scheduler = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config)
100
+ pipe = pipe.to("cuda")
101
+
102
+ image = pipe(prompt, num_inference_steps=25).images[0]
103
+
104
+ image.save("./result.jpg")
105
+ ```
106
+
107
+ ![img](https://huggingface.co/datasets/patrickvonplaten/images/resolve/main/protogen/rswf5qk9be9a1.jpg)
108
+
109
+ ## PENDING DATA FOR MERGE, RPGv2 not accounted..
110
+ ## Checkpoint Merging Data Reference
111
+
112
+ <style>
113
+ .myTable {
114
+ border-collapse:collapse;
115
+ }
116
+ .myTable th {
117
+ background-color:#663380;
118
+ color:white;
119
+ }
120
+ .myTable td, .myTable th {
121
+ padding:5px;
122
+ border:1px solid #663380;
123
+ }
124
+ </style>
125
+ <table class="myTable">
126
+ <tr>
127
+ <th>Models</th>
128
+ <th>Protogen v2.2 (Anime)</th>
129
+ <th>Protogen x3.4 (Photo)</th>
130
+ <th>Protogen x5.3 (Photo)</th>
131
+ <th>Protogen x5.8 (Sci-fi/Anime)</th>
132
+ <th>Protogen x5.9 (Dragon)</th>
133
+ <th>Protogen x7.4 (Eclipse)</th>
134
+ <th>Protogen x8.0 (Nova)</th>
135
+ <th>Protogen x8.6 (Infinity)</th>
136
+ </tr>
137
+ <tr>
138
+ <td>seek_art_mega v1</td>
139
+ <td>52.50%</td>
140
+ <td>42.76%</td>
141
+ <td>42.63%</td>
142
+ <td></td>
143
+ <td></td>
144
+ <td></td>
145
+ <td>25.21%</td>
146
+ <td>14.83%</td>
147
+ </tr>
148
+ <tr>
149
+ <td>modelshoot v1</td>
150
+ <td>30.00%</td>
151
+ <td>24.44%</td>
152
+ <td>24.37%</td>
153
+ <td>2.56%</td>
154
+ <td>2.05%</td>
155
+ <td>3.48%</td>
156
+ <td>22.91%</td>
157
+ <td>13.48%</td>
158
+ </tr>
159
+ <tr>
160
+ <td>elldreth v1</td>
161
+ <td>12.64%</td>
162
+ <td>10.30%</td>
163
+ <td>10.23%</td>
164
+ <td></td>
165
+ <td></td>
166
+ <td></td>
167
+ <td>6.06%</td>
168
+ <td>3.57%</td>
169
+ </tr>
170
+ <tr>
171
+ <td>photoreal v2</td>
172
+ <td></td>
173
+ <td></td>
174
+ <td>10.00%</td>
175
+ <td>48.64%</td>
176
+ <td>38.91%</td>
177
+ <td>66.33%</td>
178
+ <td>20.49%</td>
179
+ <td>12.06%</td>
180
+ </tr>
181
+ <tr>
182
+ <td>analogdiffusion v1</td>
183
+ <td></td>
184
+ <td>4.75%</td>
185
+ <td>4.50%</td>
186
+ <td></td>
187
+ <td></td>
188
+ <td></td>
189
+ <td>1.75%</td>
190
+ <td>1.03%</td>
191
+ </tr>
192
+ <tr>
193
+ <td>openjourney v2</td>
194
+ <td></td>
195
+ <td>4.51%</td>
196
+ <td>4.28%</td>
197
+ <td></td>
198
+ <td></td>
199
+ <td>4.75%</td>
200
+ <td>2.26%</td>
201
+ <td>1.33%</td>
202
+ </tr>
203
+ <tr>
204
+ <td>hassan1.4</td>
205
+ <td>2.63%</td>
206
+ <td>2.14%</td>
207
+ <td>2.13%</td>
208
+ <td></td>
209
+ <td></td>
210
+ <td></td>
211
+ <td>1.26%</td>
212
+ <td>0.74%</td>
213
+ </tr>
214
+ <tr>
215
+ <td>f222</td>
216
+ <td>2.23%</td>
217
+ <td>1.82%</td>
218
+ <td>1.81%</td>
219
+ <td></td>
220
+ <td></td>
221
+ <td></td>
222
+ <td>1.07%</td>
223
+ <td>0.63%</td>
224
+ </tr>
225
+ <tr>
226
+ <td>hasdx</td>
227
+ <td></td>
228
+ <td></td>
229
+ <td></td>
230
+ <td>20.00%</td>
231
+ <td>16.00%</td>
232
+ <td>4.07%</td>
233
+ <td>5.01%</td>
234
+ <td>2.95%</td>
235
+ </tr>
236
+ <tr>
237
+ <td>moistmix</td>
238
+ <td></td>
239
+ <td></td>
240
+ <td></td>
241
+ <td>16.00%</td>
242
+ <td>12.80%</td>
243
+ <td>3.86%</td>
244
+ <td>4.08%</td>
245
+ <td>2.40%</td>
246
+ </tr>
247
+ <tr>
248
+ <td>roboDiffusion v1</td>
249
+ <td></td>
250
+ <td>4.29%</td>
251
+ <td></td>
252
+ <td>12.80%</td>
253
+ <td>10.24%</td>
254
+ <td>3.67%</td>
255
+ <td>4.41%</td>
256
+ <td>2.60%</td>
257
+ </tr>
258
+ <tr>
259
+ <td>RPG v3</td>
260
+ <td></td>
261
+ <td>5.00%</td>
262
+ <td></td>
263
+ <td></td>
264
+ <td>20.00%</td>
265
+ <td>4.29%</td>
266
+ <td>4.29%</td>
267
+ <td>2.52%</td>
268
+ </tr>
269
+ <tr>
270
+ <td>anything&everything</td>
271
+ <td></td>
272
+ <td></td>
273
+ <td></td>
274
+ <td></td>
275
+ <td></td>
276
+ <td>4.51%</td>
277
+ <td>0.56%</td>
278
+ <td>0.33%</td>
279
+ </tr>
280
+ <tr>
281
+ <td>dreamlikediff v1</td>
282
+ <td></td>
283
+ <td></td>
284
+ <td></td>
285
+ <td></td>
286
+ <td></td>
287
+ <td>5.0%</td>
288
+ <td>0.63%</td>
289
+ <td>0.37%</td>
290
+ </tr>
291
+ <tr>
292
+ <td>sci-fidiff v1</td>
293
+ <td></td>
294
+ <td></td>
295
+ <td></td>
296
+ <td></td>
297
+ <td></td>
298
+ <td></td>
299
+ <td></td>
300
+ <td>3.10%</td>
301
+ </tr>
302
+ <tr>
303
+ <td>synthwavepunk v2</td>
304
+ <td></td>
305
+ <td></td>
306
+ <td></td>
307
+ <td></td>
308
+ <td></td>
309
+ <td></td>
310
+ <td></td>
311
+ <td>3.26%</td>
312
+ </tr>
313
+ <tr>
314
+ <td>mashupv2</td>
315
+ <td></td>
316
+ <td></td>
317
+ <td></td>
318
+ <td></td>
319
+ <td></td>
320
+ <td></td>
321
+ <td></td>
322
+ <td>11.51%</td>
323
+ </tr>
324
+ <tr>
325
+ <td>dreamshaper 252</td>
326
+ <td></td>
327
+ <td></td>
328
+ <td></td>
329
+ <td></td>
330
+ <td></td>
331
+ <td></td>
332
+ <td></td>
333
+ <td>4.04%</td>
334
+ </tr>
335
+ <tr>
336
+ <td>comicdiff v2</td>
337
+ <td></td>
338
+ <td></td>
339
+ <td></td>
340
+ <td></td>
341
+ <td></td>
342
+ <td></td>
343
+ <td></td>
344
+ <td>4.25%</td>
345
+ </tr>
346
+ <tr>
347
+ <td>artEros</td>
348
+ <td></td>
349
+ <td></td>
350
+ <td></td>
351
+ <td></td>
352
+ <td></td>
353
+ <td></td>
354
+ <td></td>
355
+ <td>15.00%</td>
356
+ </tr>
357
+ </table>
358
+
359
+ ## License
360
+
361
+ By downloading you agree to the terms of these licenses
362
+
363
+ <a href="https://huggingface.co/spaces/CompVis/stable-diffusion-license">CreativeML Open RAIL-M</a>
364
+
365
+ <a href="https://huggingface.co/coreco/seek.art_MEGA/blob/main/LICENSE.txt">Seek Art Mega License</a>
feature_extractor/preprocessor_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": {
3
+ "height": 224,
4
+ "width": 224
5
+ },
6
+ "do_center_crop": true,
7
+ "do_convert_rgb": true,
8
+ "do_normalize": true,
9
+ "do_rescale": true,
10
+ "do_resize": true,
11
+ "feature_extractor_type": "CLIPFeatureExtractor",
12
+ "image_mean": [
13
+ 0.48145466,
14
+ 0.4578275,
15
+ 0.40821073
16
+ ],
17
+ "image_processor_type": "CLIPFeatureExtractor",
18
+ "image_std": [
19
+ 0.26862954,
20
+ 0.26130258,
21
+ 0.27577711
22
+ ],
23
+ "resample": 3,
24
+ "rescale_factor": 0.00392156862745098,
25
+ "size": {
26
+ "shortest_edge": 224
27
+ }
28
+ }
model_index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_class_name": "StableDiffusionPipeline", "_diffusers_version": "0.12.0.dev0", "feature_extractor": ["transformers", "CLIPImageProcessor"], "requires_safety_checker": false, "safety_checker": [null, null], "scheduler": ["diffusers", "PNDMScheduler"], "text_encoder": ["transformers", "CLIPTextModel"], "tokenizer": ["transformers", "CLIPTokenizer"], "unet": ["diffusers", "UNet2DConditionModel"], "vae": ["diffusers", "AutoencoderKL"]}
safety_checker/config.json ADDED
@@ -0,0 +1,181 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_commit_hash": "cb41f3a270d63d454d385fc2e4f571c487c253c5",
3
+ "_name_or_path": "CompVis/stable-diffusion-safety-checker",
4
+ "architectures": [
5
+ "StableDiffusionSafetyChecker"
6
+ ],
7
+ "initializer_factor": 1.0,
8
+ "logit_scale_init_value": 2.6592,
9
+ "model_type": "clip",
10
+ "projection_dim": 768,
11
+ "text_config": {
12
+ "_name_or_path": "",
13
+ "add_cross_attention": false,
14
+ "architectures": null,
15
+ "attention_dropout": 0.0,
16
+ "bad_words_ids": null,
17
+ "begin_suppress_tokens": null,
18
+ "bos_token_id": 0,
19
+ "chunk_size_feed_forward": 0,
20
+ "cross_attention_hidden_size": null,
21
+ "decoder_start_token_id": null,
22
+ "diversity_penalty": 0.0,
23
+ "do_sample": false,
24
+ "dropout": 0.0,
25
+ "early_stopping": false,
26
+ "encoder_no_repeat_ngram_size": 0,
27
+ "eos_token_id": 2,
28
+ "exponential_decay_length_penalty": null,
29
+ "finetuning_task": null,
30
+ "forced_bos_token_id": null,
31
+ "forced_eos_token_id": null,
32
+ "hidden_act": "quick_gelu",
33
+ "hidden_size": 768,
34
+ "id2label": {
35
+ "0": "LABEL_0",
36
+ "1": "LABEL_1"
37
+ },
38
+ "initializer_factor": 1.0,
39
+ "initializer_range": 0.02,
40
+ "intermediate_size": 3072,
41
+ "is_decoder": false,
42
+ "is_encoder_decoder": false,
43
+ "label2id": {
44
+ "LABEL_0": 0,
45
+ "LABEL_1": 1
46
+ },
47
+ "layer_norm_eps": 1e-05,
48
+ "length_penalty": 1.0,
49
+ "max_length": 20,
50
+ "max_position_embeddings": 77,
51
+ "min_length": 0,
52
+ "model_type": "clip_text_model",
53
+ "no_repeat_ngram_size": 0,
54
+ "num_attention_heads": 12,
55
+ "num_beam_groups": 1,
56
+ "num_beams": 1,
57
+ "num_hidden_layers": 12,
58
+ "num_return_sequences": 1,
59
+ "output_attentions": false,
60
+ "output_hidden_states": false,
61
+ "output_scores": false,
62
+ "pad_token_id": 1,
63
+ "prefix": null,
64
+ "problem_type": null,
65
+ "projection_dim": 512,
66
+ "pruned_heads": {},
67
+ "remove_invalid_values": false,
68
+ "repetition_penalty": 1.0,
69
+ "return_dict": true,
70
+ "return_dict_in_generate": false,
71
+ "sep_token_id": null,
72
+ "suppress_tokens": null,
73
+ "task_specific_params": null,
74
+ "temperature": 1.0,
75
+ "tf_legacy_loss": false,
76
+ "tie_encoder_decoder": false,
77
+ "tie_word_embeddings": true,
78
+ "tokenizer_class": null,
79
+ "top_k": 50,
80
+ "top_p": 1.0,
81
+ "torch_dtype": null,
82
+ "torchscript": false,
83
+ "transformers_version": "4.26.0.dev0",
84
+ "typical_p": 1.0,
85
+ "use_bfloat16": false,
86
+ "vocab_size": 49408
87
+ },
88
+ "text_config_dict": {
89
+ "hidden_size": 768,
90
+ "intermediate_size": 3072,
91
+ "num_attention_heads": 12,
92
+ "num_hidden_layers": 12
93
+ },
94
+ "torch_dtype": "float32",
95
+ "transformers_version": null,
96
+ "vision_config": {
97
+ "_name_or_path": "",
98
+ "add_cross_attention": false,
99
+ "architectures": null,
100
+ "attention_dropout": 0.0,
101
+ "bad_words_ids": null,
102
+ "begin_suppress_tokens": null,
103
+ "bos_token_id": null,
104
+ "chunk_size_feed_forward": 0,
105
+ "cross_attention_hidden_size": null,
106
+ "decoder_start_token_id": null,
107
+ "diversity_penalty": 0.0,
108
+ "do_sample": false,
109
+ "dropout": 0.0,
110
+ "early_stopping": false,
111
+ "encoder_no_repeat_ngram_size": 0,
112
+ "eos_token_id": null,
113
+ "exponential_decay_length_penalty": null,
114
+ "finetuning_task": null,
115
+ "forced_bos_token_id": null,
116
+ "forced_eos_token_id": null,
117
+ "hidden_act": "quick_gelu",
118
+ "hidden_size": 1024,
119
+ "id2label": {
120
+ "0": "LABEL_0",
121
+ "1": "LABEL_1"
122
+ },
123
+ "image_size": 224,
124
+ "initializer_factor": 1.0,
125
+ "initializer_range": 0.02,
126
+ "intermediate_size": 4096,
127
+ "is_decoder": false,
128
+ "is_encoder_decoder": false,
129
+ "label2id": {
130
+ "LABEL_0": 0,
131
+ "LABEL_1": 1
132
+ },
133
+ "layer_norm_eps": 1e-05,
134
+ "length_penalty": 1.0,
135
+ "max_length": 20,
136
+ "min_length": 0,
137
+ "model_type": "clip_vision_model",
138
+ "no_repeat_ngram_size": 0,
139
+ "num_attention_heads": 16,
140
+ "num_beam_groups": 1,
141
+ "num_beams": 1,
142
+ "num_channels": 3,
143
+ "num_hidden_layers": 24,
144
+ "num_return_sequences": 1,
145
+ "output_attentions": false,
146
+ "output_hidden_states": false,
147
+ "output_scores": false,
148
+ "pad_token_id": null,
149
+ "patch_size": 14,
150
+ "prefix": null,
151
+ "problem_type": null,
152
+ "projection_dim": 512,
153
+ "pruned_heads": {},
154
+ "remove_invalid_values": false,
155
+ "repetition_penalty": 1.0,
156
+ "return_dict": true,
157
+ "return_dict_in_generate": false,
158
+ "sep_token_id": null,
159
+ "suppress_tokens": null,
160
+ "task_specific_params": null,
161
+ "temperature": 1.0,
162
+ "tf_legacy_loss": false,
163
+ "tie_encoder_decoder": false,
164
+ "tie_word_embeddings": true,
165
+ "tokenizer_class": null,
166
+ "top_k": 50,
167
+ "top_p": 1.0,
168
+ "torch_dtype": null,
169
+ "torchscript": false,
170
+ "transformers_version": "4.26.0.dev0",
171
+ "typical_p": 1.0,
172
+ "use_bfloat16": false
173
+ },
174
+ "vision_config_dict": {
175
+ "hidden_size": 1024,
176
+ "intermediate_size": 4096,
177
+ "num_attention_heads": 16,
178
+ "num_hidden_layers": 24,
179
+ "patch_size": 14
180
+ }
181
+ }
safety_checker/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16d28f2b37109f222cdc33620fdd262102ac32112be0352a7f77e9614b35a394
3
+ size 1216064769
scheduler/scheduler_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "PNDMScheduler",
3
+ "_diffusers_version": "0.12.0.dev0",
4
+ "beta_end": 0.012,
5
+ "beta_schedule": "scaled_linear",
6
+ "beta_start": 0.00085,
7
+ "clip_sample": false,
8
+ "num_train_timesteps": 1000,
9
+ "prediction_type": "epsilon",
10
+ "set_alpha_to_one": false,
11
+ "skip_prk_steps": true,
12
+ "steps_offset": 1,
13
+ "trained_betas": null
14
+ }
text_encoder/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "openai/clip-vit-large-patch14",
3
+ "architectures": [
4
+ "CLIPTextModel"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 0,
8
+ "dropout": 0.0,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "quick_gelu",
11
+ "hidden_size": 768,
12
+ "initializer_factor": 1.0,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 77,
17
+ "model_type": "clip_text_model",
18
+ "num_attention_heads": 12,
19
+ "num_hidden_layers": 12,
20
+ "pad_token_id": 1,
21
+ "projection_dim": 768,
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.26.0.dev0",
24
+ "vocab_size": 49408
25
+ }
text_encoder/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:562a8a1222c3e3f73b802a3c52d866f97a79325a1a3189ec2fe49e5f54bc5a7b
3
+ size 492307041
tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": {
4
+ "__type": "AddedToken",
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false
10
+ },
11
+ "do_lower_case": true,
12
+ "eos_token": {
13
+ "__type": "AddedToken",
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ },
20
+ "errors": "replace",
21
+ "model_max_length": 77,
22
+ "name_or_path": "openai/clip-vit-large-patch14",
23
+ "pad_token": "<|endoftext|>",
24
+ "special_tokens_map_file": "./special_tokens_map.json",
25
+ "tokenizer_class": "CLIPTokenizer",
26
+ "unk_token": {
27
+ "__type": "AddedToken",
28
+ "content": "<|endoftext|>",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false
33
+ }
34
+ }
tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
unet/config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.12.0.dev0",
4
+ "act_fn": "silu",
5
+ "attention_head_dim": 8,
6
+ "block_out_channels": [
7
+ 320,
8
+ 640,
9
+ 1280,
10
+ 1280
11
+ ],
12
+ "center_input_sample": false,
13
+ "class_embed_type": null,
14
+ "cross_attention_dim": 768,
15
+ "down_block_types": [
16
+ "CrossAttnDownBlock2D",
17
+ "CrossAttnDownBlock2D",
18
+ "CrossAttnDownBlock2D",
19
+ "DownBlock2D"
20
+ ],
21
+ "downsample_padding": 1,
22
+ "dual_cross_attention": false,
23
+ "flip_sin_to_cos": true,
24
+ "freq_shift": 0,
25
+ "in_channels": 4,
26
+ "layers_per_block": 2,
27
+ "mid_block_scale_factor": 1,
28
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
29
+ "norm_eps": 1e-05,
30
+ "norm_num_groups": 32,
31
+ "num_class_embeds": null,
32
+ "only_cross_attention": false,
33
+ "out_channels": 4,
34
+ "resnet_time_scale_shift": "default",
35
+ "sample_size": 64,
36
+ "up_block_types": [
37
+ "UpBlock2D",
38
+ "CrossAttnUpBlock2D",
39
+ "CrossAttnUpBlock2D",
40
+ "CrossAttnUpBlock2D"
41
+ ],
42
+ "upcast_attention": false,
43
+ "use_linear_projection": false
44
+ }
unet/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:926c30ee1b8fb52ec8983427e9b2a23ab67ed29fab23ea5eb48c221cc331afbf
3
+ size 3438366373
vae/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.12.0.dev0",
4
+ "act_fn": "silu",
5
+ "block_out_channels": [
6
+ 128,
7
+ 256,
8
+ 512,
9
+ 512
10
+ ],
11
+ "down_block_types": [
12
+ "DownEncoderBlock2D",
13
+ "DownEncoderBlock2D",
14
+ "DownEncoderBlock2D",
15
+ "DownEncoderBlock2D"
16
+ ],
17
+ "in_channels": 3,
18
+ "latent_channels": 4,
19
+ "layers_per_block": 2,
20
+ "norm_num_groups": 32,
21
+ "out_channels": 3,
22
+ "sample_size": 512,
23
+ "scaling_factor": 0.18215,
24
+ "up_block_types": [
25
+ "UpDecoderBlock2D",
26
+ "UpDecoderBlock2D",
27
+ "UpDecoderBlock2D",
28
+ "UpDecoderBlock2D"
29
+ ]
30
+ }
vae/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e9214a656c2445a921065a40861f6adfbe0aa8e0219785e5866f9eef0d5716f
3
+ size 334711857