zuleo commited on
Commit
879197e
·
1 Parent(s): 074756e

initial commit to filmation model

Browse files
.gitattributes CHANGED
@@ -32,3 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
+ *.png filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,94 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: creativeml-openrail-m
3
+ tags:
4
+ - stable-diffusion
5
+ - stable-diffusion-diffusers
6
+ - text-to-image
7
+ - image-to-image
8
+ - art
9
+ - artistic
10
+ - diffusers
11
+ - filmation
12
+ ---
13
+
14
+ # Filmation MOTU Diffusion
15
+
16
+
17
+ Fine-tuned Stable Diffusion model, based of ```SD 1.5```, trained with art from Masters of the Universe.
18
+
19
+ ![Detailed Samples](https://huggingface.co/zuleo/effeffIX-concept-diffusion/resolve/main/booth1.png)
20
+
21
+ ## Model Usage
22
+
23
+ Use the token ```filmation motu style```
24
+
25
+
26
+ ---
27
+
28
+ ☕ If you enjoy this model, buy us a coffee [![Buy a coffee](https://badgen.net/badge/icon/kofi?icon=kofi&label=buy%20us%20a%20coffee)](https://ko-fi.com/3eegames)
29
+
30
+ ---
31
+
32
+
33
+ ## 🧾 Prompt example:
34
+
35
+ **The MOTU styled woman**
36
+
37
+ ```award winning masterpiece illustration of a woman, filmation motu style, trending on artstation, highly detailed, rendered with Unreal Engine, 8k```
38
+
39
+ [Negative prompt](#❎-negative-prompt-template)
40
+
41
+ _Steps: 40, Sampler: DPM++ 2S a Karras, CFG scale: 7, Seed: 311713104, Size: 512x512, Model hash: c0302f10_
42
+
43
+ ---
44
+
45
+ **Refined He-Man**
46
+
47
+ ```award winning masterpiece illustration of a man, filmation motu style, trending on artstation, highly detailed```
48
+
49
+ [Negative prompt](#❎-negative-prompt-template)
50
+
51
+ _Steps: 40, Sampler: DPM++ 2S a Karras, CFG scale: 7, Seed: 195692676, Size: 512x512, Model hash: c0302f10_
52
+
53
+ ---
54
+
55
+
56
+ ## ❎ Negative Prompt Template
57
+
58
+ All images were rendered with the negative prompt below:
59
+
60
+ ```lowres, bad anatomy, bad hands, text, error, missing fingers, extra digit, fewer digits, cropped, worst quality, low quality, normal quality, jpeg artifacts, signature, watermark, username, blurry, artist name```
61
+
62
+ ## 🧨 Diffusers
63
+
64
+ This model can be used just like any other Stable Diffusion model. For more information,
65
+ please have a look at the [Stable Diffusion](https://huggingface.co/docs/diffusers/api/pipelines/stable_diffusion).
66
+
67
+ Export the model:
68
+ - [ONNX](https://huggingface.co/docs/diffusers/optimization/onnx)
69
+ - [MPS](https://huggingface.co/docs/diffusers/optimization/mps)
70
+ - [FLAX/JAX](https://huggingface.co/blog/stable_diffusion_jax)
71
+
72
+ ```python
73
+ from diffusers import StableDiffusionPipeline
74
+ import torch
75
+
76
+ model_id = "zuleo/filmation-motu"
77
+ pipe = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float16)
78
+ pipe = pipe.to("cuda")
79
+
80
+ prompt = "award winning masterpiece illustration of a man, filmation motu style"
81
+ image = pipe(prompt).images[0]
82
+
83
+ image.save("./filmation_man.png")
84
+ ```
85
+
86
+ ## License
87
+
88
+ This model is open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage.
89
+ The CreativeML OpenRAIL License specifies:
90
+
91
+ - You can't use the model to deliberately produce nor share illegal or harmful outputs or content
92
+ - The authors claims no rights on the outputs you generate, you are free to use them and are accountable for their use which must not go against the provisions set in the license
93
+ - You may re-distribute the weights and use the model commercially and/or as a service. If you do, please be aware you have to include the same use restrictions as the ones in the license and share a copy of the CreativeML OpenRAIL-M to all your users (please read the license entirely and carefully)
94
+ [Please read the full license here](https://huggingface.co/spaces/CompVis/stable-diffusion-license)
booth1.png ADDED

Git LFS Details

  • SHA256: dc14d31291de3c6a10b76322d74a96ce11b8f985bdcdf698f2c59d8f3df4cdb1
  • Pointer size: 131 Bytes
  • Size of remote file: 597 kB
filmationmotu_v1.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72af5b79a2b73d1cf43ff53a1f8074a9bf07beac9f794154de87a9615045babb
3
+ size 4098020782
model_index.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.9.0",
4
+ "feature_extractor": [
5
+ null,
6
+ null
7
+ ],
8
+ "requires_safety_checker": null,
9
+ "safety_checker": [
10
+ null,
11
+ null
12
+ ],
13
+ "scheduler": [
14
+ "diffusers",
15
+ "DDIMScheduler"
16
+ ],
17
+ "text_encoder": [
18
+ "transformers",
19
+ "CLIPTextModel"
20
+ ],
21
+ "tokenizer": [
22
+ "transformers",
23
+ "CLIPTokenizer"
24
+ ],
25
+ "unet": [
26
+ "diffusers",
27
+ "UNet2DConditionModel"
28
+ ],
29
+ "vae": [
30
+ "diffusers",
31
+ "AutoencoderKL"
32
+ ]
33
+ }
scheduler/scheduler_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "DDIMScheduler",
3
+ "_diffusers_version": "0.9.0",
4
+ "beta_end": 0.012,
5
+ "beta_schedule": "scaled_linear",
6
+ "beta_start": 0.00085,
7
+ "clip_sample": false,
8
+ "num_train_timesteps": 1000,
9
+ "prediction_type": "epsilon",
10
+ "set_alpha_to_one": false,
11
+ "steps_offset": 1,
12
+ "trained_betas": null
13
+ }
text_encoder/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "C:\\Users\\ryguy\\Documents\\github\\stable-diffusion-webui\\models\\dreambooth\\filmationmotu\\working",
3
+ "architectures": [
4
+ "CLIPTextModel"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 0,
8
+ "dropout": 0.0,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "quick_gelu",
11
+ "hidden_size": 768,
12
+ "initializer_factor": 1.0,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 77,
17
+ "model_type": "clip_text_model",
18
+ "num_attention_heads": 12,
19
+ "num_hidden_layers": 12,
20
+ "pad_token_id": 1,
21
+ "projection_dim": 768,
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.21.0",
24
+ "vocab_size": 49408
25
+ }
text_encoder/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b43b7fdd3937a0b56131c164a15fccf193f8d112222636cdb730d69be149467
3
+ size 492308087
tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": {
4
+ "__type": "AddedToken",
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false
10
+ },
11
+ "do_lower_case": true,
12
+ "eos_token": {
13
+ "__type": "AddedToken",
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ },
20
+ "errors": "replace",
21
+ "model_max_length": 77,
22
+ "name_or_path": "C:\\Users\\ryguy\\Documents\\github\\stable-diffusion-webui\\models\\dreambooth\\filmationmotu\\working\\tokenizer",
23
+ "pad_token": "<|endoftext|>",
24
+ "special_tokens_map_file": "./special_tokens_map.json",
25
+ "tokenizer_class": "CLIPTokenizer",
26
+ "unk_token": {
27
+ "__type": "AddedToken",
28
+ "content": "<|endoftext|>",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false
33
+ }
34
+ }
tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
unet/config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.9.0",
4
+ "_name_or_path": "C:\\Users\\ryguy\\Documents\\github\\stable-diffusion-webui\\models\\dreambooth\\filmationmotu\\working",
5
+ "act_fn": "silu",
6
+ "attention_head_dim": 8,
7
+ "block_out_channels": [
8
+ 320,
9
+ 640,
10
+ 1280,
11
+ 1280
12
+ ],
13
+ "center_input_sample": false,
14
+ "cross_attention_dim": 768,
15
+ "down_block_types": [
16
+ "CrossAttnDownBlock2D",
17
+ "CrossAttnDownBlock2D",
18
+ "CrossAttnDownBlock2D",
19
+ "DownBlock2D"
20
+ ],
21
+ "downsample_padding": 1,
22
+ "dual_cross_attention": false,
23
+ "flip_sin_to_cos": true,
24
+ "freq_shift": 0,
25
+ "in_channels": 4,
26
+ "layers_per_block": 2,
27
+ "mid_block_scale_factor": 1,
28
+ "norm_eps": 1e-05,
29
+ "norm_num_groups": 32,
30
+ "num_class_embeds": null,
31
+ "only_cross_attention": false,
32
+ "out_channels": 4,
33
+ "sample_size": 64,
34
+ "up_block_types": [
35
+ "UpBlock2D",
36
+ "CrossAttnUpBlock2D",
37
+ "CrossAttnUpBlock2D",
38
+ "CrossAttnUpBlock2D"
39
+ ],
40
+ "use_linear_projection": false
41
+ }
unet/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72cf296ab83ef21967e5b4bf05e159d43ce8da694cb016fe91ef2cd7dfe16e7f
3
+ size 3438364325
vae/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.9.0",
4
+ "_name_or_path": "C:\\Users\\ryguy\\Documents\\github\\stable-diffusion-webui\\models\\dreambooth\\filmationmotu\\working",
5
+ "act_fn": "silu",
6
+ "block_out_channels": [
7
+ 128,
8
+ 256,
9
+ 512,
10
+ 512
11
+ ],
12
+ "down_block_types": [
13
+ "DownEncoderBlock2D",
14
+ "DownEncoderBlock2D",
15
+ "DownEncoderBlock2D",
16
+ "DownEncoderBlock2D"
17
+ ],
18
+ "in_channels": 3,
19
+ "latent_channels": 4,
20
+ "layers_per_block": 2,
21
+ "norm_num_groups": 32,
22
+ "out_channels": 3,
23
+ "sample_size": 256,
24
+ "up_block_types": [
25
+ "UpDecoderBlock2D",
26
+ "UpDecoderBlock2D",
27
+ "UpDecoderBlock2D",
28
+ "UpDecoderBlock2D"
29
+ ]
30
+ }
vae/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cab9e6ef49532bd47bb0dbfa8f6adab999c2ee36a6c6d33791177018bddbe7a
3
+ size 167402961