FredZhang7 commited on
Commit
166c004
1 Parent(s): f23ec7a

Upload with huggingface_hub

Browse files
feature_extractor/preprocessor_config.json CHANGED
@@ -1,28 +1,28 @@
1
  {
2
- "crop_size": {
3
- "height": 224,
4
- "width": 224
5
- },
6
- "do_center_crop": true,
7
- "do_convert_rgb": true,
8
- "do_normalize": true,
9
- "do_rescale": true,
10
- "do_resize": true,
11
- "feature_extractor_type": "CLIPFeatureExtractor",
12
- "image_mean": [
13
- 0.48145466,
14
- 0.4578275,
15
- 0.40821073
16
- ],
17
- "image_processor_type": "CLIPImageProcessor",
18
- "image_std": [
19
- 0.26862954,
20
- 0.26130258,
21
- 0.27577711
22
- ],
23
- "resample": 3,
24
- "rescale_factor": 0.00392156862745098,
25
- "size": {
26
- "shortest_edge": 224
27
- }
28
- }
 
1
  {
2
+ "crop_size": {
3
+ "height": 224,
4
+ "width": 224
5
+ },
6
+ "do_center_crop": true,
7
+ "do_convert_rgb": true,
8
+ "do_normalize": true,
9
+ "do_rescale": true,
10
+ "do_resize": true,
11
+ "feature_extractor_type": "CLIPFeatureExtractor",
12
+ "image_mean": [
13
+ 0.48145466,
14
+ 0.4578275,
15
+ 0.40821073
16
+ ],
17
+ "image_processor_type": "CLIPImageProcessor",
18
+ "image_std": [
19
+ 0.26862954,
20
+ 0.26130258,
21
+ 0.27577711
22
+ ],
23
+ "resample": 3,
24
+ "rescale_factor": 0.00392156862745098,
25
+ "size": {
26
+ "shortest_edge": 224
27
+ }
28
+ }
model_index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
- "_diffusers_version": "0.10.2",
4
  "feature_extractor": [
5
  null,
6
  null
@@ -12,7 +12,7 @@
12
  ],
13
  "scheduler": [
14
  "diffusers",
15
- "DDIMScheduler"
16
  ],
17
  "text_encoder": [
18
  "transformers",
@@ -30,4 +30,4 @@
30
  "diffusers",
31
  "AutoencoderKL"
32
  ]
33
- }
 
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.11.1",
4
  "feature_extractor": [
5
  null,
6
  null
 
12
  ],
13
  "scheduler": [
14
  "diffusers",
15
+ "PNDMScheduler"
16
  ],
17
  "text_encoder": [
18
  "transformers",
 
30
  "diffusers",
31
  "AutoencoderKL"
32
  ]
33
+ }
scheduler/scheduler_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
- "_class_name": "DDIMScheduler",
3
- "_diffusers_version": "0.10.2",
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
@@ -11,4 +11,4 @@
11
  "skip_prk_steps": true,
12
  "steps_offset": 1,
13
  "trained_betas": null
14
- }
 
1
  {
2
+ "_class_name": "PNDMScheduler",
3
+ "_diffusers_version": "0.11.1",
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
 
11
  "skip_prk_steps": true,
12
  "steps_offset": 1,
13
  "trained_betas": null
14
+ }
text_encoder/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0a0e89c58d325d78fa948f70e2f162941223786783d6e1994d86b6f2b635ef44
3
- size 492307041
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a24f52aaf1c79a0b1264eaf0d21f016ea6f7cbe7ff6927436696e01ddca38f56
3
+ size 492305335
tokenizer/tokenizer_config.json CHANGED
@@ -19,7 +19,7 @@
19
  },
20
  "errors": "replace",
21
  "model_max_length": 77,
22
- "name_or_path": "dreamlike-art/dreamlike-diffusion-1.0",
23
  "pad_token": "<|endoftext|>",
24
  "special_tokens_map_file": "./special_tokens_map.json",
25
  "tokenizer_class": "CLIPTokenizer",
 
19
  },
20
  "errors": "replace",
21
  "model_max_length": 77,
22
+ "name_or_path": "openai/clip-vit-large-patch14",
23
  "pad_token": "<|endoftext|>",
24
  "special_tokens_map_file": "./special_tokens_map.json",
25
  "tokenizer_class": "CLIPTokenizer",
unet/config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
- "_diffusers_version": "0.10.2",
4
  "act_fn": "silu",
5
  "attention_head_dim": 8,
6
  "block_out_channels": [
@@ -10,6 +10,7 @@
10
  1280
11
  ],
12
  "center_input_sample": false,
 
13
  "cross_attention_dim": 768,
14
  "down_block_types": [
15
  "CrossAttnDownBlock2D",
@@ -24,12 +25,14 @@
24
  "in_channels": 4,
25
  "layers_per_block": 2,
26
  "mid_block_scale_factor": 1,
 
27
  "norm_eps": 1e-05,
28
  "norm_num_groups": 32,
29
  "num_class_embeds": null,
30
  "only_cross_attention": false,
31
  "out_channels": 4,
32
- "sample_size": 80,
 
33
  "up_block_types": [
34
  "UpBlock2D",
35
  "CrossAttnUpBlock2D",
@@ -38,4 +41,4 @@
38
  ],
39
  "upcast_attention": false,
40
  "use_linear_projection": false
41
- }
 
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.11.1",
4
  "act_fn": "silu",
5
  "attention_head_dim": 8,
6
  "block_out_channels": [
 
10
  1280
11
  ],
12
  "center_input_sample": false,
13
+ "class_embed_type": null,
14
  "cross_attention_dim": 768,
15
  "down_block_types": [
16
  "CrossAttnDownBlock2D",
 
25
  "in_channels": 4,
26
  "layers_per_block": 2,
27
  "mid_block_scale_factor": 1,
28
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
29
  "norm_eps": 1e-05,
30
  "norm_num_groups": 32,
31
  "num_class_embeds": null,
32
  "only_cross_attention": false,
33
  "out_channels": 4,
34
+ "resnet_time_scale_shift": "default",
35
+ "sample_size": 64,
36
  "up_block_types": [
37
  "UpBlock2D",
38
  "CrossAttnUpBlock2D",
 
41
  ],
42
  "upcast_attention": false,
43
  "use_linear_projection": false
44
+ }
unet/diffusion_pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0bf4e2d66f92e62071443dfe0d258b3d52c9e58bb40a8e730daea97e8fe58f28
3
- size 3438375973
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddb35e6290df5168888b2e6bddd2d5979141ab551fcb1dd9247bd9fe291c69d1
3
+ size 3438354725
vae/config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "AutoencoderKL",
3
- "_diffusers_version": "0.10.2",
4
  "act_fn": "silu",
5
  "block_out_channels": [
6
  128,
@@ -19,11 +19,11 @@
19
  "layers_per_block": 2,
20
  "norm_num_groups": 32,
21
  "out_channels": 3,
22
- "sample_size": 256,
23
  "up_block_types": [
24
  "UpDecoderBlock2D",
25
  "UpDecoderBlock2D",
26
  "UpDecoderBlock2D",
27
  "UpDecoderBlock2D"
28
  ]
29
- }
 
1
  {
2
  "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.11.1",
4
  "act_fn": "silu",
5
  "block_out_channels": [
6
  128,
 
19
  "layers_per_block": 2,
20
  "norm_num_groups": 32,
21
  "out_channels": 3,
22
+ "sample_size": 512,
23
  "up_block_types": [
24
  "UpDecoderBlock2D",
25
  "UpDecoderBlock2D",
26
  "UpDecoderBlock2D",
27
  "UpDecoderBlock2D"
28
  ]
29
+ }
vae/diffusion_pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f6e817ec7c5c6641c0000a1a7fc7893f6625251131f33f7bb034c55e5d949455
3
- size 167404145
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8cf5b49d164db18a485d392b2d9a9b4e3636d70613cb756d2e1bc460dd13161
3
+ size 334707217