adhikjoshi commited on
Commit
2af8cf8
1 Parent(s): a83cca5

converted using stablediffusionapi.com

Browse files
README.md CHANGED
@@ -1,39 +1,39 @@
1
  ---
2
  license: creativeml-openrail-m
3
  tags:
4
- - stablediffusionapi.com
5
  - stable-diffusion-api
6
  - text-to-image
7
  - ultra-realistic
8
  pinned: true
9
  ---
10
 
11
- # Rev Anim API Inference
12
 
13
- ![generated from stablediffusionapi.com](https://pub-8b49af329fae499aa563997f5d4068a4.r2.dev/generations/5603613411682351252.png)
14
  ## Get API Key
15
 
16
- Get API key from [Stable Diffusion API](http://stablediffusionapi.com/), No Payment needed.
17
 
18
  Replace Key in below code, change **model_id** to "rev-anim"
19
 
20
- Coding in PHP/Node/Java etc? Have a look at docs for more code examples: [View docs](https://stablediffusionapi.com/docs)
21
 
22
- Model link: [View model](https://stablediffusionapi.com/models/rev-anim)
23
 
24
- Credits: [View credits](https://civitai.com/?query=Rev%20Anim)
25
 
26
- View all models: [View Models](https://stablediffusionapi.com/models)
27
 
28
  import requests
29
  import json
30
 
31
- url = "https://stablediffusionapi.com/api/v3/dreambooth"
32
 
33
  payload = json.dumps({
34
- "key": "",
35
  "model_id": "rev-anim",
36
- "prompt": "actual 8K portrait photo of gareth person, portrait, happy colors, bright eyes, clear eyes, warm smile, smooth soft skin, big dreamy eyes, beautiful intricate colored hair, symmetrical, anime wide eyes, soft lighting, detailed face, by makoto shinkai, stanley artgerm lau, wlop, rossdraws, concept art, digital painting, looking into camera",
37
  "negative_prompt": "painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime",
38
  "width": "512",
39
  "height": "512",
 
1
  ---
2
  license: creativeml-openrail-m
3
  tags:
4
+ - modelslab.com
5
  - stable-diffusion-api
6
  - text-to-image
7
  - ultra-realistic
8
  pinned: true
9
  ---
10
 
11
+ # API Inference
12
 
13
+ ![generated from modelslab.com](https://cdn2.stablediffusionapi.com/generations/bf190b5a-fe19-437c-ba05-82f29cb1f7ad-0.png)
14
  ## Get API Key
15
 
16
+ Get API key from [ModelsLab API](http://modelslab.com), No Payment needed.
17
 
18
  Replace Key in below code, change **model_id** to "rev-anim"
19
 
20
+ Coding in PHP/Node/Java etc? Have a look at docs for more code examples: [View docs](https://modelslab.com/docs)
21
 
22
+ Try model for free: [Generate Images](https://modelslab.com/models/rev-anim)
23
 
24
+ Model link: [View model](https://modelslab.com/models/rev-anim)
25
 
26
+ View all models: [View Models](https://modelslab.com/models)
27
 
28
  import requests
29
  import json
30
 
31
+ url = "https://modelslab.com/api/v6/images/text2img"
32
 
33
  payload = json.dumps({
34
+ "key": "your_api_key",
35
  "model_id": "rev-anim",
36
+ "prompt": "ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K",
37
  "negative_prompt": "painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime",
38
  "width": "512",
39
  "height": "512",
feature_extractor/preprocessor_config.json CHANGED
@@ -24,5 +24,6 @@
24
  "rescale_factor": 0.00392156862745098,
25
  "size": {
26
  "shortest_edge": 224
27
- }
 
28
  }
 
24
  "rescale_factor": 0.00392156862745098,
25
  "size": {
26
  "shortest_edge": 224
27
+ },
28
+ "use_square_size": false
29
  }
model_index.json CHANGED
@@ -1,10 +1,14 @@
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
- "_diffusers_version": "0.16.0.dev0",
4
  "feature_extractor": [
5
  "transformers",
6
  "CLIPFeatureExtractor"
7
  ],
 
 
 
 
8
  "requires_safety_checker": true,
9
  "safety_checker": [
10
  "stable_diffusion",
 
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.25.0.dev0",
4
  "feature_extractor": [
5
  "transformers",
6
  "CLIPFeatureExtractor"
7
  ],
8
+ "image_encoder": [
9
+ null,
10
+ null
11
+ ],
12
  "requires_safety_checker": true,
13
  "safety_checker": [
14
  "stable_diffusion",
safety_checker/config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_commit_hash": "cb41f3a270d63d454d385fc2e4f571c487c253c5",
3
  "_name_or_path": "CompVis/stable-diffusion-safety-checker",
4
  "architectures": [
5
  "StableDiffusionSafetyChecker"
@@ -9,160 +8,21 @@
9
  "model_type": "clip",
10
  "projection_dim": 768,
11
  "text_config": {
12
- "_name_or_path": "",
13
- "add_cross_attention": false,
14
- "architectures": null,
15
- "attention_dropout": 0.0,
16
- "bad_words_ids": null,
17
- "begin_suppress_tokens": null,
18
- "bos_token_id": 0,
19
- "chunk_size_feed_forward": 0,
20
- "cross_attention_hidden_size": null,
21
- "decoder_start_token_id": null,
22
- "diversity_penalty": 0.0,
23
- "do_sample": false,
24
  "dropout": 0.0,
25
- "early_stopping": false,
26
- "encoder_no_repeat_ngram_size": 0,
27
- "eos_token_id": 2,
28
- "exponential_decay_length_penalty": null,
29
- "finetuning_task": null,
30
- "forced_bos_token_id": null,
31
- "forced_eos_token_id": null,
32
- "hidden_act": "quick_gelu",
33
  "hidden_size": 768,
34
- "id2label": {
35
- "0": "LABEL_0",
36
- "1": "LABEL_1"
37
- },
38
- "initializer_factor": 1.0,
39
- "initializer_range": 0.02,
40
  "intermediate_size": 3072,
41
- "is_decoder": false,
42
- "is_encoder_decoder": false,
43
- "label2id": {
44
- "LABEL_0": 0,
45
- "LABEL_1": 1
46
- },
47
- "layer_norm_eps": 1e-05,
48
- "length_penalty": 1.0,
49
- "max_length": 20,
50
- "max_position_embeddings": 77,
51
- "min_length": 0,
52
  "model_type": "clip_text_model",
53
- "no_repeat_ngram_size": 0,
54
- "num_attention_heads": 12,
55
- "num_beam_groups": 1,
56
- "num_beams": 1,
57
- "num_hidden_layers": 12,
58
- "num_return_sequences": 1,
59
- "output_attentions": false,
60
- "output_hidden_states": false,
61
- "output_scores": false,
62
- "pad_token_id": 1,
63
- "prefix": null,
64
- "problem_type": null,
65
- "projection_dim": 512,
66
- "pruned_heads": {},
67
- "remove_invalid_values": false,
68
- "repetition_penalty": 1.0,
69
- "return_dict": true,
70
- "return_dict_in_generate": false,
71
- "sep_token_id": null,
72
- "suppress_tokens": null,
73
- "task_specific_params": null,
74
- "temperature": 1.0,
75
- "tf_legacy_loss": false,
76
- "tie_encoder_decoder": false,
77
- "tie_word_embeddings": true,
78
- "tokenizer_class": null,
79
- "top_k": 50,
80
- "top_p": 1.0,
81
- "torch_dtype": null,
82
- "torchscript": false,
83
- "transformers_version": "4.28.1",
84
- "typical_p": 1.0,
85
- "use_bfloat16": false,
86
- "vocab_size": 49408
87
  },
88
- "torch_dtype": "float32",
89
- "transformers_version": null,
90
  "vision_config": {
91
- "_name_or_path": "",
92
- "add_cross_attention": false,
93
- "architectures": null,
94
- "attention_dropout": 0.0,
95
- "bad_words_ids": null,
96
- "begin_suppress_tokens": null,
97
- "bos_token_id": null,
98
- "chunk_size_feed_forward": 0,
99
- "cross_attention_hidden_size": null,
100
- "decoder_start_token_id": null,
101
- "diversity_penalty": 0.0,
102
- "do_sample": false,
103
  "dropout": 0.0,
104
- "early_stopping": false,
105
- "encoder_no_repeat_ngram_size": 0,
106
- "eos_token_id": null,
107
- "exponential_decay_length_penalty": null,
108
- "finetuning_task": null,
109
- "forced_bos_token_id": null,
110
- "forced_eos_token_id": null,
111
- "hidden_act": "quick_gelu",
112
  "hidden_size": 1024,
113
- "id2label": {
114
- "0": "LABEL_0",
115
- "1": "LABEL_1"
116
- },
117
- "image_size": 224,
118
- "initializer_factor": 1.0,
119
- "initializer_range": 0.02,
120
  "intermediate_size": 4096,
121
- "is_decoder": false,
122
- "is_encoder_decoder": false,
123
- "label2id": {
124
- "LABEL_0": 0,
125
- "LABEL_1": 1
126
- },
127
- "layer_norm_eps": 1e-05,
128
- "length_penalty": 1.0,
129
- "max_length": 20,
130
- "min_length": 0,
131
  "model_type": "clip_vision_model",
132
- "no_repeat_ngram_size": 0,
133
  "num_attention_heads": 16,
134
- "num_beam_groups": 1,
135
- "num_beams": 1,
136
- "num_channels": 3,
137
  "num_hidden_layers": 24,
138
- "num_return_sequences": 1,
139
- "output_attentions": false,
140
- "output_hidden_states": false,
141
- "output_scores": false,
142
- "pad_token_id": null,
143
- "patch_size": 14,
144
- "prefix": null,
145
- "problem_type": null,
146
- "projection_dim": 512,
147
- "pruned_heads": {},
148
- "remove_invalid_values": false,
149
- "repetition_penalty": 1.0,
150
- "return_dict": true,
151
- "return_dict_in_generate": false,
152
- "sep_token_id": null,
153
- "suppress_tokens": null,
154
- "task_specific_params": null,
155
- "temperature": 1.0,
156
- "tf_legacy_loss": false,
157
- "tie_encoder_decoder": false,
158
- "tie_word_embeddings": true,
159
- "tokenizer_class": null,
160
- "top_k": 50,
161
- "top_p": 1.0,
162
- "torch_dtype": null,
163
- "torchscript": false,
164
- "transformers_version": "4.28.1",
165
- "typical_p": 1.0,
166
- "use_bfloat16": false
167
  }
168
  }
 
1
  {
 
2
  "_name_or_path": "CompVis/stable-diffusion-safety-checker",
3
  "architectures": [
4
  "StableDiffusionSafetyChecker"
 
8
  "model_type": "clip",
9
  "projection_dim": 768,
10
  "text_config": {
 
 
 
 
 
 
 
 
 
 
 
 
11
  "dropout": 0.0,
 
 
 
 
 
 
 
 
12
  "hidden_size": 768,
 
 
 
 
 
 
13
  "intermediate_size": 3072,
 
 
 
 
 
 
 
 
 
 
 
14
  "model_type": "clip_text_model",
15
+ "num_attention_heads": 12
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  },
17
+ "torch_dtype": "float16",
18
+ "transformers_version": "4.35.2",
19
  "vision_config": {
 
 
 
 
 
 
 
 
 
 
 
 
20
  "dropout": 0.0,
 
 
 
 
 
 
 
 
21
  "hidden_size": 1024,
 
 
 
 
 
 
 
22
  "intermediate_size": 4096,
 
 
 
 
 
 
 
 
 
 
23
  "model_type": "clip_vision_model",
 
24
  "num_attention_heads": 16,
 
 
 
25
  "num_hidden_layers": 24,
26
+ "patch_size": 14
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27
  }
28
  }
safety_checker/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:16d28f2b37109f222cdc33620fdd262102ac32112be0352a7f77e9614b35a394
3
- size 1216064769
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e09c5d39c29d798137c7aa6f529bd44abe3616b054740a02d896f0381ca4f4f9
3
+ size 608099578
scheduler/scheduler_config.json CHANGED
@@ -1,18 +1,15 @@
1
  {
2
  "_class_name": "PNDMScheduler",
3
- "_diffusers_version": "0.16.0.dev0",
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
7
  "clip_sample": false,
8
- "clip_sample_range": 1.0,
9
- "dynamic_thresholding_ratio": 0.995,
10
  "num_train_timesteps": 1000,
11
  "prediction_type": "epsilon",
12
- "sample_max_value": 1.0,
13
  "set_alpha_to_one": false,
14
  "skip_prk_steps": true,
15
  "steps_offset": 1,
16
- "thresholding": false,
17
  "trained_betas": null
18
  }
 
1
  {
2
  "_class_name": "PNDMScheduler",
3
+ "_diffusers_version": "0.25.0.dev0",
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
7
  "clip_sample": false,
 
 
8
  "num_train_timesteps": 1000,
9
  "prediction_type": "epsilon",
 
10
  "set_alpha_to_one": false,
11
  "skip_prk_steps": true,
12
  "steps_offset": 1,
13
+ "timestep_spacing": "leading",
14
  "trained_betas": null
15
  }
text_encoder/config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_name_or_path": "openai/clip-vit-large-patch14",
3
  "architectures": [
4
  "CLIPTextModel"
5
  ],
@@ -19,7 +18,7 @@
19
  "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
  "projection_dim": 768,
22
- "torch_dtype": "float32",
23
- "transformers_version": "4.28.1",
24
  "vocab_size": 49408
25
  }
 
1
  {
 
2
  "architectures": [
3
  "CLIPTextModel"
4
  ],
 
18
  "num_hidden_layers": 12,
19
  "pad_token_id": 1,
20
  "projection_dim": 768,
21
+ "torch_dtype": "float16",
22
+ "transformers_version": "4.35.2",
23
  "vocab_size": 49408
24
  }
text_encoder/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aad7965eb37e3fc7f0619f382ee4c4d448d76df8d257ad958a48b8a871bea3fd
3
- size 492307041
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f1f74ea903f40a0ac121fbd4bee54f0d4e1021f5afef69ceca44cd0e93f1955
3
+ size 246185562
tokenizer/special_tokens_map.json CHANGED
@@ -13,7 +13,13 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": "<|endoftext|>",
 
 
 
 
 
 
17
  "unk_token": {
18
  "content": "<|endoftext|>",
19
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": {
17
+ "content": "<|endoftext|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
  "unk_token": {
24
  "content": "<|endoftext|>",
25
  "lstrip": false,
tokenizer/tokenizer_config.json CHANGED
@@ -1,33 +1,30 @@
1
  {
2
  "add_prefix_space": false,
3
- "bos_token": {
4
- "__type": "AddedToken",
5
- "content": "<|startoftext|>",
6
- "lstrip": false,
7
- "normalized": true,
8
- "rstrip": false,
9
- "single_word": false
 
 
 
 
 
 
 
 
 
 
10
  },
 
11
  "clean_up_tokenization_spaces": true,
12
  "do_lower_case": true,
13
- "eos_token": {
14
- "__type": "AddedToken",
15
- "content": "<|endoftext|>",
16
- "lstrip": false,
17
- "normalized": true,
18
- "rstrip": false,
19
- "single_word": false
20
- },
21
  "errors": "replace",
22
  "model_max_length": 77,
23
  "pad_token": "<|endoftext|>",
24
  "tokenizer_class": "CLIPTokenizer",
25
- "unk_token": {
26
- "__type": "AddedToken",
27
- "content": "<|endoftext|>",
28
- "lstrip": false,
29
- "normalized": true,
30
- "rstrip": false,
31
- "single_word": false
32
- }
33
  }
 
1
  {
2
  "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "49406": {
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "49407": {
13
+ "content": "<|endoftext|>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ }
20
  },
21
+ "bos_token": "<|startoftext|>",
22
  "clean_up_tokenization_spaces": true,
23
  "do_lower_case": true,
24
+ "eos_token": "<|endoftext|>",
 
 
 
 
 
 
 
25
  "errors": "replace",
26
  "model_max_length": 77,
27
  "pad_token": "<|endoftext|>",
28
  "tokenizer_class": "CLIPTokenizer",
29
+ "unk_token": "<|endoftext|>"
 
 
 
 
 
 
 
30
  }
unet/config.json CHANGED
@@ -1,8 +1,12 @@
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
- "_diffusers_version": "0.16.0.dev0",
4
  "act_fn": "silu",
 
 
 
5
  "attention_head_dim": 8,
 
6
  "block_out_channels": [
7
  320,
8
  640,
@@ -23,8 +27,10 @@
23
  "DownBlock2D"
24
  ],
25
  "downsample_padding": 1,
 
26
  "dual_cross_attention": false,
27
  "encoder_hid_dim": null,
 
28
  "flip_sin_to_cos": true,
29
  "freq_shift": 0,
30
  "in_channels": 4,
@@ -34,6 +40,7 @@
34
  "mid_block_type": "UNetMidBlock2DCrossAttn",
35
  "norm_eps": 1e-05,
36
  "norm_num_groups": 32,
 
37
  "num_class_embeds": null,
38
  "only_cross_attention": false,
39
  "out_channels": 4,
@@ -41,11 +48,14 @@
41
  "resnet_out_scale_factor": 1.0,
42
  "resnet_skip_time_act": false,
43
  "resnet_time_scale_shift": "default",
 
44
  "sample_size": 64,
45
  "time_cond_proj_dim": null,
46
  "time_embedding_act_fn": null,
 
47
  "time_embedding_type": "positional",
48
  "timestep_post_act": null,
 
49
  "up_block_types": [
50
  "UpBlock2D",
51
  "CrossAttnUpBlock2D",
 
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.25.0.dev0",
4
  "act_fn": "silu",
5
+ "addition_embed_type": null,
6
+ "addition_embed_type_num_heads": 64,
7
+ "addition_time_embed_dim": null,
8
  "attention_head_dim": 8,
9
+ "attention_type": "default",
10
  "block_out_channels": [
11
  320,
12
  640,
 
27
  "DownBlock2D"
28
  ],
29
  "downsample_padding": 1,
30
+ "dropout": 0.0,
31
  "dual_cross_attention": false,
32
  "encoder_hid_dim": null,
33
+ "encoder_hid_dim_type": null,
34
  "flip_sin_to_cos": true,
35
  "freq_shift": 0,
36
  "in_channels": 4,
 
40
  "mid_block_type": "UNetMidBlock2DCrossAttn",
41
  "norm_eps": 1e-05,
42
  "norm_num_groups": 32,
43
+ "num_attention_heads": null,
44
  "num_class_embeds": null,
45
  "only_cross_attention": false,
46
  "out_channels": 4,
 
48
  "resnet_out_scale_factor": 1.0,
49
  "resnet_skip_time_act": false,
50
  "resnet_time_scale_shift": "default",
51
+ "reverse_transformer_layers_per_block": null,
52
  "sample_size": 64,
53
  "time_cond_proj_dim": null,
54
  "time_embedding_act_fn": null,
55
+ "time_embedding_dim": null,
56
  "time_embedding_type": "positional",
57
  "timestep_post_act": null,
58
+ "transformer_layers_per_block": 1,
59
  "up_block_types": [
60
  "UpBlock2D",
61
  "CrossAttnUpBlock2D",
unet/diffusion_pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d8e4501ae1ba2c9950a2c06cbdd74a3317a4b4efbce54557163d114945760a0a
3
- size 3438366373
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:202bd5c12326be4cdc91b990d714cd734a9177b0b7d503328e8e91519dc9ab7c
3
+ size 1719324918
vae/config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "AutoencoderKL",
3
- "_diffusers_version": "0.16.0.dev0",
4
  "act_fn": "silu",
5
  "block_out_channels": [
6
  128,
@@ -14,6 +14,7 @@
14
  "DownEncoderBlock2D",
15
  "DownEncoderBlock2D"
16
  ],
 
17
  "in_channels": 3,
18
  "latent_channels": 4,
19
  "layers_per_block": 2,
 
1
  {
2
  "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.25.0.dev0",
4
  "act_fn": "silu",
5
  "block_out_channels": [
6
  128,
 
14
  "DownEncoderBlock2D",
15
  "DownEncoderBlock2D"
16
  ],
17
+ "force_upcast": true,
18
  "in_channels": 3,
19
  "latent_channels": 4,
20
  "layers_per_block": 2,
vae/diffusion_pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c7b7ae67f39449fa9e14d44bbc2010587bc241f47a5507278d0493bdf808b148
3
- size 334711857
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf87681032cdc39493102d540edf55c3ad0e06cddf7b8f7709f1cd96662f1098
3
+ size 167404866