adhikjoshi commited on
Commit
dc69aea
1 Parent(s): 65903a7

converted using stablediffusionapi.com

Browse files
README.md CHANGED
@@ -8,9 +8,9 @@ tags:
8
  pinned: true
9
  ---
10
 
11
- # hassaku-hentai API Inference
12
 
13
- ![generated from stablediffusionapi.com](https://cdn.stablediffusionapi.com/generations/17497589161684967880.png)
14
  ## Get API Key
15
 
16
  Get API key from [Stable Diffusion API](http://stablediffusionapi.com/), No Payment needed.
@@ -19,21 +19,23 @@ Replace Key in below code, change **model_id** to "hassaku-hentai"
19
 
20
  Coding in PHP/Node/Java etc? Have a look at docs for more code examples: [View docs](https://stablediffusionapi.com/docs)
21
 
 
 
22
  Model link: [View model](https://stablediffusionapi.com/models/hassaku-hentai)
23
 
24
- Credits: [View credits](https://civitai.com/?query=hassaku-hentai)
25
 
26
  View all models: [View Models](https://stablediffusionapi.com/models)
27
 
28
  import requests
29
  import json
30
 
31
- url = "https://stablediffusionapi.com/api/v3/dreambooth"
32
 
33
  payload = json.dumps({
34
- "key": "",
35
  "model_id": "hassaku-hentai",
36
- "prompt": "actual 8K portrait photo of gareth person, portrait, happy colors, bright eyes, clear eyes, warm smile, smooth soft skin, big dreamy eyes, beautiful intricate colored hair, symmetrical, anime wide eyes, soft lighting, detailed face, by makoto shinkai, stanley artgerm lau, wlop, rossdraws, concept art, digital painting, looking into camera",
37
  "negative_prompt": "painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime",
38
  "width": "512",
39
  "height": "512",
 
8
  pinned: true
9
  ---
10
 
11
+ # Hassaku (hentai model) API Inference
12
 
13
+ ![generated from stablediffusionapi.com](image)
14
  ## Get API Key
15
 
16
  Get API key from [Stable Diffusion API](http://stablediffusionapi.com/), No Payment needed.
 
19
 
20
  Coding in PHP/Node/Java etc? Have a look at docs for more code examples: [View docs](https://stablediffusionapi.com/docs)
21
 
22
+ Try model for free: [Generate Images](https://stablediffusionapi.com/models/hassaku-hentai)
23
+
24
  Model link: [View model](https://stablediffusionapi.com/models/hassaku-hentai)
25
 
26
+ Credits: [View credits](https://civitai.com/?query=Hassaku%20%28hentai%20model%29)
27
 
28
  View all models: [View Models](https://stablediffusionapi.com/models)
29
 
30
  import requests
31
  import json
32
 
33
+ url = "https://stablediffusionapi.com/api/v4/dreambooth"
34
 
35
  payload = json.dumps({
36
+ "key": "your_api_key",
37
  "model_id": "hassaku-hentai",
38
+ "prompt": "ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K",
39
  "negative_prompt": "painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime",
40
  "width": "512",
41
  "height": "512",
model_index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
- "_diffusers_version": "0.15.1",
4
  "feature_extractor": [
5
  "transformers",
6
  "CLIPFeatureExtractor"
 
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.17.1",
4
  "feature_extractor": [
5
  "transformers",
6
  "CLIPFeatureExtractor"
safety_checker/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c82ebf3724a1fa2a998e47ff5d7278a1970a556cd4efdd0bbf4556c94630f678
3
- size 1216065210
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ba98d19732abb6b1a90fc030f61430678f9b48f5aae79be5de8e005107c8563
3
+ size 1216065214
scheduler/scheduler_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "PNDMScheduler",
3
- "_diffusers_version": "0.15.1",
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
@@ -9,10 +9,12 @@
9
  "dynamic_thresholding_ratio": 0.995,
10
  "num_train_timesteps": 1000,
11
  "prediction_type": "epsilon",
 
12
  "sample_max_value": 1.0,
13
  "set_alpha_to_one": false,
14
  "skip_prk_steps": true,
15
  "steps_offset": 1,
16
  "thresholding": false,
 
17
  "trained_betas": null
18
  }
 
1
  {
2
  "_class_name": "PNDMScheduler",
3
+ "_diffusers_version": "0.17.1",
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
 
9
  "dynamic_thresholding_ratio": 0.995,
10
  "num_train_timesteps": 1000,
11
  "prediction_type": "epsilon",
12
+ "rescale_betas_zero_snr": false,
13
  "sample_max_value": 1.0,
14
  "set_alpha_to_one": false,
15
  "skip_prk_steps": true,
16
  "steps_offset": 1,
17
  "thresholding": false,
18
+ "timestep_spacing": "leading",
19
  "trained_betas": null
20
  }
text_encoder/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e90e5e57d980b90c540c3d8919420f01bde1f9b625270fab27ad7c0dd117fee0
3
- size 492307482
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7f9f7f6f28583d722efe1c5958e3b6e4fcf6adc64275faacca4c79ff5a4c76b
3
+ size 492307486
unet/config.json CHANGED
@@ -1,7 +1,9 @@
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
- "_diffusers_version": "0.15.1",
4
  "act_fn": "silu",
 
 
5
  "attention_head_dim": 8,
6
  "block_out_channels": [
7
  320,
@@ -25,6 +27,7 @@
25
  "downsample_padding": 1,
26
  "dual_cross_attention": false,
27
  "encoder_hid_dim": null,
 
28
  "flip_sin_to_cos": true,
29
  "freq_shift": 0,
30
  "in_channels": 4,
@@ -44,6 +47,7 @@
44
  "sample_size": 64,
45
  "time_cond_proj_dim": null,
46
  "time_embedding_act_fn": null,
 
47
  "time_embedding_type": "positional",
48
  "timestep_post_act": null,
49
  "up_block_types": [
 
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.17.1",
4
  "act_fn": "silu",
5
+ "addition_embed_type": null,
6
+ "addition_embed_type_num_heads": 64,
7
  "attention_head_dim": 8,
8
  "block_out_channels": [
9
  320,
 
27
  "downsample_padding": 1,
28
  "dual_cross_attention": false,
29
  "encoder_hid_dim": null,
30
+ "encoder_hid_dim_type": null,
31
  "flip_sin_to_cos": true,
32
  "freq_shift": 0,
33
  "in_channels": 4,
 
47
  "sample_size": 64,
48
  "time_cond_proj_dim": null,
49
  "time_embedding_act_fn": null,
50
+ "time_embedding_dim": null,
51
  "time_embedding_type": "positional",
52
  "timestep_post_act": null,
53
  "up_block_types": [
unet/diffusion_pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf42ca381e6516377dd0678173eb6f174c7b50fb5be55c145911fabf3757b7df
3
- size 3438366834
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be1ce8558546bdb1e8cf53d1cff61e3c3bdf96a17c18d87775abbbf77f5ba1c4
3
+ size 3438366838
vae/config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "AutoencoderKL",
3
- "_diffusers_version": "0.15.1",
4
  "act_fn": "silu",
5
  "block_out_channels": [
6
  128,
 
1
  {
2
  "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.17.1",
4
  "act_fn": "silu",
5
  "block_out_channels": [
6
  128,
vae/diffusion_pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e0301f1d3a7a1bad075c5eb155226a0940e35940cdca1859e218548115847c9c
3
- size 334712318
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0d02cecc6067b653e619572075d36c21ea58de1868d927d9050b96ea4a57b8c
3
+ size 334712578