Upload StableDiffusionINSTDIFFPipeline
Browse files- README.md +1 -0
- model_index.json +0 -1
- text_encoder/config.json +2 -2
- text_encoder/model.fp16.safetensors +3 -0
- unet/config.json +0 -1
- unet/diffusion_pytorch_model.fp16.safetensors +3 -0
- vae/config.json +0 -1
- vae/diffusion_pytorch_model.fp16.safetensors +3 -0
README.md
CHANGED
@@ -5,6 +5,7 @@ language:
|
|
5 |
tags:
|
6 |
- instancediffusion
|
7 |
- layout-to-image
|
|
|
8 |
---
|
9 |
|
10 |
# Diffusers 🧨 port of [InstanceDiffusion: Instance-level Control for Image Generation (CVPR 2024)](https://arxiv.org/abs/2402.03290)
|
|
|
5 |
tags:
|
6 |
- instancediffusion
|
7 |
- layout-to-image
|
8 |
+
library_name: diffusers
|
9 |
---
|
10 |
|
11 |
# Diffusers 🧨 port of [InstanceDiffusion: Instance-level Control for Image Generation (CVPR 2024)](https://arxiv.org/abs/2402.03290)
|
model_index.json
CHANGED
@@ -1,7 +1,6 @@
|
|
1 |
{
|
2 |
"_class_name": "StableDiffusionINSTDIFFPipeline",
|
3 |
"_diffusers_version": "0.32.0.dev0",
|
4 |
-
"_name_or_path": "./instancediffusion_sd15/",
|
5 |
"feature_extractor": [
|
6 |
null,
|
7 |
null
|
|
|
1 |
{
|
2 |
"_class_name": "StableDiffusionINSTDIFFPipeline",
|
3 |
"_diffusers_version": "0.32.0.dev0",
|
|
|
4 |
"feature_extractor": [
|
5 |
null,
|
6 |
null
|
text_encoder/config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"CLIPTextModel"
|
5 |
],
|
@@ -19,7 +19,7 @@
|
|
19 |
"num_hidden_layers": 12,
|
20 |
"pad_token_id": 1,
|
21 |
"projection_dim": 768,
|
22 |
-
"torch_dtype": "
|
23 |
"transformers_version": "4.46.0",
|
24 |
"vocab_size": 49408
|
25 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "openai/clip-vit-large-patch14",
|
3 |
"architectures": [
|
4 |
"CLIPTextModel"
|
5 |
],
|
|
|
19 |
"num_hidden_layers": 12,
|
20 |
"pad_token_id": 1,
|
21 |
"projection_dim": 768,
|
22 |
+
"torch_dtype": "float16",
|
23 |
"transformers_version": "4.46.0",
|
24 |
"vocab_size": 49408
|
25 |
}
|
text_encoder/model.fp16.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:660c6f5b1abae9dc498ac2d21e1347d2abdb0cf6c0c0c8576cd796491d9a6cdd
|
3 |
+
size 246144152
|
unet/config.json
CHANGED
@@ -1,7 +1,6 @@
|
|
1 |
{
|
2 |
"_class_name": "UNet2DConditionModel",
|
3 |
"_diffusers_version": "0.32.0.dev0",
|
4 |
-
"_name_or_path": "./instancediffusion_sd15/unet",
|
5 |
"act_fn": "silu",
|
6 |
"addition_embed_type": null,
|
7 |
"addition_embed_type_num_heads": 64,
|
|
|
1 |
{
|
2 |
"_class_name": "UNet2DConditionModel",
|
3 |
"_diffusers_version": "0.32.0.dev0",
|
|
|
4 |
"act_fn": "silu",
|
5 |
"addition_embed_type": null,
|
6 |
"addition_embed_type_num_heads": 64,
|
unet/diffusion_pytorch_model.fp16.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:26305ecece83dca73cc72801c3e0c364754065a45c40e20cce17e738d20936f6
|
3 |
+
size 2456817634
|
vae/config.json
CHANGED
@@ -1,7 +1,6 @@
|
|
1 |
{
|
2 |
"_class_name": "AutoencoderKL",
|
3 |
"_diffusers_version": "0.32.0.dev0",
|
4 |
-
"_name_or_path": "./instancediffusion_sd15/vae",
|
5 |
"act_fn": "silu",
|
6 |
"block_out_channels": [
|
7 |
128,
|
|
|
1 |
{
|
2 |
"_class_name": "AutoencoderKL",
|
3 |
"_diffusers_version": "0.32.0.dev0",
|
|
|
4 |
"act_fn": "silu",
|
5 |
"block_out_channels": [
|
6 |
128,
|
vae/diffusion_pytorch_model.fp16.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4fbcf0ebe55a0984f5a5e00d8c4521d52359af7229bb4d81890039d2aa16dd7c
|
3 |
+
size 167335342
|