Add files using upload-large-folder tool
Browse files
README.md
CHANGED
|
@@ -1,22 +1,16 @@
|
|
| 1 |
-
<!-- README Version: v1.1 -->
|
| 2 |
-
|
| 3 |
---
|
| 4 |
license: apache-2.0
|
| 5 |
library_name: diffusers
|
| 6 |
pipeline_tag: text-to-image
|
| 7 |
tags:
|
| 8 |
-
- text-to-image
|
| 9 |
- flux
|
| 10 |
-
-
|
| 11 |
- image-generation
|
| 12 |
- fp16
|
| 13 |
-
- diffusion
|
| 14 |
-
- stable-diffusion
|
| 15 |
-
- ip-adapter
|
| 16 |
-
- style-transfer
|
| 17 |
-
base_model: black-forest-labs/FLUX.1-dev
|
| 18 |
---
|
| 19 |
|
|
|
|
|
|
|
| 20 |
# FLUX.1-dev FP16 Model Repository
|
| 21 |
|
| 22 |
High-quality text-to-image generation model from Black Forest Labs in FP16 precision format. FLUX.1-dev delivers state-of-the-art image synthesis with exceptional prompt adherence, visual quality, and detail preservation.
|
|
@@ -52,12 +46,10 @@ flux-dev-fp16/
|
|
| 52 |
β βββ t5xxl_fp16.safetensors (9.2 GB) # T5-XXL encoder (alternate location)
|
| 53 |
βββ clip_vision/
|
| 54 |
β βββ clip_vision_h.safetensors (1.2 GB) # CLIP vision encoder
|
| 55 |
-
|
| 56 |
-
|
| 57 |
-
βββ ipadapter-flux/
|
| 58 |
-
βββ ip-adapter.bin (5.0 GB) # IP-Adapter for image prompting
|
| 59 |
|
| 60 |
-
Total Repository Size: 72 GB
|
| 61 |
```
|
| 62 |
|
| 63 |
**Model Components**:
|
|
@@ -65,7 +57,6 @@ Total Repository Size: 72 GB
|
|
| 65 |
- **Text Encoders**: CLIP-L, CLIP-G, T5-XXL for advanced text understanding
|
| 66 |
- **Vision Encoder**: CLIP vision model for image understanding capabilities
|
| 67 |
- **VAE**: `flux-vae-bf16.safetensors` (160 MB) - Variational autoencoder for latent/image conversion
|
| 68 |
-
- **IP-Adapter**: `ip-adapter.bin` (5.0 GB) - Image prompt adapter for style transfer and image conditioning
|
| 69 |
|
| 70 |
## Hardware Requirements
|
| 71 |
|
|
@@ -188,45 +179,6 @@ image = pipe(
|
|
| 188 |
image.save("optimized_output.png")
|
| 189 |
```
|
| 190 |
|
| 191 |
-
### IP-Adapter Image Prompting
|
| 192 |
-
|
| 193 |
-
```python
|
| 194 |
-
import torch
|
| 195 |
-
from diffusers import FluxPipeline
|
| 196 |
-
from ip_adapter import IPAdapter
|
| 197 |
-
|
| 198 |
-
# Load FLUX pipeline
|
| 199 |
-
pipe = FluxPipeline.from_single_file(
|
| 200 |
-
"E:/huggingface/flux-dev-fp16/checkpoints/flux/flux1-dev-fp16.safetensors",
|
| 201 |
-
torch_dtype=torch.float16
|
| 202 |
-
)
|
| 203 |
-
pipe.to("cuda")
|
| 204 |
-
|
| 205 |
-
# Load IP-Adapter for image conditioning
|
| 206 |
-
ip_adapter = IPAdapter(
|
| 207 |
-
pipe,
|
| 208 |
-
image_encoder_path="E:/huggingface/flux-dev-fp16/clip_vision",
|
| 209 |
-
ip_ckpt="E:/huggingface/flux-dev-fp16/ipadapter-flux/ip-adapter.bin",
|
| 210 |
-
device="cuda"
|
| 211 |
-
)
|
| 212 |
-
|
| 213 |
-
# Load reference image for style/composition transfer
|
| 214 |
-
reference_image = "reference_style.jpg"
|
| 215 |
-
|
| 216 |
-
# Generate image with text prompt + image reference
|
| 217 |
-
image = ip_adapter.generate(
|
| 218 |
-
pil_image=reference_image,
|
| 219 |
-
prompt="A landscape in the style of the reference image",
|
| 220 |
-
num_inference_steps=50,
|
| 221 |
-
guidance_scale=7.5,
|
| 222 |
-
scale=0.6, # IP-Adapter influence strength (0.0-1.0)
|
| 223 |
-
height=1024,
|
| 224 |
-
width=1024
|
| 225 |
-
)[0]
|
| 226 |
-
|
| 227 |
-
image.save("style_transfer_output.png")
|
| 228 |
-
```
|
| 229 |
-
|
| 230 |
## Model Specifications
|
| 231 |
|
| 232 |
| Specification | Details |
|
|
@@ -247,7 +199,6 @@ image.save("style_transfer_output.png")
|
|
| 247 |
- Multi-aspect ratio generation
|
| 248 |
- Img2img workflows
|
| 249 |
- Inpainting and outpainting
|
| 250 |
-
- IP-Adapter image prompting and style transfer
|
| 251 |
- ControlNet compatibility
|
| 252 |
- LoRA fine-tuning support
|
| 253 |
|
|
@@ -329,5 +280,5 @@ If you use FLUX.1-dev in your research or projects, please cite:
|
|
| 329 |
|
| 330 |
**Model Version**: FLUX.1-dev
|
| 331 |
**Precision**: FP16
|
| 332 |
-
**Repository Version**: v1.
|
| 333 |
-
**Last Updated**: 2025-10-
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
library_name: diffusers
|
| 4 |
pipeline_tag: text-to-image
|
| 5 |
tags:
|
|
|
|
| 6 |
- flux
|
| 7 |
+
- text-to-image
|
| 8 |
- image-generation
|
| 9 |
- fp16
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 10 |
---
|
| 11 |
|
| 12 |
+
<!-- README Version: v1.2 -->
|
| 13 |
+
|
| 14 |
# FLUX.1-dev FP16 Model Repository
|
| 15 |
|
| 16 |
High-quality text-to-image generation model from Black Forest Labs in FP16 precision format. FLUX.1-dev delivers state-of-the-art image synthesis with exceptional prompt adherence, visual quality, and detail preservation.
|
|
|
|
| 46 |
β βββ t5xxl_fp16.safetensors (9.2 GB) # T5-XXL encoder (alternate location)
|
| 47 |
βββ clip_vision/
|
| 48 |
β βββ clip_vision_h.safetensors (1.2 GB) # CLIP vision encoder
|
| 49 |
+
βββ vae/flux/
|
| 50 |
+
βββ flux-vae-bf16.safetensors (160 MB) # VAE decoder in BF16 precision
|
|
|
|
|
|
|
| 51 |
|
| 52 |
+
Total Repository Size: ~72 GB
|
| 53 |
```
|
| 54 |
|
| 55 |
**Model Components**:
|
|
|
|
| 57 |
- **Text Encoders**: CLIP-L, CLIP-G, T5-XXL for advanced text understanding
|
| 58 |
- **Vision Encoder**: CLIP vision model for image understanding capabilities
|
| 59 |
- **VAE**: `flux-vae-bf16.safetensors` (160 MB) - Variational autoencoder for latent/image conversion
|
|
|
|
| 60 |
|
| 61 |
## Hardware Requirements
|
| 62 |
|
|
|
|
| 179 |
image.save("optimized_output.png")
|
| 180 |
```
|
| 181 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 182 |
## Model Specifications
|
| 183 |
|
| 184 |
| Specification | Details |
|
|
|
|
| 199 |
- Multi-aspect ratio generation
|
| 200 |
- Img2img workflows
|
| 201 |
- Inpainting and outpainting
|
|
|
|
| 202 |
- ControlNet compatibility
|
| 203 |
- LoRA fine-tuning support
|
| 204 |
|
|
|
|
| 280 |
|
| 281 |
**Model Version**: FLUX.1-dev
|
| 282 |
**Precision**: FP16
|
| 283 |
+
**Repository Version**: v1.2
|
| 284 |
+
**Last Updated**: 2025-10-14
|