wangkanai commited on
Commit
e8413da
Β·
verified Β·
1 Parent(s): 1b86108

Add files using upload-large-folder tool

Browse files
Files changed (1) hide show
  1. README.md +8 -57
README.md CHANGED
@@ -1,22 +1,16 @@
1
- <!-- README Version: v1.1 -->
2
-
3
  ---
4
  license: apache-2.0
5
  library_name: diffusers
6
  pipeline_tag: text-to-image
7
  tags:
8
- - text-to-image
9
  - flux
10
- - flux.1-dev
11
  - image-generation
12
  - fp16
13
- - diffusion
14
- - stable-diffusion
15
- - ip-adapter
16
- - style-transfer
17
- base_model: black-forest-labs/FLUX.1-dev
18
  ---
19
 
 
 
20
  # FLUX.1-dev FP16 Model Repository
21
 
22
  High-quality text-to-image generation model from Black Forest Labs in FP16 precision format. FLUX.1-dev delivers state-of-the-art image synthesis with exceptional prompt adherence, visual quality, and detail preservation.
@@ -52,12 +46,10 @@ flux-dev-fp16/
52
  β”‚ └── t5xxl_fp16.safetensors (9.2 GB) # T5-XXL encoder (alternate location)
53
  β”œβ”€β”€ clip_vision/
54
  β”‚ └── clip_vision_h.safetensors (1.2 GB) # CLIP vision encoder
55
- β”œβ”€β”€ vae/flux/
56
- β”‚ └── flux-vae-bf16.safetensors (160 MB) # VAE decoder in BF16 precision
57
- └── ipadapter-flux/
58
- └── ip-adapter.bin (5.0 GB) # IP-Adapter for image prompting
59
 
60
- Total Repository Size: 72 GB
61
  ```
62
 
63
  **Model Components**:
@@ -65,7 +57,6 @@ Total Repository Size: 72 GB
65
  - **Text Encoders**: CLIP-L, CLIP-G, T5-XXL for advanced text understanding
66
  - **Vision Encoder**: CLIP vision model for image understanding capabilities
67
  - **VAE**: `flux-vae-bf16.safetensors` (160 MB) - Variational autoencoder for latent/image conversion
68
- - **IP-Adapter**: `ip-adapter.bin` (5.0 GB) - Image prompt adapter for style transfer and image conditioning
69
 
70
  ## Hardware Requirements
71
 
@@ -188,45 +179,6 @@ image = pipe(
188
  image.save("optimized_output.png")
189
  ```
190
 
191
- ### IP-Adapter Image Prompting
192
-
193
- ```python
194
- import torch
195
- from diffusers import FluxPipeline
196
- from ip_adapter import IPAdapter
197
-
198
- # Load FLUX pipeline
199
- pipe = FluxPipeline.from_single_file(
200
- "E:/huggingface/flux-dev-fp16/checkpoints/flux/flux1-dev-fp16.safetensors",
201
- torch_dtype=torch.float16
202
- )
203
- pipe.to("cuda")
204
-
205
- # Load IP-Adapter for image conditioning
206
- ip_adapter = IPAdapter(
207
- pipe,
208
- image_encoder_path="E:/huggingface/flux-dev-fp16/clip_vision",
209
- ip_ckpt="E:/huggingface/flux-dev-fp16/ipadapter-flux/ip-adapter.bin",
210
- device="cuda"
211
- )
212
-
213
- # Load reference image for style/composition transfer
214
- reference_image = "reference_style.jpg"
215
-
216
- # Generate image with text prompt + image reference
217
- image = ip_adapter.generate(
218
- pil_image=reference_image,
219
- prompt="A landscape in the style of the reference image",
220
- num_inference_steps=50,
221
- guidance_scale=7.5,
222
- scale=0.6, # IP-Adapter influence strength (0.0-1.0)
223
- height=1024,
224
- width=1024
225
- )[0]
226
-
227
- image.save("style_transfer_output.png")
228
- ```
229
-
230
  ## Model Specifications
231
 
232
  | Specification | Details |
@@ -247,7 +199,6 @@ image.save("style_transfer_output.png")
247
  - Multi-aspect ratio generation
248
  - Img2img workflows
249
  - Inpainting and outpainting
250
- - IP-Adapter image prompting and style transfer
251
  - ControlNet compatibility
252
  - LoRA fine-tuning support
253
 
@@ -329,5 +280,5 @@ If you use FLUX.1-dev in your research or projects, please cite:
329
 
330
  **Model Version**: FLUX.1-dev
331
  **Precision**: FP16
332
- **Repository Version**: v1.0
333
- **Last Updated**: 2025-10-13
 
 
 
1
  ---
2
  license: apache-2.0
3
  library_name: diffusers
4
  pipeline_tag: text-to-image
5
  tags:
 
6
  - flux
7
+ - text-to-image
8
  - image-generation
9
  - fp16
 
 
 
 
 
10
  ---
11
 
12
+ <!-- README Version: v1.2 -->
13
+
14
  # FLUX.1-dev FP16 Model Repository
15
 
16
  High-quality text-to-image generation model from Black Forest Labs in FP16 precision format. FLUX.1-dev delivers state-of-the-art image synthesis with exceptional prompt adherence, visual quality, and detail preservation.
 
46
  β”‚ └── t5xxl_fp16.safetensors (9.2 GB) # T5-XXL encoder (alternate location)
47
  β”œβ”€β”€ clip_vision/
48
  β”‚ └── clip_vision_h.safetensors (1.2 GB) # CLIP vision encoder
49
+ └── vae/flux/
50
+ └── flux-vae-bf16.safetensors (160 MB) # VAE decoder in BF16 precision
 
 
51
 
52
+ Total Repository Size: ~72 GB
53
  ```
54
 
55
  **Model Components**:
 
57
  - **Text Encoders**: CLIP-L, CLIP-G, T5-XXL for advanced text understanding
58
  - **Vision Encoder**: CLIP vision model for image understanding capabilities
59
  - **VAE**: `flux-vae-bf16.safetensors` (160 MB) - Variational autoencoder for latent/image conversion
 
60
 
61
  ## Hardware Requirements
62
 
 
179
  image.save("optimized_output.png")
180
  ```
181
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
182
  ## Model Specifications
183
 
184
  | Specification | Details |
 
199
  - Multi-aspect ratio generation
200
  - Img2img workflows
201
  - Inpainting and outpainting
 
202
  - ControlNet compatibility
203
  - LoRA fine-tuning support
204
 
 
280
 
281
  **Model Version**: FLUX.1-dev
282
  **Precision**: FP16
283
+ **Repository Version**: v1.2
284
+ **Last Updated**: 2025-10-14