wangkanai commited on
Commit
c22a27f
·
verified ·
1 Parent(s): 677953d

Upload folder using huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +104 -3
README.md CHANGED
@@ -1,3 +1,104 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: diffusers
4
+ pipeline_tag: text-to-image
5
+ tags:
6
+ - text-to-image
7
+ - flux
8
+ - flux.1-dev
9
+ - image-generation
10
+ - stable-diffusion
11
+ - fp16
12
+ - full-precision
13
+ base_model: black-forest-labs/FLUX.1-dev
14
+ ---
15
+
16
+ # FLUX.1-dev FP16 Model Collection
17
+
18
+ This repository contains the FP16 (full precision) variant of the FLUX.1-dev text-to-image generation model. This streamlined collection includes only FP16 precision models optimized for quality.
19
+
20
+ ## Model Description
21
+
22
+ FLUX.1-dev is a state-of-the-art text-to-image generation model that produces high-quality images from text prompts. This FP16 collection provides the best quality output with full precision weights.
23
+
24
+ ## Repository Contents
25
+
26
+ **Total Size**: ~72GB
27
+
28
+ ### Diffusion Models
29
+ - `diffusion_models/flux1-dev-fp16.safetensors` (23GB) - Main diffusion model
30
+ - `checkpoints/flux1-dev-fp16.safetensors` (23GB) - Checkpoint format
31
+
32
+ ### Text Encoders
33
+ - `text_encoders/clip_g.safetensors` (1.3GB) - CLIP-G text encoder
34
+ - `text_encoders/clip_l.safetensors` (235MB) - CLIP-L text encoder
35
+ - `text_encoders/clip-vit-large.safetensors` (1.6GB) - CLIP ViT-Large encoder
36
+ - `text_encoders/t5xxl_fp16.safetensors` (9.2GB) - T5-XXL FP16 text encoder
37
+ - `clip/t5xxl_fp16.safetensors` (9.2GB) - T5-XXL alternative path
38
+
39
+ ### Vision Models
40
+ - `clip_vision/clip_vision_h.safetensors` (1.2GB) - CLIP Vision H model
41
+
42
+ ## Hardware Requirements
43
+
44
+ - **VRAM**: 16GB+ recommended for optimal performance
45
+ - **Disk Space**: 72GB
46
+ - **Precision**: FP16 (full precision)
47
+ - **Memory**: 32GB+ system RAM recommended
48
+
49
+ ## Usage
50
+
51
+ ```python
52
+ from diffusers import FluxPipeline
53
+ import torch
54
+
55
+ # Load the FP16 model
56
+ pipe = FluxPipeline.from_pretrained(
57
+ "path/to/flux-dev-fp16",
58
+ torch_dtype=torch.float16
59
+ )
60
+
61
+ pipe.to("cuda")
62
+
63
+ # Generate an image
64
+ image = pipe(
65
+ prompt="a beautiful mountain landscape at sunset",
66
+ num_inference_steps=50,
67
+ guidance_scale=7.5
68
+ ).images[0]
69
+
70
+ image.save("output.png")
71
+ ```
72
+
73
+ ## Model Precision Trade-offs
74
+
75
+ **FP16 (This Collection)**:
76
+ - Best quality output
77
+ - Full precision weights
78
+ - Requires more VRAM (16GB+)
79
+ - Slower inference compared to FP8
80
+ - Recommended for: Quality-focused applications, professional use
81
+
82
+ **Alternatives**:
83
+ - FP8: ~50% smaller, faster inference, minimal quality loss
84
+ - GGUF: Quantized variants for memory-constrained scenarios
85
+
86
+ ## License
87
+
88
+ This model is released under the Apache 2.0 license.
89
+
90
+ ## Citation
91
+
92
+ ```bibtex
93
+ @software{flux1-dev,
94
+ author = {Black Forest Labs},
95
+ title = {FLUX.1-dev},
96
+ year = {2024},
97
+ publisher = {Hugging Face},
98
+ url = {https://huggingface.co/black-forest-labs/FLUX.1-dev}
99
+ }
100
+ ```
101
+
102
+ ## Model Card Contact
103
+
104
+ For questions or issues with this model collection, please refer to the original FLUX.1-dev model card and repository.