wikeeyang commited on
Commit
2f79c79
1 Parent(s): afcfc42

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -5
README.md CHANGED
@@ -20,20 +20,28 @@ library_name: diffusers
20
 
21
  [Also on CivitAI](https://civitai.com/models/941929)
22
 
23
- 可能是目前快速出图(10步以内) Flux 模型中,出图质量最好、细节最丰富的基础模型。
24
 
25
- May be the Best Quality Step 6-10 Model, In some details, it surpasses the Flux.1 Dev model and approaches the Flux.1 Pro model.
26
 
27
  Based on **[Flux-Fusion-V2](https://huggingface.co/Anibaaal/Flux-Fusion-V2-4step-merge-gguf-nf4/tree/main)**, Merge of **[flux-dev-de-distill](https://huggingface.co/nyanko7/flux-dev-de-distill/tree/main)**, finetuned by **[ComfyUI](https://github.com/comfyanonymous/ComfyUI)**, **[Block_Patcher_ComfyUI](https://github.com/cubiq/Block_Patcher_ComfyUI)**, **[ComfyUI_essentials](https://github.com/cubiq/ComfyUI_essentials)** and other tools.
28
  Recommended 6-10 steps. Greatly improved quality compared to other Flux.1 model.
29
 
30
  ![](./compare.jpg)
31
 
32
- 应网友要求,特提供 GGUF Q8_0 量化版本模型文件,针对该版本模型文件,由于时间关系,未做详细测试,可能出图细节与 fp8 模型略微有细小差异,但应该相差不大,同时,模型文件加载方式和模型转换方法,已在下面附加描述,测试图片效果如下:
33
 
34
- At the request, GGUF Q8_0 quantified version of the model file is provided, for this version of the model file, due to the time relationship, no detailed test has been done, the details of the drawing may be slightly different from the fp8 model, but it should not be much different, at the same time, the model file loading method and model conversion method have been described below.
35
 
36
- ![](./gguf-sample.jpg)
 
 
 
 
 
 
 
 
37
 
38
  # Recommend:
39
 
@@ -66,6 +74,8 @@ https://github.com/city96/ComfyUI-GGUF, Native support GGUF Quantization Model.
66
 
67
  https://github.com/leejet/stable-diffusion.cpp, Provider pure C/C++ GGUF model convert scripts.
68
 
 
 
69
  ## LICENSE
70
 
71
  The weights fall under the [FLUX.1 [dev]](https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md) Non-Commercial License.
 
20
 
21
  [Also on CivitAI](https://civitai.com/models/941929)
22
 
23
+ **可能是目前快速出图(10步以内) Flux 微调模型中,遵循原版 Flux.1 Dev 风格,提示词还原能力强、出图质量最好、出图细节超越 Flux.1 Dev 模型,最接近 Flux.1 Pro 的基础模型。**
24
 
25
+ **May be the Best Quality Step 6-10 Model, In some details, it surpasses the Flux.1 Dev model and approaches the Flux.1 Pro model. and have good ability of prompt following, good of the original Flux.1 Dev style following.**
26
 
27
  Based on **[Flux-Fusion-V2](https://huggingface.co/Anibaaal/Flux-Fusion-V2-4step-merge-gguf-nf4/tree/main)**, Merge of **[flux-dev-de-distill](https://huggingface.co/nyanko7/flux-dev-de-distill/tree/main)**, finetuned by **[ComfyUI](https://github.com/comfyanonymous/ComfyUI)**, **[Block_Patcher_ComfyUI](https://github.com/cubiq/Block_Patcher_ComfyUI)**, **[ComfyUI_essentials](https://github.com/cubiq/ComfyUI_essentials)** and other tools.
28
  Recommended 6-10 steps. Greatly improved quality compared to other Flux.1 model.
29
 
30
  ![](./compare.jpg)
31
 
32
+ GGUF Q8_0 量化版本模型文件,经过几天的测试,没问题。其出图质量与 fp8 相同,在某些细节体现方面,略优于 fp8 格式,GGUF 模型文件加载方式和模型转换方法,已在下面附加描述。另:应网友要求,正在准备 GGUF Q4_1 格式模型,初步测试情况如下图:
33
 
34
+ GGUF Q8_0 quantized model file, had tested, no problem. The quality of the image is the same as fp8 format, and in some details, it is slightly better than the fp8 format, and the loading method of GGUF model and conversion method are described below. Another: At the request, the GGUF Q4_1 format model is being prepared, and the preliminary test situation is as follows:.
35
 
36
+ ![](./Q8_0&Q4_1.jpg)
37
+
38
+ 初步测试结果看:Q4_1 模型在人物等较大对象的输出时,表现较好,但在需要体现更多细节时,发现细节丢失较多,因此,建议网友们如果硬件允许,尽量选择 fp8/Q8_0 模型,Q4_1 模型经测试完成后会尽快上传,NF4格式由于已经失去支持,所以不会提供。
39
+
40
+ According to the preliminary test results, the Q4_1 model performs well in the output of larger objects such as characters, but when it is necessary more details, it is found that more details are lost, so it is recommended try to choose the FP8/Q8_0 model if the hardware allows, and the Q4_1 model will be uploaded as soon as possible. The NF4 format will not be available as it has deprecated.
41
+
42
+ 过度量化将失去本高精细模型的优势,所以,除 Q4_1 外,将不再提供别的量化版本,如有需要,朋友们可根据下面提示信息,自己下载 fp8 后量化。
43
+
44
+ Over-quantization will lose the advantages of this high-precision model, so in addition to Q4_1, no other quantization will be provided, if necessary, you can download FP8 and quantizate it according to the following tips.
45
 
46
  # Recommend:
47
 
 
74
 
75
  https://github.com/leejet/stable-diffusion.cpp, Provider pure C/C++ GGUF model convert scripts.
76
 
77
+ Attn: For easy convert to GGUF Q5/Q4, you can use https://github.com/ruSauron/to-gguf-bat script, download it and put to the same directory with sd.exe file, then just pull my fp8.safetensors model file to bat file in exploer, will pop a CMD windows, and follow the menu to conver the one you want.
78
+
79
  ## LICENSE
80
 
81
  The weights fall under the [FLUX.1 [dev]](https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md) Non-Commercial License.