scribbyotx commited on
Commit
8467b7d
1 Parent(s): f3fa5f7

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +170 -0
README.md ADDED
@@ -0,0 +1,170 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: openrail++
3
+ tags:
4
+ - art
5
+ - stable diffusion
6
+ - ControlNet
7
+ - SDXL
8
+ - Diffusion-XL
9
+ pipeline_tag: text-to-image
10
+ ---
11
+ # MistoLine
12
+ ## Control Every Line!
13
+
14
+ ![Intro Image](assets/intro.png)
15
+ [GitHub Repo](https://github.com/TheMistoAI/MistoLine)
16
+
17
+ ## NEWS!!!!! Anyline-preprocessor is released!!!!
18
+ [Anyline Repo](https://github.com/TheMistoAI/ComfyUI-Anyline)
19
+
20
+ **MistoLine: A Versatile and Robust SDXL-ControlNet Model for Adaptable Line Art Conditioning.**
21
+
22
+ MistoLine is an SDXL-ControlNet model that can adapt to any type of line art input, demonstrating high accuracy and excellent stability. It can generate high-quality images (with a short side greater than 1024px) based on user-provided line art of various types, including hand-drawn sketches, different ControlNet line preprocessors, and model-generated outlines. MistoLine eliminates the need to select different ControlNet models for different line preprocessors, as it exhibits strong generalization capabilities across diverse line art conditions.
23
+
24
+ We developed MistoLine by employing a novel line preprocessing algorithm **[Anyline](https://github.com/TheMistoAI/ComfyUI-Anyline)** and retraining the ControlNet model based on the Unet of stabilityai/ stable-diffusion-xl-base-1.0, along with innovations in large model training engineering. MistoLine showcases superior performance across
25
+ different types of line art inputs, surpassing existing ControlNet models in terms of detail restoration, prompt alignment, and stability, particularly in more complex scenarios.
26
+
27
+ MistoLine maintains consistency with the ControlNet architecture released by @lllyasviel, as illustrated in the following schematic diagram:
28
+ ![ControlNet architecture](assets/controlnet_1.png)
29
+ ![ControlNet architecture](assets/controlnet_2.png)
30
+ *reference:https://github.com/lllyasviel/ControlNet*
31
+
32
+ More information about ControlNet can be found in the following references:
33
+ https://github.com/lllyasviel/ControlNet
34
+ https://huggingface.co/docs/diffusers/main/en/api/pipelines/controlnet_sdxl
35
+
36
+ The model is compatible with most SDXL models, except for PlaygroundV2.5, CosXL, and SDXL-Lightning(maybe). It can be used in conjunction with LCM and other ControlNet models.
37
+
38
+ The following usage of this model is not allowed:
39
+ * Violating laws and regulations
40
+ * Harming or exploiting minors
41
+ * Creating and spreading false information
42
+ * Infringing on others' privacy
43
+ * Defaming or harassing others
44
+ * Automated decision-making that harms others' legal rights
45
+ * Discrimination based on social behavior or personal characteristics
46
+ * Exploiting the vulnerabilities of specific groups to mislead their behavior
47
+ * Discrimination based on legally protected characteristics
48
+ * Providing medical advice and diagnostic results
49
+ * Improperly generating and using information for purposes such as law enforcement and immigration
50
+
51
+ If you use or distribute this model for commercial purposes, you must comply with the following conditions:
52
+ 1. Clearly acknowledge the contribution of TheMisto.ai to this model in the documentation, website, or other prominent and visible locations of your product.
53
+ Example: "This product uses the MistoLine-SDXL-ControlNet developed by TheMisto.ai."
54
+ 2. If your product includes about screens, readme files, or other similar display areas, you must include the above attribution information in those areas.
55
+ 3. If your product does not have the aforementioned areas, you must include the attribution information in other reasonable locations within the product to ensure that end-users can notice it.
56
+ 4. You must not imply in any way that TheMisto.ai endorses or promotes your product. The use of the attribution information is solely to indicate the origin of this model.
57
+ If you have any questions about how to provide attribution in specific cases, please contact info@themisto.ai.
58
+
59
+ 署名条款
60
+ 如果您在商业用途中使用或分发本模型,您必须满足以下条件:
61
+ 1. 在产品的文档,网站,或其他主要可见位置,明确提及 TheMisto.ai 对本软件的贡献。
62
+ 示例: "本产品使用了 TheMisto.ai 开发的 MistoLine-SDXL-ControlNet。"
63
+ 2. 如果您的产品包含有关屏幕,说明文件,或其他类似的显示区域,您必须在这些区域中包含上述署名信息。
64
+ 3. 如果您的产品没有上述区域,您必须在产品的其他合理位置包含署名信息,以确保最终用户能够注意到。
65
+ 4. 您不得以任何方式暗示 TheMisto.ai 为您的产品背书或促销。署名信息的使用仅用于表明本模型的来源。
66
+ 如果您对如何在特定情况下提供署名有任何疑问,请联系info@themisto.ai。
67
+
68
+ The model output is not censored and the authors do not endorse the opinions in the generated content. Use at your own risk.
69
+
70
+ ## Apply with Different Line Preprocessors
71
+ ![preprocessors](assets/preprocessors.png)
72
+
73
+ ## Compere with Other Controlnets
74
+ ![comparison](assets/comparison.png)
75
+
76
+ ## Application Examples
77
+
78
+ ### Sketch Rendering
79
+ *The following case only utilized MistoLine as the controlnet:*
80
+ ![Sketch Rendering](assets/sketch_rendering.png)
81
+
82
+ ### Model Rendering
83
+ *The following case only utilized Anyline as the preprocessor and MistoLine as the controlnet.*
84
+ ![Model Rendering](assets/model_rendering.png)
85
+
86
+ ## ComfyUI Recommended Parameters
87
+ ```
88
+ sampler steps:30
89
+ CFG:7.0
90
+ sampler_name:dpmpp_2m_sde
91
+ scheduler:karras
92
+ denoise:0.93
93
+ controlnet_strength:1.0
94
+ stargt_percent:0.0
95
+ end_percent:0.9
96
+ ```
97
+ ## Diffusers pipeline
98
+ Make sure to first install the libraries:
99
+ ```
100
+ pip install accelerate transformers safetensors opencv-python diffusers
101
+ ```
102
+ And then we're ready to go:
103
+ ```
104
+ from diffusers import ControlNetModel, StableDiffusionXLControlNetPipeline, AutoencoderKL
105
+ from diffusers.utils import load_image
106
+ from PIL import Image
107
+ import torch
108
+ import numpy as np
109
+ import cv2
110
+
111
+ prompt = "aerial view, a futuristic research complex in a bright foggy jungle, hard lighting"
112
+ negative_prompt = 'low quality, bad quality, sketches'
113
+
114
+ image = load_image("https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/hf-logo.png")
115
+
116
+ controlnet_conditioning_scale = 0.5
117
+
118
+ controlnet = ControlNetModel.from_pretrained(
119
+ "TheMistoAI/MistoLine",
120
+ torch_dtype=torch.float16,
121
+ variant="fp16",
122
+ )
123
+ vae = AutoencoderKL.from_pretrained("madebyollin/sdxl-vae-fp16-fix", torch_dtype=torch.float16)
124
+ pipe = StableDiffusionXLControlNetPipeline.from_pretrained(
125
+ "stabilityai/stable-diffusion-xl-base-1.0",
126
+ controlnet=controlnet,
127
+ vae=vae,
128
+ torch_dtype=torch.float16,
129
+ )
130
+ pipe.enable_model_cpu_offload()
131
+
132
+ image = np.array(image)
133
+ image = cv2.Canny(image, 100, 200)
134
+ image = image[:, :, None]
135
+ image = np.concatenate([image, image, image], axis=2)
136
+ image = Image.fromarray(image)
137
+
138
+ images = pipe(
139
+ prompt, negative_prompt=negative_prompt, image=image, controlnet_conditioning_scale=controlnet_conditioning_scale,
140
+ ).images
141
+
142
+ images[0].save(f"hug_lab.png")
143
+ ```
144
+
145
+
146
+ ## Checkpoints
147
+ * mistoLine_rank256.safetensors : General usage version, for ComfyUI and AUTOMATIC1111-WebUI.
148
+ * mistoLine_fp16.safetensors : FP16 weights, for ComfyUI and AUTOMATIC1111-WebUI.
149
+
150
+ ## !!!mistoLine_rank256.safetensors better than mistoLine_fp16.safetensors
151
+ ## !!!mistoLine_rank256.safetensors 表现更加出色!!
152
+
153
+ ## ComfyUI Usage
154
+ ![ComfyUI](assets/comfyui.png)
155
+
156
+ ## 中国(大陆地区)便捷下载地址:
157
+ 链接:https://pan.baidu.com/s/1DbZWmGJ40Uzr3Iz9RNBG_w?pwd=8mzs
158
+ 提取码:8mzs
159
+
160
+ ## Citation
161
+ ```
162
+ @misc{
163
+ title={Adding Conditional Control to Text-to-Image Diffusion Models},
164
+ author={Lvmin Zhang, Anyi Rao, Maneesh Agrawala},
165
+ year={2023},
166
+ eprint={2302.05543},
167
+ archivePrefix={arXiv},
168
+ primaryClass={cs.CV}
169
+ }
170
+ ```