|
--- |
|
license: apache-2.0 |
|
tags: |
|
- Text-to-Image |
|
- ControlNet |
|
- Diffusers |
|
- Stable Diffusion |
|
--- |
|
|
|
# **ControlNet++: All-in-one ControlNet for image generations and editing!** |
|
 |
|
|
|
## Network Arichitecture |
|
 |
|
|
|
## Advantages about the model |
|
- Use bucket training like novelai, can generate high resolutions images of any aspect ratio |
|
- Use large amount of high quality data(over 10000000 images), the dataset covers a diversity of situation |
|
- Use re-captioned prompt like DALLE.3, use CogVLM to generate detailed description, good prompt following ability |
|
- Use many useful tricks during training. Including but not limited to date augmentation, mutiple loss, multi resolution |
|
- Use almost the same parameter compared with original ControlNet. No obvious increase in network parameter or computation. |
|
- Support 10+ control conditions, no obvious performance drop on any single condition compared with training independently |
|
- Support multi condition generation, condition fusion is learned during training. No need to set hyperparameter or design prompts. |
|
- Compatible with other opensource SDXL models, such as BluePencilXL, CounterfeitXL. Compatible with other Lora models. |
|
|
|
|
|
***We design a new architecture that can support 10+ control types in condition text-to-image generation and can generate high resolution images visually comparable with |
|
midjourney***. The network is based on the original ControlNet architecture, we propose two new modules to: 1 Extend the original ControlNet to support different image |
|
conditions using the same network parameter. 2 Support multiple conditions input without increasing computation offload, which is especially important for designers |
|
who want to edit image in detail, different conditions use the same condition encoder, without adding extra computations or parameters. We do thoroughly experiments |
|
on SDXL and achieve superior performance both in control ability and aesthetic score. We release the method and the model to the open source community to make everyone |
|
can enjoy it. |
|
|
|
Inference scripts and more details can found: https://github.com/xinsir6/ControlNetPlus/tree/main |
|
|
|
**If you find it useful, please give me a star, thank you very much** |
|
|
|
|
|
## Visual Examples |
|
### Openpose |
|
 |
|
 |
|
 |
|
 |
|
 |
|
### Depth |
|
 |
|
 |
|
 |
|
 |
|
 |
|
### Canny |
|
 |
|
 |
|
 |
|
 |
|
 |
|
### Lineart |
|
 |
|
 |
|
 |
|
 |
|
 |
|
### AnimeLineart |
|
 |
|
 |
|
 |
|
 |
|
 |
|
### Mlsd |
|
 |
|
 |
|
 |
|
 |
|
 |
|
### Scribble |
|
 |
|
 |
|
 |
|
 |
|
 |
|
### Hed |
|
 |
|
 |
|
 |
|
 |
|
 |
|
### Pidi(Softedge) |
|
 |
|
 |
|
 |
|
 |
|
 |
|
### Teed |
|
 |
|
 |
|
 |
|
 |
|
 |
|
### Segment |
|
 |
|
 |
|
 |
|
 |
|
 |
|
### Normal |
|
 |
|
 |
|
 |
|
 |
|
 |
|
|
|
## Multi Control Visual Examples |
|
### Openpose + Canny |
|
 |
|
 |
|
 |
|
 |
|
 |
|
 |
|
|
|
### Openpose + Depth |
|
 |
|
 |
|
 |
|
 |
|
 |
|
 |
|
|
|
### Openpose + Scribble |
|
 |
|
 |
|
 |
|
 |
|
 |
|
 |
|
|
|
### Openpose + Normal |
|
 |
|
 |
|
 |
|
 |
|
 |
|
 |
|
|
|
### Openpose + Segment |
|
 |
|
 |
|
 |
|
 |
|
 |
|
 |