Update README.md
Browse files
README.md
CHANGED
@@ -7,4 +7,30 @@ language:
|
|
7 |
- en
|
8 |
- zh
|
9 |
---
|
10 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
- en
|
8 |
- zh
|
9 |
---
|
10 |
+
# Hunyuan-DiT : A Powerful Multi-Resolution Diffusion Transformer with Fine-Grained Chinese Understanding
|
11 |
+
|
12 |
+
This repo contains PyTorch model definitions, pre-trained weights and inference/sampling code for our paper exploring Hunyuan-DiT. You can find more visualizations on our [project page](https://dit.hunyuan.tencent.com/).
|
13 |
+
|
14 |
+
> [**Hunyuan-DiT: A Powerful Multi-Resolution Diffusion Transformer with Fine-Grained Chinese Understanding**](https://arxiv.org/abs/2405.08748) <br>
|
15 |
+
|
16 |
+
> [**DialogGen: Multi-modal Interactive Dialogue System for Multi-turn Text-to-Image Generation**](https://arxiv.org/abs/2403.08857) <br>
|
17 |
+
|
18 |
+
## π₯π₯π₯ News!!
|
19 |
+
* Jul 15, 2024: π HunYuanDiT and Shakker.Ai have jointly launched a fine-tuning event based on the HunYuanDiT 1.2 model. By publishing a lora or fine-tuned model based on HunYuanDiT, you can earn up to $230 bonus from Shakker.Ai. See [Shakker.Ai](https://www.shakker.ai/activitys/shaker-the-world-hunyuan) for more details.
|
20 |
+
* Jul 15, 2024: :tada: Update ComfyUI to support standardized workflows and compatibility with weights from t2i module and Lora training for versions 1.1/1.2, as well as those trained by Kohya or the official script. See [ComfyUI](https://github.com/Tencent/HunyuanDiT/tree/main/comfyui-hydit) for details.
|
21 |
+
* Jul 15, 2024: :zap: We offer Docker environments for CUDA 11/12, allowing you to bypass complex installations and play with a single click! See [dockers](https://github.com/Tencent/HunyuanDiT?tab=readme-ov-file#installation-guide-for-linux) for details.
|
22 |
+
* Jul 08, 2024: :tada: HYDiT-v1.2 version is released. Please check [HunyuanDiT-v1.2](https://huggingface.co/Tencent-Hunyuan/HunyuanDiT-v1.2) and [Distillation-v1.2](https://huggingface.co/Tencent-Hunyuan/Distillation-v1.2) for more details.
|
23 |
+
* Jul 03, 2024: :tada: Kohya-hydit version now available for v1.1 and v1.2 models, with GUI for inference. Official Kohya version is under review. See [kohya](https://github.com/Tencent/HunyuanDiT/tree/main/kohya_ss-hydit) for details.
|
24 |
+
* Jun 27, 2024: :art: Hunyuan-Captioner is released, providing fine-grained caption for training data. See [mllm](https://github.com/Tencent/HunyuanDiT/tree/main/mllm) for details.
|
25 |
+
* Jun 27, 2024: :tada: Support LoRa and ControlNet in diffusers. See [diffusers](https://github.com/Tencent/HunyuanDiT/tree/main/diffusers) for details.
|
26 |
+
* Jun 27, 2024: :tada: 6GB GPU VRAM Inference scripts are released. See [lite](https://github.com/Tencent/HunyuanDiT/tree/main/lite) for details.
|
27 |
+
* Jun 19, 2024: :tada: ControlNet is released, supporting canny, pose and depth control. See [training/inference codes](https://github.com/Tencent/HunyuanDiT?tab=readme-ov-file#controlnet) for details.
|
28 |
+
* Jun 13, 2024: :zap: HYDiT-v1.1 version is released, which mitigates the issue of image oversaturation and alleviates the watermark issue. Please check [HunyuanDiT-v1.1](https://huggingface.co/Tencent-Hunyuan/HunyuanDiT-v1.1) and
|
29 |
+
[Distillation-v1.1](https://huggingface.co/Tencent-Hunyuan/Distillation-v1.1) for more details.
|
30 |
+
* Jun 13, 2024: :truck: The training code is released, offering [full-parameter training](https://github.com/Tencent/HunyuanDiT?tab=readme-ov-file#full-parameter-training) and [LoRA training](https://github.com/Tencent/HunyuanDiT?tab=readme-ov-file#lora).
|
31 |
+
* Jun 06, 2024: :tada: Hunyuan-DiT is now available in ComfyUI. Please check [ComfyUI](https://github.com/Tencent/HunyuanDiT?tab=readme-ov-file#using-comfyui) for more details.
|
32 |
+
* Jun 06, 2024: π We introduce Distillation version for Hunyuan-DiT acceleration, which achieves **50%** acceleration on NVIDIA GPUs. Please check [Distillation](https://huggingface.co/Tencent-Hunyuan/Distillation) for more details.
|
33 |
+
* Jun 05, 2024: π€ Hunyuan-DiT is now available in π€ Diffusers! Please check the [example](https://github.com/Tencent/HunyuanDiT?tab=readme-ov-file#using--diffusers) below.
|
34 |
+
* Jun 04, 2024: :globe_with_meridians: Support Tencent Cloud links to download the pretrained models! Please check the [links](https://github.com/Tencent/HunyuanDiT?tab=readme-ov-file#-download-pretrained-models) below.
|
35 |
+
* May 22, 2024: π We introduce TensorRT version for Hunyuan-DiT acceleration, which achieves **47%** acceleration on NVIDIA GPUs. Please check [TensorRT-libs](https://huggingface.co/Tencent-Hunyuan/TensorRT-libs) for instructions.
|
36 |
+
* May 22, 2024: π¬ We support demo running multi-turn text2image generation now. Please check the [script](https://github.com/Tencent/HunyuanDiT?tab=readme-ov-file#using-gradio) below.
|