chfm
/

chfm SeeSee21 commited on
Commit
1536e1d
·
0 Parent(s):

Duplicate from SeeSee21/Z-Anime

Browse files

Co-authored-by: Sebastian Böhnke <SeeSee21@users.noreply.huggingface.co>

This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +48 -0
  2. README.md +520 -0
  3. aio/z-anime-base-aio-bf16.safetensors +3 -0
  4. aio/z-anime-base-aio-fp8.safetensors +3 -0
  5. aio/z-anime-distill-4step-aio-bf16.safetensors +3 -0
  6. aio/z-anime-distill-4step-aio-fp8.safetensors +3 -0
  7. aio/z-anime-distill-8step-aio-bf16.safetensors +3 -0
  8. aio/z-anime-distill-8step-aio-fp8.safetensors +3 -0
  9. config.json +131 -0
  10. diffusers/model_index.json +24 -0
  11. diffusers/scheduler/scheduler_config.json +18 -0
  12. diffusers/text_encoder/config.json +68 -0
  13. diffusers/text_encoder/generation_config.json +6 -0
  14. diffusers/text_encoder/model.safetensors +3 -0
  15. diffusers/tokenizer/added_tokens.json +28 -0
  16. diffusers/tokenizer/chat_template.jinja +89 -0
  17. diffusers/tokenizer/merges.txt +0 -0
  18. diffusers/tokenizer/special_tokens_map.json +31 -0
  19. diffusers/tokenizer/tokenizer_config.json +239 -0
  20. diffusers/tokenizer/vocab.json +0 -0
  21. diffusers/transformer/config.json +32 -0
  22. diffusers/transformer/diffusion_pytorch_model-00001-of-00002.safetensors +3 -0
  23. diffusers/transformer/diffusion_pytorch_model-00002-of-00002.safetensors +3 -0
  24. diffusers/transformer/diffusion_pytorch_model.safetensors.index.json +528 -0
  25. diffusers/vae/config.json +38 -0
  26. diffusers/vae/diffusion_pytorch_model.safetensors +3 -0
  27. diffusion_models/z-anime-base-bf16.safetensors +3 -0
  28. diffusion_models/z-anime-base-fp8.safetensors +3 -0
  29. diffusion_models/z-anime-distill-4step-bf16.safetensors +3 -0
  30. diffusion_models/z-anime-distill-4step-fp8.safetensors +3 -0
  31. diffusion_models/z-anime-distill-8step-bf16.safetensors +3 -0
  32. diffusion_models/z-anime-distill-8step-fp8.safetensors +3 -0
  33. gguf/z-anime-base-q4_k_s.gguf +3 -0
  34. gguf/z-anime-base-q8_0.gguf +3 -0
  35. images/1.png +3 -0
  36. images/2.png +3 -0
  37. images/3.png +3 -0
  38. images/4.png +3 -0
  39. images/5.png +3 -0
  40. images/6.png +3 -0
  41. images/7.png +3 -0
  42. images/8.png +3 -0
  43. images/9.png +3 -0
  44. images/cover.png +3 -0
  45. images/workflow-cover.png +3 -0
  46. images/workflow-overview.png +0 -0
  47. text_encoder/qwen_3_4b-bf16.safetensors +3 -0
  48. text_encoder/qwen_3_4b-engineer-v4-bf16.safetensors +3 -0
  49. text_encoder/qwen_3_4b-engineer-v4-fp8.safetensors +3 -0
  50. text_encoder/qwen_3_4b-fp8.safetensors +3 -0
.gitattributes ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ images/1.png filter=lfs diff=lfs merge=lfs -text
37
+ images/2.png filter=lfs diff=lfs merge=lfs -text
38
+ images/3.png filter=lfs diff=lfs merge=lfs -text
39
+ images/4.png filter=lfs diff=lfs merge=lfs -text
40
+ images/5.png filter=lfs diff=lfs merge=lfs -text
41
+ images/6.png filter=lfs diff=lfs merge=lfs -text
42
+ images/7.png filter=lfs diff=lfs merge=lfs -text
43
+ images/8.png filter=lfs diff=lfs merge=lfs -text
44
+ images/9.png filter=lfs diff=lfs merge=lfs -text
45
+ images/cover.png filter=lfs diff=lfs merge=lfs -text
46
+ images/workflow-cover.png filter=lfs diff=lfs merge=lfs -text
47
+ gguf/z-anime-base-q4_k_s.gguf filter=lfs diff=lfs merge=lfs -text
48
+ gguf/z-anime-base-q8_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,520 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ tags:
6
+ - text-to-image
7
+ - image-generation
8
+ - diffusion
9
+ - anime
10
+ - z-image
11
+ - z-anime
12
+ - comfyui
13
+ - gguf
14
+ - fp8
15
+ - bf16
16
+ - aio
17
+ pipeline_tag: text-to-image
18
+ library_name: diffusers
19
+ base_model:
20
+ - Tongyi-MAI/Z-Image
21
+ base_model_relation: finetune
22
+ ---
23
+
24
+ # 🎌 Z-Anime | Full Anime Fine-Tune on Z-Image Base
25
+
26
+ <div align="center">
27
+ <img src="images/cover.png" width="380" alt="Z-Anime" />
28
+ </div>
29
+
30
+ <div align="center">
31
+
32
+ **Full Fine-Tune • Rich Aesthetics • Strong Diversity • Full Negative Prompt Support**
33
+ **BF16 & FP8 & GGUF & AIO • Natural Language Prompts • 8GB VRAM**
34
+
35
+ </div>
36
+
37
+ ---
38
+
39
+ ## 🖼️ Preview Gallery
40
+
41
+ <table>
42
+ <tr>
43
+ <td><img src="images/1.png" alt="Z-Anime preview 1" /></td>
44
+ <td><img src="images/2.png" alt="Z-Anime preview 2" /></td>
45
+ <td><img src="images/3.png" alt="Z-Anime preview 3" /></td>
46
+ </tr>
47
+ <tr>
48
+ <td><img src="images/4.png" alt="Z-Anime preview 4" /></td>
49
+ <td><img src="images/5.png" alt="Z-Anime preview 5" /></td>
50
+ <td><img src="images/6.png" alt="Z-Anime preview 6" /></td>
51
+ </tr>
52
+ <tr>
53
+ <td><img src="images/7.png" alt="Z-Anime preview 7" /></td>
54
+ <td><img src="images/8.png" alt="Z-Anime preview 8" /></td>
55
+ <td><img src="images/9.png" alt="Z-Anime preview 9" /></td>
56
+ </tr>
57
+ </table>
58
+
59
+ ---
60
+
61
+ ## ✨ What is Z-Anime?
62
+
63
+ **Z-Anime** is a full fine-tune of Alibaba's **Z-Image Base** architecture — **not a LoRA merge**, but a fully trained anime-focused model family built from the ground up.
64
+
65
+ Built on the **S3-DiT (Single-Stream Diffusion Transformer, 6B parameters)**, Z-Anime inherits the strong foundation of Z-Image Base: rich diversity, strong controllability, full negative prompt support, and a high ceiling for fine-tuning — now adapted for anime-style generation.
66
+
67
+ This repository contains the full **Z-Anime family**:
68
+
69
+ | Variant | Focus | Best For |
70
+ |---|---|---|
71
+ | 🎌 **Z-Anime Base** | Highest quality | Final renders, full control |
72
+ | ⚡ **Z-Anime Distill-8-Step** | Speed + quality balance | Everyday generation |
73
+ | 🚀 **Z-Anime Distill-4-Step** | Maximum speed | Fast iteration, batches |
74
+ | 📦 **GGUF Variants** | Lower memory usage | Low VRAM / CPU / AMD-friendly workflows |
75
+ | 📦 **AIO Variants** | Single-file convenience | Easy ComfyUI setup |
76
+ | 🐍 **Diffusers Folder** | `from_pretrained()` ready | Python pipelines, further fine-tuning |
77
+
78
+ ---
79
+
80
+ ## 🎯 Key Features
81
+
82
+ - ✅ Full fine-tune on Z-Image Base — **not** a LoRA merge
83
+ - ✅ Rich anime aesthetics with strong style diversity
84
+ - ✅ Natural language prompting — works best with descriptive prompts, not tag lists
85
+ - ✅ High diversity across characters, poses, compositions, and layouts
86
+ - ✅ LoRA training ready — strong base for further fine-tuning
87
+ - ✅ Partially NSFW capable
88
+ - ✅ 8GB VRAM compatible
89
+ - ✅ GGUF variants available
90
+ - ✅ AIO variants available (Base, 4-Step, 8-Step)
91
+
92
+ ---
93
+
94
+ ## 🗺️ Z-Anime Roadmap
95
+
96
+ ### ✅ Released
97
+
98
+ #### 🎌 Z-Anime Base
99
+ Full fine-tune on Z-Image Base — **BF16 & FP8**
100
+
101
+ #### ⚡ Z-Anime Distill-8-Step
102
+ **BF16 & FP8** — fast anime generation in **8 steps**, **CFG 1.0**
103
+
104
+ #### 🚀 Z-Anime Distill-4-Step
105
+ **BF16 & FP8** — ultra-fast anime generation in **4 steps**, **CFG 1.0**
106
+
107
+ #### 📦 GGUF Variants
108
+ Available for **low VRAM**, **CPU inference**, and **AMD-friendly** workflows.
109
+
110
+ - **Z-Anime-Base-Q8_0** — Q8_0 quantization (**~6.73 GB**)
111
+ - **Z-Anime-Base-Q4_K_S** — Q4_K_S quantization (**~4.2 GB**)
112
+
113
+ #### 📦 AIO Variants
114
+ All-in-one checkpoints with **image model + VAE + Text Encoder integrated** in a single file.
115
+ Available for **Base**, **Distill-4-Step** and **Distill-8-Step** — each in **BF16 & FP8**.
116
+
117
+ #### 🧩 VAE & Text Encoder
118
+ The required **VAE** (`ae.safetensors`) and **Text Encoder** (`qwen_3_4b.safetensors`) are also included in this repository for users running the standard (non-AIO) variants.
119
+
120
+ #### 🐍 Diffusers Folder
121
+ The full **Diffusers-format folder** (`diffusers/`) is included — drop-in compatible with `ZImagePipeline.from_pretrained()` for Python users who want to run inference outside ComfyUI or use Z-Anime as a starting point for further fine-tuning.
122
+
123
+ More updates coming — follow to stay notified! 🎌
124
+
125
+ ---
126
+
127
+ ## 📦 Versions Overview
128
+
129
+ ### 🟢 BF16 (~12GB)
130
+ Maximum precision. **BFloat16** format with minimal quality compromise. Best for final renders, careful work, and LoRA training.
131
+
132
+ ### 🟡 FP8 (~6GB)
133
+ Recommended for most users. Smaller files, faster downloads, and excellent quality with only minor tradeoffs compared to BF16.
134
+
135
+ ### 🔵 GGUF
136
+ Optimized for lightweight inference setups, especially useful for low VRAM, CPU inference, or alternative backends.
137
+
138
+ ### 🟣 AIO
139
+ All-in-one checkpoints with **image model + Text Encoder + VAE integrated** into a single file for the easiest setup. Available for Base, Distill-4-Step and Distill-8-Step.
140
+
141
+ ---
142
+
143
+ ## 🎌 Z-Anime Base
144
+
145
+ The foundation of the Z-Anime family.
146
+
147
+ A full fine-tune with the **highest quality ceiling**, the **widest creative range**, and **full negative prompt support**.
148
+
149
+ ### Recommended Settings
150
+
151
+ ```yaml
152
+ steps: 28-50
153
+ cfg: 3.0-5.0 # up to 9.0 possible
154
+ sampler: euler_ancestral
155
+ scheduler: beta
156
+ negative_prompt: strongly recommended
157
+ ```
158
+
159
+ ### CFG Guide
160
+
161
+ - **3.0–5.0** → sweet spot for balanced quality and creativity
162
+ - **5.0–7.0** → tighter prompt adherence
163
+ - **7.0–9.0** → maximum control, but watch for oversaturation
164
+ - **Above 9.0** → not recommended
165
+
166
+ Negative prompts have **full effect** on Z-Anime Base and are highly recommended.
167
+
168
+ ---
169
+
170
+ ## ⚡ Z-Anime Distill-8-Step
171
+
172
+ The sweet spot of the family.
173
+
174
+ Distilled from Z-Anime Base, this version delivers strong anime results in just **8 steps** while keeping most of the quality.
175
+
176
+ ### Recommended Settings
177
+
178
+ ```yaml
179
+ steps: 8
180
+ cfg: 1.0 # max ~1.5
181
+ sampler: euler_ancestral
182
+ scheduler: beta
183
+ negative_prompt: limited effect
184
+ ```
185
+
186
+ ### CFG Guide
187
+
188
+ - Best at **CFG 1.0**
189
+ - Small increases to **1.3–1.5** are possible
190
+ - Do **not** go above **1.5** — artifacts may appear
191
+
192
+ Negative prompts have only **limited effect** at this distillation level. If your workflow includes **ConditioningZeroOut**, prefer that instead of a large negative prompt.
193
+
194
+ ---
195
+
196
+ ## 🚀 Z-Anime Distill-4-Step
197
+
198
+ The fastest Z-Anime variant.
199
+
200
+ Built for **maximum throughput** — ideal for rapid prototyping, quick batch generation, and speed-focused workflows.
201
+
202
+ ### Recommended Settings
203
+
204
+ ```yaml
205
+ steps: 4
206
+ cfg: 1.0 # max ~1.5
207
+ sampler: euler_ancestral
208
+ scheduler: beta
209
+ negative_prompt: limited effect
210
+ ```
211
+
212
+ ### Tips for 4-Step
213
+
214
+ - Stay at **CFG 1.0** for the most stable results
215
+ - Put the most important visual details **early** in the prompt
216
+ - An optional upscaler such as hires fix or SeedVR2 can help recover fine detail
217
+
218
+ ---
219
+
220
+ ## 📐 Resolution Guide
221
+
222
+ | Use Case | Resolution |
223
+ |---|---|
224
+ | Portrait / character art | **832 × 1216** |
225
+ | Landscape / scenes / backgrounds | **1216 × 832** |
226
+ | Square / general purpose | **1024 × 1024** |
227
+ | Tall / full body / wallpaper | **768 × 1344** |
228
+ | Cinematic / wide scenes | **1920 × 1088** |
229
+ | Detailed portraits | **1024 × 1536** |
230
+
231
+ **Supported range:** approximately **512 × 512 to 2048 × 2048**, any aspect ratio.
232
+ All main variants are designed to run on **8GB VRAM**.
233
+
234
+ ---
235
+
236
+ ## 💡 Prompting Guide
237
+
238
+ **Natural language works best — not tag lists.**
239
+
240
+ ### ✅ Good
241
+
242
+ ```text
243
+ A young anime girl with long silver hair and golden eyes, wearing a traditional shrine maiden outfit with white haori and red hakama. She stands in a sunlit bamboo forest, cherry blossoms falling softly around her. Warm afternoon light filtering through the trees, detailed fabric shading, expressive face, calm serene expression, high quality anime illustration with fine line work.
244
+ ```
245
+
246
+ ### ❌ Avoid
247
+
248
+ ```text
249
+ anime girl, silver hair, shrine maiden, bamboo, cherry blossom, warm light
250
+ ```
251
+
252
+ ### Character Portraits
253
+
254
+ ```text
255
+ Detailed anime portrait of [character], soft rim lighting, expressive eyes with detailed reflections, fine hair strands, clean linework, professional anime illustration quality.
256
+ ```
257
+
258
+ ### Action Scenes
259
+
260
+ ```text
261
+ Dynamic anime [scene], dramatic angle, motion energy, speed lines, particle effects, cinematic composition, detailed shading, high quality anime art.
262
+ ```
263
+
264
+ ### Backgrounds & Landscapes
265
+
266
+ ```text
267
+ Anime [location] at [time of day], [lighting], [atmosphere], beautiful background art, wallpaper quality, highly detailed environment.
268
+ ```
269
+
270
+ ---
271
+
272
+ ## 🔧 Installation
273
+
274
+ ### Step 1 — Download the version you want
275
+
276
+ Choose between:
277
+
278
+ - **Standard / Distill models** in **BF16** or **FP8** (+ VAE + Text Encoder)
279
+ - **GGUF variants** for low VRAM / CPU / AMD-friendly inference (+ VAE + Text Encoder)
280
+ - **AIO variants** for single-file convenience (no extra VAE / Text Encoder needed)
281
+
282
+ ### Step 2 — Place the files
283
+
284
+ #### Standard BF16 / FP8 models
285
+ ```text
286
+ ComfyUI/models/diffusion_models/
287
+ ├── z-anime-base-bf16.safetensors
288
+ ├── z-anime-base-fp8.safetensors
289
+ ├── z-anime-distill-8step-bf16.safetensors
290
+ ├── z-anime-distill-8step-fp8.safetensors
291
+ ├── z-anime-distill-4step-bf16.safetensors
292
+ └── z-anime-distill-4step-fp8.safetensors
293
+ ```
294
+
295
+ #### GGUF variants
296
+ ```text
297
+ ComfyUI/models/unet/
298
+ ├── z-anime-base-q8_0.gguf
299
+ └── z-anime-base-q4_k_s.gguf
300
+ ```
301
+
302
+ #### Text Encoder
303
+ Two text encoders are included — pick **one**:
304
+
305
+ ```text
306
+ ComfyUI/models/clip/
307
+ └── qwen_3_4b-bf16.safetensors # default (Z-Image standard, BF16)
308
+ or
309
+ └── qwen_3_4b-fp8.safetensors # default (Z-Image standard, FP8)
310
+ or
311
+ └── qwen_3_4b-engineer-v4-bf16.safetensors # alternative (Engineer V4, BF16)
312
+ or
313
+ └── qwen_3_4b-engineer-v4-fp8.safetensors # alternative (Engineer V4, FP8)
314
+ ```
315
+
316
+ - **Default (`qwen_3_4b-*`)** — the standard Z-Image text encoder, repackaged as a single `.safetensors` file (BF16 + FP8). This is what the model was trained against.
317
+ - **Engineer V4 (`qwen_3_4b-engineer-v4-*`)** — an alternative full fine-tune of the Z-Image text encoder by **BennyDaBall**, drop-in compatible. Often produces more varied outputs from the same seed. See *Credits* below for the original repo.
318
+
319
+ #### VAE
320
+ ```text
321
+ ComfyUI/models/vae/
322
+ └── ae.safetensors
323
+ ```
324
+
325
+ #### AIO variants
326
+ For the AIO versions, you only need the single checkpoint file — no extra VAE or Text Encoder required:
327
+
328
+ ```text
329
+ ComfyUI/models/checkpoints/
330
+ ├── z-anime-base-aio-bf16.safetensors
331
+ ├── z-anime-base-aio-fp8.safetensors
332
+ ├── z-anime-distill-8step-aio-bf16.safetensors
333
+ ├── z-anime-distill-8step-aio-fp8.safetensors
334
+ ├── z-anime-distill-4step-aio-bf16.safetensors
335
+ └── z-anime-distill-4step-aio-fp8.safetensors
336
+ ```
337
+
338
+ ### Step 3 — Load in ComfyUI
339
+
340
+ #### For standard BF16 / FP8 versions
341
+ Use:
342
+
343
+ - **Load Diffusion Model** for the model file
344
+ - **CLIP Loader** for the text encoder
345
+ - **VAE Loader** for the VAE
346
+
347
+ #### For GGUF versions
348
+ - Load the **GGUF model from the `models/unet/` folder**
349
+ - Use the same **CLIP** and **VAE** files as above
350
+
351
+ #### For AIO versions
352
+ Use a standard **Checkpoint Loader** — no extra CLIP or VAE loading required.
353
+
354
+ ---
355
+
356
+ ## 📦 Custom Nodes
357
+
358
+ - **rgthree-comfy**
359
+ - **ComfyUI-Lora-Manager**
360
+ - **ComfyUI-GGUF** *(only for the GGUF variants)*
361
+ - **ComfyUI-SeedVR2_VideoUpscaler** *(optional, only for SeedVR2 upscale)*
362
+
363
+ ---
364
+
365
+ ## 🐍 Using the Diffusers Folder
366
+
367
+ For Python users, the full Diffusers-format folder is included under `diffusers/` and can be loaded directly with the `subfolder` argument:
368
+
369
+ ```python
370
+ import torch
371
+ from diffusers import ZImagePipeline
372
+
373
+ pipe = ZImagePipeline.from_pretrained(
374
+ "SeeSee21/Z-Anime",
375
+ subfolder="diffusers",
376
+ torch_dtype=torch.bfloat16,
377
+ ).to("cuda")
378
+
379
+ image = pipe(
380
+ prompt="A young anime girl with long silver hair and golden eyes, "
381
+ "shrine maiden outfit, sunlit bamboo forest, cherry blossoms, "
382
+ "professional anime illustration, fine line work.",
383
+ num_inference_steps=40,
384
+ guidance_scale=4.0,
385
+ ).images[0]
386
+
387
+ image.save("z-anime-output.png")
388
+ ```
389
+
390
+ This format is also a clean starting point for further fine-tuning (LoRA or full fine-tune) with frameworks like **OneTrainer**, **diffusers**, or **kohya-ss**.
391
+
392
+ ---
393
+
394
+ ## 🧩 Official Workflow
395
+
396
+ <div align="center">
397
+ <img src="images/workflow-cover.png" width="380" alt="Z-Anime Workflow" />
398
+ </div>
399
+
400
+ A ready-to-use ComfyUI workflow that supports **all variants** (Base / Distill-8 / Distill-4, BF16 / FP8 / GGUF / AIO) is included in [`workflows/Z-Anime-Workflow-v1.json`](workflows/Z-Anime-Workflow-v1.json).
401
+
402
+ It includes:
403
+ - 📦 Model switch (Diffusion / GGUF / AIO loaders — toggle one at a time)
404
+ - 📖 Optional LoRA loader
405
+ - ✍️ Positive + Negative prompt nodes (with default anime negative)
406
+ - 📐 Resolution presets
407
+ - 🎨 Generate + 🔼 Optional 1.5× upscale with side-by-side compare
408
+ - 📚 Built-in MarkdownNote guide with settings per variant
409
+
410
+ <div align="center">
411
+ <img src="images/workflow-overview.png" alt="Z-Anime Workflow overview" />
412
+ </div>
413
+
414
+ ---
415
+
416
+ ## 📁 Repository Structure
417
+
418
+ ```text
419
+ Z-Anime/
420
+ ├── README.md
421
+ ├── config.json
422
+
423
+ ├── diffusion_models/
424
+ │ ├── z-anime-base-bf16.safetensors
425
+ │ ├── z-anime-base-fp8.safetensors
426
+ │ ├── z-anime-distill-8step-bf16.safetensors
427
+ │ ├── z-anime-distill-8step-fp8.safetensors
428
+ │ ├── z-anime-distill-4step-bf16.safetensors
429
+ │ └── z-anime-distill-4step-fp8.safetensors
430
+
431
+ ├── gguf/
432
+ │ ├── z-anime-base-q8_0.gguf
433
+ │ └── z-anime-base-q4_k_s.gguf
434
+
435
+ ├── aio/
436
+ │ ├── z-anime-base-aio-bf16.safetensors
437
+ │ ├── z-anime-base-aio-fp8.safetensors
438
+ │ ├── z-anime-distill-8step-aio-bf16.safetensors
439
+ │ ├── z-anime-distill-8step-aio-fp8.safetensors
440
+ │ ├── z-anime-distill-4step-aio-bf16.safetensors
441
+ │ └── z-anime-distill-4step-aio-fp8.safetensors
442
+
443
+ ├── text_encoder/
444
+ │ ├── qwen_3_4b-bf16.safetensors # default
445
+ │ ├── qwen_3_4b-fp8.safetensors # default
446
+ │ ├── qwen_3_4b-engineer-v4-bf16.safetensors # alternative (BennyDaBall)
447
+ │ └── qwen_3_4b-engineer-v4-fp8.safetensors # alternative (BennyDaBall)
448
+
449
+ ├── vae/
450
+ │ └── ae.safetensors
451
+
452
+ ├── diffusers/
453
+ │ ├── model_index.json
454
+ │ ├── scheduler/
455
+ │ ├── tokenizer/
456
+ │ ├── text_encoder/
457
+ │ ├── transformer/ (sharded safetensors + index)
458
+ │ └── vae/
459
+
460
+ ├── images/
461
+ │ ├── cover.png
462
+ │ ├── workflow-cover.png
463
+ │ ├── workflow-overview.png
464
+ │ ├── 1.png
465
+ │ ├── 2.png
466
+ │ ├── 3.png
467
+ │ ├── 4.png
468
+ │ ├── 5.png
469
+ │ ├── 6.png
470
+ │ ├── 7.png
471
+ │ ├── 8.png
472
+ │ └── 9.png
473
+ └── workflows/
474
+ └── Z-Anime-Workflow-v1.json
475
+ ```
476
+
477
+ ---
478
+
479
+ ## 📈 Version History
480
+
481
+ ### v1.0 — Initial Release
482
+
483
+ - **Z-Anime Base** released in **BF16 & FP8**
484
+ - **Z-Anime Distill-8-Step** released in **BF16 & FP8**
485
+ - **Z-Anime Distill-4-Step** released in **BF16 & FP8**
486
+ - **GGUF variants added**
487
+ - **Z-Anime-Base-Q8_0** — Q8_0 quantization (~6.73 GB)
488
+ - **Z-Anime-Base-Q4_K_S** — Q4_K_S quantization (~4.2 GB)
489
+ - **AIO variants added** — Base, Distill-4-Step and Distill-8-Step (each in BF16 & FP8)
490
+ - **VAE** (`ae.safetensors`) and **Text Encoder** (`qwen_3_4b.safetensors`) included
491
+ - Optimized for **euler_ancestral**, **euler + beta**, and simple practical use across the family
492
+
493
+ ---
494
+
495
+ ## 🔗 Links
496
+
497
+ - **CivitAI Page:** [civitai.red/models/2483351](https://civitai.red/models/2483351)
498
+ - **Base Model:** [Tongyi-MAI/Z-Image](https://huggingface.co/Tongyi-MAI/Z-Image)
499
+ - **Author:** [SeeSee21 on Hugging Face](https://huggingface.co/SeeSee21)
500
+
501
+ ---
502
+
503
+ ## 🙏 Credits
504
+
505
+ - **Base Architecture:** Tongyi Lab (Alibaba) — Z-Image
506
+ - **Fine-Tune:** SeeSee21
507
+ - **License:** Apache 2.0
508
+ - **Architecture:** S3-DiT (Single-Stream Diffusion Transformer, 6B parameters)
509
+ - **Base Model:** [`Tongyi-MAI/Z-Image`](https://huggingface.co/Tongyi-MAI/Z-Image)
510
+ - **Engineer V4 Text Encoder:** [`BennyDaBall/Qwen3-4b-Z-Image-Engineer-V4`](https://huggingface.co/BennyDaBall/Qwen3-4b-Z-Image-Engineer-V4) — full fine-tune with SMART training, included as alternative text encoder
511
+
512
+ ---
513
+
514
+ ## ❤️ Notes
515
+
516
+ Z-Anime is an experimental anime-focused model family built to explore what a full fine-tune on Z-Image Base can achieve in this space.
517
+
518
+ It is already strong for anime aesthetics, character work, and fast iteration, and future versions will continue to improve diversity, character handling, prompting flexibility, and overall quality.
519
+
520
+ **Z-Anime — anime at its finest, powered by Z-Image Base. 🎌**
aio/z-anime-base-aio-bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90a8b6c67557cee05cccf2d8e596adcbf52ded4754879203d2b275a03f5e9096
3
+ size 20690181512
aio/z-anime-base-aio-fp8.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef778cee4e515ff0c6e930cdcd1315e32ab96380730ae5d21ee3e03f2d414e81
3
+ size 10512806304
aio/z-anime-distill-4step-aio-bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f165264e86d641d9e4c630fc2e624cdae3363aa5c27ad434a72b34cea173c558
3
+ size 20690181536
aio/z-anime-distill-4step-aio-fp8.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5e47fee7d57caf036cf1678a49dce99353f9546e60b09d8da29fd14f85b9a09
3
+ size 10512806320
aio/z-anime-distill-8step-aio-bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2bd85af9db1f587f41775b5c1acc3443d482bc00ce2f8229d4cac27eaba986e6
3
+ size 20690181536
aio/z-anime-distill-8step-aio-fp8.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a342f2651ae36f905de8cc4b110daf0120f0a2ab38ab1ca79d50bca61b61eb60
3
+ size 10512806320
config.json ADDED
@@ -0,0 +1,131 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "z-anime",
3
+ "architecture": "S3-DiT",
4
+ "parameters": "6B",
5
+ "license": "apache-2.0",
6
+ "base_model": "Tongyi-MAI/Z-Image",
7
+ "base_model_relation": "finetune",
8
+ "author": "SeeSee21",
9
+ "pipeline_tag": "text-to-image",
10
+ "prompting": {
11
+ "style": "natural-language",
12
+ "negative_prompt_support": {
13
+ "base": "full",
14
+ "distill_8step": "limited",
15
+ "distill_4step": "limited"
16
+ }
17
+ },
18
+ "variants": {
19
+ "base": {
20
+ "bf16": "diffusion_models/z-anime-base-bf16.safetensors",
21
+ "fp8": "diffusion_models/z-anime-base-fp8.safetensors",
22
+ "aio_bf16": "aio/z-anime-base-aio-bf16.safetensors",
23
+ "aio_fp8": "aio/z-anime-base-aio-fp8.safetensors",
24
+ "recommended_settings": {
25
+ "steps": "28-50",
26
+ "cfg": "3.0-5.0",
27
+ "sampler": "euler_ancestral",
28
+ "scheduler": "beta"
29
+ }
30
+ },
31
+ "distill_8step": {
32
+ "bf16": "diffusion_models/z-anime-distill-8step-bf16.safetensors",
33
+ "fp8": "diffusion_models/z-anime-distill-8step-fp8.safetensors",
34
+ "aio_bf16": "aio/z-anime-distill-8step-aio-bf16.safetensors",
35
+ "aio_fp8": "aio/z-anime-distill-8step-aio-fp8.safetensors",
36
+ "recommended_settings": {
37
+ "steps": 8,
38
+ "cfg": 1.0,
39
+ "sampler": "euler_ancestral",
40
+ "scheduler": "beta"
41
+ }
42
+ },
43
+ "distill_4step": {
44
+ "bf16": "diffusion_models/z-anime-distill-4step-bf16.safetensors",
45
+ "fp8": "diffusion_models/z-anime-distill-4step-fp8.safetensors",
46
+ "aio_bf16": "aio/z-anime-distill-4step-aio-bf16.safetensors",
47
+ "aio_fp8": "aio/z-anime-distill-4step-aio-fp8.safetensors",
48
+ "recommended_settings": {
49
+ "steps": 4,
50
+ "cfg": 1.0,
51
+ "sampler": "euler_ancestral",
52
+ "scheduler": "beta"
53
+ }
54
+ },
55
+ "gguf": {
56
+ "q8_0": {
57
+ "file": "gguf/z-anime-base-q8_0.gguf",
58
+ "description": "Q8_0 quantization",
59
+ "size": "~6.73 GB"
60
+ },
61
+ "q4_k_s": {
62
+ "file": "gguf/z-anime-base-q4_k_s.gguf",
63
+ "description": "Q4_K_S quantization",
64
+ "size": "~4.2 GB"
65
+ }
66
+ }
67
+ },
68
+ "diffusers_folder": {
69
+ "path": "diffusers/",
70
+ "pipeline_class": "ZImagePipeline",
71
+ "usage": "ZImagePipeline.from_pretrained('SeeSee21/Z-Anime', subfolder='diffusers', torch_dtype=torch.bfloat16)",
72
+ "components": [
73
+ "model_index.json",
74
+ "scheduler/",
75
+ "tokenizer/",
76
+ "text_encoder/",
77
+ "transformer/",
78
+ "vae/"
79
+ ]
80
+ },
81
+ "components": {
82
+ "text_encoders": {
83
+ "default": {
84
+ "bf16": "text_encoder/qwen_3_4b-bf16.safetensors",
85
+ "fp8": "text_encoder/qwen_3_4b-fp8.safetensors",
86
+ "description": "Standard Z-Image text encoder, repackaged as a single safetensors",
87
+ "comfyui_path": "ComfyUI/models/clip/"
88
+ },
89
+ "engineer_v4": {
90
+ "bf16": "text_encoder/qwen_3_4b-engineer-v4-bf16.safetensors",
91
+ "fp8": "text_encoder/qwen_3_4b-engineer-v4-fp8.safetensors",
92
+ "description": "Alternative full fine-tune by BennyDaBall (SMART training, more varied outputs)",
93
+ "source": "https://huggingface.co/BennyDaBall/Qwen3-4b-Z-Image-Engineer-V4",
94
+ "comfyui_path": "ComfyUI/models/clip/"
95
+ }
96
+ },
97
+ "vae": {
98
+ "file": "vae/ae.safetensors",
99
+ "description": "Z-Image VAE (slightly trained alongside Z-Anime)",
100
+ "comfyui_path": "ComfyUI/models/vae/"
101
+ }
102
+ },
103
+ "comfyui_paths": {
104
+ "diffusion_models": "ComfyUI/models/diffusion_models/",
105
+ "unet": "ComfyUI/models/unet/",
106
+ "clip": "ComfyUI/models/clip/",
107
+ "vae": "ComfyUI/models/vae/",
108
+ "checkpoints": "ComfyUI/models/checkpoints/"
109
+ },
110
+ "requirements": {
111
+ "custom_nodes": [
112
+ "rgthree-comfy",
113
+ "ComfyUI-Lora-Manager",
114
+ "ComfyUI-SeedVR2_VideoUpscaler (optional)"
115
+ ]
116
+ },
117
+ "supported_vram": "8GB+",
118
+ "links": {
119
+ "civitai": "https://civitai.red/models/2483351",
120
+ "base_model": "https://huggingface.co/Tongyi-MAI/Z-Image",
121
+ "engineer_v4": "https://huggingface.co/BennyDaBall/Qwen3-4b-Z-Image-Engineer-V4",
122
+ "author": "https://huggingface.co/SeeSee21"
123
+ },
124
+ "notes": [
125
+ "BF16 and FP8 are the main release formats.",
126
+ "GGUF variants are intended for lower-memory or alternative inference setups.",
127
+ "AIO variants include the Text Encoder and VAE in a single file (Base, 4-Step, 8-Step).",
128
+ "Two text encoders are included: the standard Z-Image one (default) and BennyDaBall's Engineer V4 (alternative).",
129
+ "The diffusers/ subfolder is a full diffusers-format checkpoint loadable via ZImagePipeline.from_pretrained(repo, subfolder='diffusers')."
130
+ ]
131
+ }
diffusers/model_index.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "ZImagePipeline",
3
+ "_diffusers_version": "0.37.0.dev0",
4
+ "scheduler": [
5
+ "diffusers",
6
+ "FlowMatchEulerDiscreteScheduler"
7
+ ],
8
+ "text_encoder": [
9
+ "transformers",
10
+ "Qwen3ForCausalLM"
11
+ ],
12
+ "tokenizer": [
13
+ "transformers",
14
+ "Qwen2Tokenizer"
15
+ ],
16
+ "transformer": [
17
+ "diffusers",
18
+ "ZImageTransformer2DModel"
19
+ ],
20
+ "vae": [
21
+ "diffusers",
22
+ "AutoencoderKL"
23
+ ]
24
+ }
diffusers/scheduler/scheduler_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "FlowMatchEulerDiscreteScheduler",
3
+ "_diffusers_version": "0.37.0.dev0",
4
+ "base_image_seq_len": 256,
5
+ "base_shift": 0.5,
6
+ "invert_sigmas": false,
7
+ "max_image_seq_len": 4096,
8
+ "max_shift": 1.15,
9
+ "num_train_timesteps": 1000,
10
+ "shift": 6.0,
11
+ "shift_terminal": null,
12
+ "stochastic_sampling": false,
13
+ "time_shift_type": "exponential",
14
+ "use_beta_sigmas": false,
15
+ "use_dynamic_shifting": false,
16
+ "use_exponential_sigmas": false,
17
+ "use_karras_sigmas": false
18
+ }
diffusers/text_encoder/config.json ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "dtype": "bfloat16",
9
+ "eos_token_id": 151645,
10
+ "head_dim": 128,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 2560,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 9728,
15
+ "layer_types": [
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention"
52
+ ],
53
+ "max_position_embeddings": 40960,
54
+ "max_window_layers": 36,
55
+ "model_type": "qwen3",
56
+ "num_attention_heads": 32,
57
+ "num_hidden_layers": 36,
58
+ "num_key_value_heads": 8,
59
+ "rms_norm_eps": 1e-06,
60
+ "rope_scaling": null,
61
+ "rope_theta": 1000000,
62
+ "sliding_window": null,
63
+ "tie_word_embeddings": true,
64
+ "transformers_version": "4.56.2",
65
+ "use_cache": true,
66
+ "use_sliding_window": false,
67
+ "vocab_size": 151936
68
+ }
diffusers/text_encoder/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151643,
4
+ "eos_token_id": 151645,
5
+ "transformers_version": "4.56.2"
6
+ }
diffusers/text_encoder/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f78cf5e0d7ebbad18659d1f7192a4159273a794e51852be4d0a067a1f6f943db
3
+ size 4800649802
diffusers/tokenizer/added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
diffusers/tokenizer/chat_template.jinja ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
18
+ {%- for message in messages[::-1] %}
19
+ {%- set index = (messages|length - 1) - loop.index0 %}
20
+ {%- if ns.multi_step_tool and message.role == "user" and message.content is string and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
21
+ {%- set ns.multi_step_tool = false %}
22
+ {%- set ns.last_query_index = index %}
23
+ {%- endif %}
24
+ {%- endfor %}
25
+ {%- for message in messages %}
26
+ {%- if message.content is string %}
27
+ {%- set content = message.content %}
28
+ {%- else %}
29
+ {%- set content = '' %}
30
+ {%- endif %}
31
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
32
+ {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
33
+ {%- elif message.role == "assistant" %}
34
+ {%- set reasoning_content = '' %}
35
+ {%- if message.reasoning_content is string %}
36
+ {%- set reasoning_content = message.reasoning_content %}
37
+ {%- else %}
38
+ {%- if '</think>' in content %}
39
+ {%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
40
+ {%- set content = content.split('</think>')[-1].lstrip('\n') %}
41
+ {%- endif %}
42
+ {%- endif %}
43
+ {%- if loop.index0 > ns.last_query_index %}
44
+ {%- if loop.last or (not loop.last and reasoning_content) %}
45
+ {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
46
+ {%- else %}
47
+ {{- '<|im_start|>' + message.role + '\n' + content }}
48
+ {%- endif %}
49
+ {%- else %}
50
+ {{- '<|im_start|>' + message.role + '\n' + content }}
51
+ {%- endif %}
52
+ {%- if message.tool_calls %}
53
+ {%- for tool_call in message.tool_calls %}
54
+ {%- if (loop.first and content) or (not loop.first) %}
55
+ {{- '\n' }}
56
+ {%- endif %}
57
+ {%- if tool_call.function %}
58
+ {%- set tool_call = tool_call.function %}
59
+ {%- endif %}
60
+ {{- '<tool_call>\n{"name": "' }}
61
+ {{- tool_call.name }}
62
+ {{- '", "arguments": ' }}
63
+ {%- if tool_call.arguments is string %}
64
+ {{- tool_call.arguments }}
65
+ {%- else %}
66
+ {{- tool_call.arguments | tojson }}
67
+ {%- endif %}
68
+ {{- '}\n</tool_call>' }}
69
+ {%- endfor %}
70
+ {%- endif %}
71
+ {{- '<|im_end|>\n' }}
72
+ {%- elif message.role == "tool" %}
73
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
74
+ {{- '<|im_start|>user' }}
75
+ {%- endif %}
76
+ {{- '\n<tool_response>\n' }}
77
+ {{- content }}
78
+ {{- '\n</tool_response>' }}
79
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
80
+ {{- '<|im_end|>\n' }}
81
+ {%- endif %}
82
+ {%- endif %}
83
+ {%- endfor %}
84
+ {%- if add_generation_prompt %}
85
+ {{- '<|im_start|>assistant\n' }}
86
+ {%- if enable_thinking is defined and enable_thinking is false %}
87
+ {{- '<think>\n\n</think>\n\n' }}
88
+ {%- endif %}
89
+ {%- endif %}
diffusers/tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
diffusers/tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
diffusers/tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,239 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "151665": {
182
+ "content": "<tool_response>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": false
188
+ },
189
+ "151666": {
190
+ "content": "</tool_response>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": false
196
+ },
197
+ "151667": {
198
+ "content": "<think>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": false
204
+ },
205
+ "151668": {
206
+ "content": "</think>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": false
212
+ }
213
+ },
214
+ "additional_special_tokens": [
215
+ "<|im_start|>",
216
+ "<|im_end|>",
217
+ "<|object_ref_start|>",
218
+ "<|object_ref_end|>",
219
+ "<|box_start|>",
220
+ "<|box_end|>",
221
+ "<|quad_start|>",
222
+ "<|quad_end|>",
223
+ "<|vision_start|>",
224
+ "<|vision_end|>",
225
+ "<|vision_pad|>",
226
+ "<|image_pad|>",
227
+ "<|video_pad|>"
228
+ ],
229
+ "bos_token": null,
230
+ "clean_up_tokenization_spaces": false,
231
+ "eos_token": "<|im_end|>",
232
+ "errors": "replace",
233
+ "extra_special_tokens": {},
234
+ "model_max_length": 131072,
235
+ "pad_token": "<|endoftext|>",
236
+ "split_special_tokens": false,
237
+ "tokenizer_class": "Qwen2Tokenizer",
238
+ "unk_token": null
239
+ }
diffusers/tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
diffusers/transformer/config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "ZImageTransformer2DModel",
3
+ "_diffusers_version": "0.37.0.dev0",
4
+ "all_f_patch_size": [
5
+ 1
6
+ ],
7
+ "all_patch_size": [
8
+ 2
9
+ ],
10
+ "axes_dims": [
11
+ 32,
12
+ 48,
13
+ 48
14
+ ],
15
+ "axes_lens": [
16
+ 1536,
17
+ 512,
18
+ 512
19
+ ],
20
+ "cap_feat_dim": 2560,
21
+ "dim": 3840,
22
+ "in_channels": 16,
23
+ "n_heads": 30,
24
+ "n_kv_heads": 30,
25
+ "n_layers": 30,
26
+ "n_refiner_layers": 2,
27
+ "norm_eps": 1e-05,
28
+ "qk_norm": true,
29
+ "rope_theta": 256.0,
30
+ "siglip_feat_dim": null,
31
+ "t_scale": 1000.0
32
+ }
diffusers/transformer/diffusion_pytorch_model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30522af3f6e6cf0cb53638b428a702516cc9dbebc3992d99837ff9125e244d70
3
+ size 9973727144
diffusers/transformer/diffusion_pytorch_model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca4cd5ab296a3ea3ff81226be27c612aeaa591921d91c512459991716f8426aa
3
+ size 2336146728
diffusers/transformer/diffusion_pytorch_model.safetensors.index.json ADDED
@@ -0,0 +1,528 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 12309817472
4
+ },
5
+ "weight_map": {
6
+ "all_final_layer.2-1.adaLN_modulation.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
7
+ "all_final_layer.2-1.adaLN_modulation.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
8
+ "all_final_layer.2-1.linear.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
9
+ "all_final_layer.2-1.linear.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
10
+ "all_x_embedder.2-1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
11
+ "all_x_embedder.2-1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
12
+ "cap_embedder.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
13
+ "cap_embedder.1.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
14
+ "cap_embedder.1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
15
+ "cap_pad_token": "diffusion_pytorch_model-00001-of-00002.safetensors",
16
+ "context_refiner.0.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
17
+ "context_refiner.0.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
18
+ "context_refiner.0.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
19
+ "context_refiner.0.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
20
+ "context_refiner.0.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
21
+ "context_refiner.0.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
22
+ "context_refiner.0.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
23
+ "context_refiner.0.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
24
+ "context_refiner.0.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
25
+ "context_refiner.0.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
26
+ "context_refiner.0.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
27
+ "context_refiner.0.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
28
+ "context_refiner.0.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
29
+ "context_refiner.1.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
30
+ "context_refiner.1.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
31
+ "context_refiner.1.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
32
+ "context_refiner.1.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
33
+ "context_refiner.1.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
34
+ "context_refiner.1.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
35
+ "context_refiner.1.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
36
+ "context_refiner.1.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
37
+ "context_refiner.1.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
38
+ "context_refiner.1.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
39
+ "context_refiner.1.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
40
+ "context_refiner.1.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
41
+ "context_refiner.1.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
42
+ "layers.0.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
43
+ "layers.0.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
44
+ "layers.0.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
45
+ "layers.0.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
46
+ "layers.0.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
47
+ "layers.0.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
48
+ "layers.0.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
49
+ "layers.0.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
50
+ "layers.0.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
51
+ "layers.0.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
52
+ "layers.0.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
53
+ "layers.0.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
54
+ "layers.0.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
55
+ "layers.0.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
56
+ "layers.0.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
57
+ "layers.1.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
58
+ "layers.1.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
59
+ "layers.1.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
60
+ "layers.1.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
61
+ "layers.1.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
62
+ "layers.1.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
63
+ "layers.1.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
64
+ "layers.1.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
65
+ "layers.1.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
66
+ "layers.1.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
67
+ "layers.1.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
68
+ "layers.1.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
69
+ "layers.1.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
70
+ "layers.1.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
71
+ "layers.1.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
72
+ "layers.10.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
73
+ "layers.10.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
74
+ "layers.10.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
75
+ "layers.10.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
76
+ "layers.10.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
77
+ "layers.10.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
78
+ "layers.10.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
79
+ "layers.10.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
80
+ "layers.10.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
81
+ "layers.10.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
82
+ "layers.10.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
83
+ "layers.10.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
84
+ "layers.10.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
85
+ "layers.10.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
86
+ "layers.10.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
87
+ "layers.11.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
88
+ "layers.11.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
89
+ "layers.11.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
90
+ "layers.11.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
91
+ "layers.11.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
92
+ "layers.11.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
93
+ "layers.11.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
94
+ "layers.11.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
95
+ "layers.11.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
96
+ "layers.11.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
97
+ "layers.11.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
98
+ "layers.11.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
99
+ "layers.11.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
100
+ "layers.11.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
101
+ "layers.11.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
102
+ "layers.12.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
103
+ "layers.12.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
104
+ "layers.12.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
105
+ "layers.12.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
106
+ "layers.12.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
107
+ "layers.12.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
108
+ "layers.12.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
109
+ "layers.12.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
110
+ "layers.12.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
111
+ "layers.12.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
112
+ "layers.12.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
113
+ "layers.12.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
114
+ "layers.12.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
115
+ "layers.12.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
116
+ "layers.12.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
117
+ "layers.13.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
118
+ "layers.13.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
119
+ "layers.13.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
120
+ "layers.13.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
121
+ "layers.13.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
122
+ "layers.13.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
123
+ "layers.13.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
124
+ "layers.13.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
125
+ "layers.13.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
126
+ "layers.13.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
127
+ "layers.13.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
128
+ "layers.13.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
129
+ "layers.13.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
130
+ "layers.13.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
131
+ "layers.13.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
132
+ "layers.14.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
133
+ "layers.14.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
134
+ "layers.14.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
135
+ "layers.14.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
136
+ "layers.14.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
137
+ "layers.14.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
138
+ "layers.14.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
139
+ "layers.14.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
140
+ "layers.14.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
141
+ "layers.14.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
142
+ "layers.14.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
143
+ "layers.14.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
144
+ "layers.14.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
145
+ "layers.14.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
146
+ "layers.14.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
147
+ "layers.15.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
148
+ "layers.15.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
149
+ "layers.15.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
150
+ "layers.15.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
151
+ "layers.15.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
152
+ "layers.15.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
153
+ "layers.15.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
154
+ "layers.15.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
155
+ "layers.15.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
156
+ "layers.15.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
157
+ "layers.15.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
158
+ "layers.15.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
159
+ "layers.15.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
160
+ "layers.15.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
161
+ "layers.15.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
162
+ "layers.16.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
163
+ "layers.16.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
164
+ "layers.16.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
165
+ "layers.16.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
166
+ "layers.16.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
167
+ "layers.16.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
168
+ "layers.16.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
169
+ "layers.16.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
170
+ "layers.16.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
171
+ "layers.16.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
172
+ "layers.16.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
173
+ "layers.16.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
174
+ "layers.16.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
175
+ "layers.16.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
176
+ "layers.16.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
177
+ "layers.17.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
178
+ "layers.17.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
179
+ "layers.17.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
180
+ "layers.17.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
181
+ "layers.17.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
182
+ "layers.17.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
183
+ "layers.17.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
184
+ "layers.17.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
185
+ "layers.17.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
186
+ "layers.17.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
187
+ "layers.17.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
188
+ "layers.17.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
189
+ "layers.17.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
190
+ "layers.17.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
191
+ "layers.17.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
192
+ "layers.18.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
193
+ "layers.18.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
194
+ "layers.18.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
195
+ "layers.18.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
196
+ "layers.18.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
197
+ "layers.18.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
198
+ "layers.18.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
199
+ "layers.18.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
200
+ "layers.18.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
201
+ "layers.18.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
202
+ "layers.18.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
203
+ "layers.18.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
204
+ "layers.18.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
205
+ "layers.18.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
206
+ "layers.18.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
207
+ "layers.19.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
208
+ "layers.19.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
209
+ "layers.19.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
210
+ "layers.19.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
211
+ "layers.19.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
212
+ "layers.19.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
213
+ "layers.19.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
214
+ "layers.19.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
215
+ "layers.19.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
216
+ "layers.19.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
217
+ "layers.19.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
218
+ "layers.19.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
219
+ "layers.19.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
220
+ "layers.19.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
221
+ "layers.19.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
222
+ "layers.2.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
223
+ "layers.2.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
224
+ "layers.2.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
225
+ "layers.2.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
226
+ "layers.2.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
227
+ "layers.2.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
228
+ "layers.2.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
229
+ "layers.2.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
230
+ "layers.2.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
231
+ "layers.2.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
232
+ "layers.2.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
233
+ "layers.2.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
234
+ "layers.2.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
235
+ "layers.2.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
236
+ "layers.2.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
237
+ "layers.20.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
238
+ "layers.20.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
239
+ "layers.20.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
240
+ "layers.20.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
241
+ "layers.20.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
242
+ "layers.20.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
243
+ "layers.20.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
244
+ "layers.20.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
245
+ "layers.20.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
246
+ "layers.20.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
247
+ "layers.20.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
248
+ "layers.20.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
249
+ "layers.20.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
250
+ "layers.20.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
251
+ "layers.20.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
252
+ "layers.21.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
253
+ "layers.21.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
254
+ "layers.21.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
255
+ "layers.21.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
256
+ "layers.21.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
257
+ "layers.21.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
258
+ "layers.21.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
259
+ "layers.21.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
260
+ "layers.21.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
261
+ "layers.21.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
262
+ "layers.21.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
263
+ "layers.21.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
264
+ "layers.21.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
265
+ "layers.21.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
266
+ "layers.21.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
267
+ "layers.22.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
268
+ "layers.22.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
269
+ "layers.22.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
270
+ "layers.22.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
271
+ "layers.22.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
272
+ "layers.22.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
273
+ "layers.22.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
274
+ "layers.22.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
275
+ "layers.22.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
276
+ "layers.22.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
277
+ "layers.22.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
278
+ "layers.22.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
279
+ "layers.22.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
280
+ "layers.22.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
281
+ "layers.22.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
282
+ "layers.23.adaLN_modulation.0.bias": "diffusion_pytorch_model-00002-of-00002.safetensors",
283
+ "layers.23.adaLN_modulation.0.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
284
+ "layers.23.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
285
+ "layers.23.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
286
+ "layers.23.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
287
+ "layers.23.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
288
+ "layers.23.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
289
+ "layers.23.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
290
+ "layers.23.attention_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
291
+ "layers.23.attention_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
292
+ "layers.23.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
293
+ "layers.23.feed_forward.w2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
294
+ "layers.23.feed_forward.w3.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
295
+ "layers.23.ffn_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
296
+ "layers.23.ffn_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
297
+ "layers.24.adaLN_modulation.0.bias": "diffusion_pytorch_model-00002-of-00002.safetensors",
298
+ "layers.24.adaLN_modulation.0.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
299
+ "layers.24.attention.norm_k.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
300
+ "layers.24.attention.norm_q.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
301
+ "layers.24.attention.to_k.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
302
+ "layers.24.attention.to_out.0.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
303
+ "layers.24.attention.to_q.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
304
+ "layers.24.attention.to_v.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
305
+ "layers.24.attention_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
306
+ "layers.24.attention_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
307
+ "layers.24.feed_forward.w1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
308
+ "layers.24.feed_forward.w2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
309
+ "layers.24.feed_forward.w3.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
310
+ "layers.24.ffn_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
311
+ "layers.24.ffn_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
312
+ "layers.25.adaLN_modulation.0.bias": "diffusion_pytorch_model-00002-of-00002.safetensors",
313
+ "layers.25.adaLN_modulation.0.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
314
+ "layers.25.attention.norm_k.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
315
+ "layers.25.attention.norm_q.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
316
+ "layers.25.attention.to_k.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
317
+ "layers.25.attention.to_out.0.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
318
+ "layers.25.attention.to_q.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
319
+ "layers.25.attention.to_v.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
320
+ "layers.25.attention_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
321
+ "layers.25.attention_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
322
+ "layers.25.feed_forward.w1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
323
+ "layers.25.feed_forward.w2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
324
+ "layers.25.feed_forward.w3.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
325
+ "layers.25.ffn_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
326
+ "layers.25.ffn_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
327
+ "layers.26.adaLN_modulation.0.bias": "diffusion_pytorch_model-00002-of-00002.safetensors",
328
+ "layers.26.adaLN_modulation.0.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
329
+ "layers.26.attention.norm_k.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
330
+ "layers.26.attention.norm_q.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
331
+ "layers.26.attention.to_k.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
332
+ "layers.26.attention.to_out.0.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
333
+ "layers.26.attention.to_q.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
334
+ "layers.26.attention.to_v.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
335
+ "layers.26.attention_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
336
+ "layers.26.attention_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
337
+ "layers.26.feed_forward.w1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
338
+ "layers.26.feed_forward.w2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
339
+ "layers.26.feed_forward.w3.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
340
+ "layers.26.ffn_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
341
+ "layers.26.ffn_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
342
+ "layers.27.adaLN_modulation.0.bias": "diffusion_pytorch_model-00002-of-00002.safetensors",
343
+ "layers.27.adaLN_modulation.0.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
344
+ "layers.27.attention.norm_k.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
345
+ "layers.27.attention.norm_q.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
346
+ "layers.27.attention.to_k.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
347
+ "layers.27.attention.to_out.0.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
348
+ "layers.27.attention.to_q.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
349
+ "layers.27.attention.to_v.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
350
+ "layers.27.attention_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
351
+ "layers.27.attention_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
352
+ "layers.27.feed_forward.w1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
353
+ "layers.27.feed_forward.w2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
354
+ "layers.27.feed_forward.w3.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
355
+ "layers.27.ffn_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
356
+ "layers.27.ffn_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
357
+ "layers.28.adaLN_modulation.0.bias": "diffusion_pytorch_model-00002-of-00002.safetensors",
358
+ "layers.28.adaLN_modulation.0.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
359
+ "layers.28.attention.norm_k.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
360
+ "layers.28.attention.norm_q.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
361
+ "layers.28.attention.to_k.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
362
+ "layers.28.attention.to_out.0.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
363
+ "layers.28.attention.to_q.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
364
+ "layers.28.attention.to_v.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
365
+ "layers.28.attention_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
366
+ "layers.28.attention_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
367
+ "layers.28.feed_forward.w1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
368
+ "layers.28.feed_forward.w2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
369
+ "layers.28.feed_forward.w3.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
370
+ "layers.28.ffn_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
371
+ "layers.28.ffn_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
372
+ "layers.29.adaLN_modulation.0.bias": "diffusion_pytorch_model-00002-of-00002.safetensors",
373
+ "layers.29.adaLN_modulation.0.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
374
+ "layers.29.attention.norm_k.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
375
+ "layers.29.attention.norm_q.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
376
+ "layers.29.attention.to_k.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
377
+ "layers.29.attention.to_out.0.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
378
+ "layers.29.attention.to_q.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
379
+ "layers.29.attention.to_v.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
380
+ "layers.29.attention_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
381
+ "layers.29.attention_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
382
+ "layers.29.feed_forward.w1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
383
+ "layers.29.feed_forward.w2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
384
+ "layers.29.feed_forward.w3.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
385
+ "layers.29.ffn_norm1.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
386
+ "layers.29.ffn_norm2.weight": "diffusion_pytorch_model-00002-of-00002.safetensors",
387
+ "layers.3.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
388
+ "layers.3.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
389
+ "layers.3.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
390
+ "layers.3.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
391
+ "layers.3.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
392
+ "layers.3.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
393
+ "layers.3.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
394
+ "layers.3.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
395
+ "layers.3.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
396
+ "layers.3.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
397
+ "layers.3.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
398
+ "layers.3.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
399
+ "layers.3.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
400
+ "layers.3.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
401
+ "layers.3.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
402
+ "layers.4.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
403
+ "layers.4.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
404
+ "layers.4.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
405
+ "layers.4.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
406
+ "layers.4.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
407
+ "layers.4.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
408
+ "layers.4.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
409
+ "layers.4.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
410
+ "layers.4.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
411
+ "layers.4.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
412
+ "layers.4.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
413
+ "layers.4.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
414
+ "layers.4.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
415
+ "layers.4.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
416
+ "layers.4.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
417
+ "layers.5.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
418
+ "layers.5.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
419
+ "layers.5.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
420
+ "layers.5.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
421
+ "layers.5.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
422
+ "layers.5.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
423
+ "layers.5.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
424
+ "layers.5.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
425
+ "layers.5.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
426
+ "layers.5.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
427
+ "layers.5.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
428
+ "layers.5.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
429
+ "layers.5.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
430
+ "layers.5.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
431
+ "layers.5.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
432
+ "layers.6.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
433
+ "layers.6.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
434
+ "layers.6.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
435
+ "layers.6.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
436
+ "layers.6.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
437
+ "layers.6.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
438
+ "layers.6.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
439
+ "layers.6.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
440
+ "layers.6.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
441
+ "layers.6.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
442
+ "layers.6.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
443
+ "layers.6.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
444
+ "layers.6.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
445
+ "layers.6.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
446
+ "layers.6.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
447
+ "layers.7.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
448
+ "layers.7.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
449
+ "layers.7.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
450
+ "layers.7.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
451
+ "layers.7.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
452
+ "layers.7.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
453
+ "layers.7.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
454
+ "layers.7.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
455
+ "layers.7.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
456
+ "layers.7.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
457
+ "layers.7.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
458
+ "layers.7.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
459
+ "layers.7.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
460
+ "layers.7.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
461
+ "layers.7.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
462
+ "layers.8.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
463
+ "layers.8.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
464
+ "layers.8.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
465
+ "layers.8.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
466
+ "layers.8.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
467
+ "layers.8.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
468
+ "layers.8.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
469
+ "layers.8.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
470
+ "layers.8.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
471
+ "layers.8.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
472
+ "layers.8.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
473
+ "layers.8.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
474
+ "layers.8.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
475
+ "layers.8.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
476
+ "layers.8.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
477
+ "layers.9.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
478
+ "layers.9.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
479
+ "layers.9.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
480
+ "layers.9.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
481
+ "layers.9.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
482
+ "layers.9.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
483
+ "layers.9.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
484
+ "layers.9.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
485
+ "layers.9.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
486
+ "layers.9.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
487
+ "layers.9.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
488
+ "layers.9.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
489
+ "layers.9.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
490
+ "layers.9.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
491
+ "layers.9.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
492
+ "noise_refiner.0.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
493
+ "noise_refiner.0.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
494
+ "noise_refiner.0.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
495
+ "noise_refiner.0.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
496
+ "noise_refiner.0.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
497
+ "noise_refiner.0.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
498
+ "noise_refiner.0.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
499
+ "noise_refiner.0.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
500
+ "noise_refiner.0.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
501
+ "noise_refiner.0.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
502
+ "noise_refiner.0.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
503
+ "noise_refiner.0.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
504
+ "noise_refiner.0.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
505
+ "noise_refiner.0.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
506
+ "noise_refiner.0.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
507
+ "noise_refiner.1.adaLN_modulation.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
508
+ "noise_refiner.1.adaLN_modulation.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
509
+ "noise_refiner.1.attention.norm_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
510
+ "noise_refiner.1.attention.norm_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
511
+ "noise_refiner.1.attention.to_k.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
512
+ "noise_refiner.1.attention.to_out.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
513
+ "noise_refiner.1.attention.to_q.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
514
+ "noise_refiner.1.attention.to_v.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
515
+ "noise_refiner.1.attention_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
516
+ "noise_refiner.1.attention_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
517
+ "noise_refiner.1.feed_forward.w1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
518
+ "noise_refiner.1.feed_forward.w2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
519
+ "noise_refiner.1.feed_forward.w3.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
520
+ "noise_refiner.1.ffn_norm1.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
521
+ "noise_refiner.1.ffn_norm2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
522
+ "t_embedder.mlp.0.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
523
+ "t_embedder.mlp.0.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
524
+ "t_embedder.mlp.2.bias": "diffusion_pytorch_model-00001-of-00002.safetensors",
525
+ "t_embedder.mlp.2.weight": "diffusion_pytorch_model-00001-of-00002.safetensors",
526
+ "x_pad_token": "diffusion_pytorch_model-00001-of-00002.safetensors"
527
+ }
528
+ }
diffusers/vae/config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.37.0.dev0",
4
+ "_name_or_path": "../checkpoints/flux-dev",
5
+ "act_fn": "silu",
6
+ "block_out_channels": [
7
+ 128,
8
+ 256,
9
+ 512,
10
+ 512
11
+ ],
12
+ "down_block_types": [
13
+ "DownEncoderBlock2D",
14
+ "DownEncoderBlock2D",
15
+ "DownEncoderBlock2D",
16
+ "DownEncoderBlock2D"
17
+ ],
18
+ "force_upcast": true,
19
+ "in_channels": 3,
20
+ "latent_channels": 16,
21
+ "latents_mean": null,
22
+ "latents_std": null,
23
+ "layers_per_block": 2,
24
+ "mid_block_add_attention": true,
25
+ "norm_num_groups": 32,
26
+ "out_channels": 3,
27
+ "sample_size": 1024,
28
+ "scaling_factor": 0.3611,
29
+ "shift_factor": 0.1159,
30
+ "up_block_types": [
31
+ "UpDecoderBlock2D",
32
+ "UpDecoderBlock2D",
33
+ "UpDecoderBlock2D",
34
+ "UpDecoderBlock2D"
35
+ ],
36
+ "use_post_quant_conv": false,
37
+ "use_quant_conv": false
38
+ }
diffusers/vae/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c717328c8ad41faab2ccfd52ae17332505c6833cf176aad56e7b58f2c4d4c94
3
+ size 335306212
diffusion_models/z-anime-base-bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3012b73377b234d47b932809a9ed596fb57fd57f901dd71c0ecb13d835463518
3
+ size 12309874488
diffusion_models/z-anime-base-fp8.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d2faf5f7e223ecf4942b52b55d816b026afdb7a447cb98cf7cd297240ffad96
3
+ size 6154966640
diffusion_models/z-anime-distill-4step-bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a083f389432c5cf04bfa686a1e819f894d01e4a00bdf02e846617e17052ac77
3
+ size 12309881520
diffusion_models/z-anime-distill-4step-fp8.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2bb34d67a2a1167bb5e0eba17ccf352d1d5c6d69018bda7d71f886e323a6b0bf
3
+ size 6154968720
diffusion_models/z-anime-distill-8step-bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63b560e512306d2d97518e93cb0f3f306b014a0b4082da3a6b8a046d2677f8e1
3
+ size 12309880976
diffusion_models/z-anime-distill-8step-fp8.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:287b8474ffb15e361dbde51da5bce2770fcf038a144b5ad92083d217d549cb31
3
+ size 6154968720
gguf/z-anime-base-q4_k_s.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4b8ffb198f16c20a425853e6e3144d2d99c99d1f5f65c5b229a7e472ee167ed
3
+ size 4511516736
gguf/z-anime-base-q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f69475bfb1dbedbf9421a75270e2bbe6d7b71d71dabf6c798fc432c21daca271
3
+ size 7224707136
images/1.png ADDED

Git LFS Details

  • SHA256: 1a4d0ea0015d0954c488361107a8ed105a7ad926f14dc9b45ebcfd91c61c445f
  • Pointer size: 132 Bytes
  • Size of remote file: 3.75 MB
images/2.png ADDED

Git LFS Details

  • SHA256: 8540fda505055eb20a36c6ed0216884f35f45f3fb2f83c49e94882c83f2cf91f
  • Pointer size: 132 Bytes
  • Size of remote file: 4.98 MB
images/3.png ADDED

Git LFS Details

  • SHA256: 39a95caec8c6814fc6f2ae447e8eb74d966b566b2caf35d55ffb27562ce84e53
  • Pointer size: 132 Bytes
  • Size of remote file: 4.62 MB
images/4.png ADDED

Git LFS Details

  • SHA256: 407ed85853bcdd977ea356882e903dc92f1753f99061a5aca835a94cbc901728
  • Pointer size: 132 Bytes
  • Size of remote file: 3.9 MB
images/5.png ADDED

Git LFS Details

  • SHA256: 3afb0460afdc5ff975e908218753d9c51616c62e3d3a494f3d7e73761b765934
  • Pointer size: 132 Bytes
  • Size of remote file: 3.49 MB
images/6.png ADDED

Git LFS Details

  • SHA256: a31be0bf88a69d8c718d2c6cff3381a929dc03be659aae8067de3bb534b159b0
  • Pointer size: 132 Bytes
  • Size of remote file: 4.22 MB
images/7.png ADDED

Git LFS Details

  • SHA256: 0be2595b04526058a54a3d50bb99e15146ce24de0ef542e13e799ebce235b271
  • Pointer size: 132 Bytes
  • Size of remote file: 3.71 MB
images/8.png ADDED

Git LFS Details

  • SHA256: 479fea796476ef79034337d7ecc5358298c32ac400526e66b91a49dbb9348a5e
  • Pointer size: 132 Bytes
  • Size of remote file: 1.96 MB
images/9.png ADDED

Git LFS Details

  • SHA256: 6be70441300a704ccc51dcf097fa8480d5ec939fc974c21f4a38f844fcb3616f
  • Pointer size: 132 Bytes
  • Size of remote file: 2.16 MB
images/cover.png ADDED

Git LFS Details

  • SHA256: 95f52edced663294c26cfa89c5adbfd50c7cdfab5244f102e60d6991fa608da4
  • Pointer size: 132 Bytes
  • Size of remote file: 4 MB
images/workflow-cover.png ADDED

Git LFS Details

  • SHA256: a9714fec9db98736d851a82d52c16aa402d87841fe0f48c3d6d28d65aca922ba
  • Pointer size: 132 Bytes
  • Size of remote file: 1.91 MB
images/workflow-overview.png ADDED
text_encoder/qwen_3_4b-bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c671498573ac2f7a5501502ccce8d2b08ea6ca2f661c458e708f36b36edfc5a
3
+ size 8044982048
text_encoder/qwen_3_4b-engineer-v4-bf16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd925aad8df000d3f8b306b5f5780dd3c41c944476c5e9ca20a72ad455b2eeef
3
+ size 8044982080
text_encoder/qwen_3_4b-engineer-v4-fp8.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:deb07cdb971031fcd0af164a56468453b980278b6d4b4b6633b2f58f57fbb5e3
3
+ size 4022515040
text_encoder/qwen_3_4b-fp8.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38a245fc197f16c4025467ef46dce247d076af8f013bb8b1617013beea46d0e4
3
+ size 4022515040