ayushman-dashtoon commited on
Commit
8aab7c1
1 Parent(s): 5fc21c0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +65 -1
README.md CHANGED
@@ -144,4 +144,68 @@ tags:
144
  <div class="masonry-item">
145
  <img class="custom-image" src="https://content.dashtoon.ai/assets/DashAnimeXL_Blog/367983.png" alt="image1">
146
  </div>
147
- </div>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
144
  <div class="masonry-item">
145
  <img class="custom-image" src="https://content.dashtoon.ai/assets/DashAnimeXL_Blog/367983.png" alt="image1">
146
  </div>
147
+ </div>
148
+
149
+
150
+ DashAnimeXL V1 is a diffusion-based text-to-image generative model. The model is a finetune of [SDXL](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/tree/main) by the research team at [Dashtoon](https://dashtoon.com/create).
151
+ Please see our blog for more details.
152
+
153
+ ### Model Description
154
+
155
+ - **Developed by:** [Dashtoon](https://dashtoon.com/create)
156
+ - **Model type:** Diffusion-based text-to-image generative model
157
+ - **License:** TBD
158
+ - **Model Description:** DashAnimeXL V1 is engineered to generate high-quality anime images from textual prompts. It features enhanced hand anatomy, better concept understanding, and prompt interpretation.
159
+ - **Summary:** This model generates images based on text prompts. It follows the same architecture as [Stable Diffusion XL](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0).
160
+ - **Finetuned from model:** [SDXL](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/tree/main)
161
+
162
+ ## Using the model with 🧨 Diffusers
163
+ To use DashAnimeXL V1, install the required libraries as follows:
164
+
165
+ ```python
166
+ pip install diffusers --upgrade
167
+ pip install transformers accelerate safetensors
168
+ ```
169
+
170
+ Example script for generating images with DashAnimeXL V1:
171
+
172
+ ```python
173
+ import torch
174
+ from diffusers import (
175
+ StableDiffusionXLPipeline,
176
+ EulerAncestralDiscreteScheduler,
177
+ AutoencoderKL
178
+ )
179
+
180
+ # Load VAE component
181
+ vae = AutoencoderKL.from_pretrained(
182
+ "madebyollin/sdxl-vae-fp16-fix",
183
+ torch_dtype=torch.bfloat16
184
+ )
185
+
186
+ # Configure the pipeline
187
+ pipe = StableDiffusionXLPipeline.from_pretrained(
188
+ "Linaqruf/animagine-xl-3.0",
189
+ vae=vae,
190
+ torch_dtype=torch.bfloat16,
191
+ use_safetensors=True,
192
+ )
193
+ pipe.scheduler = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config)
194
+
195
+ if torch.cuda.is_available():
196
+ pipe.to('cuda')
197
+
198
+ # Define prompts and generate image
199
+ prompt = "anime illustration, An ink painting with a superhot, pop art style, featuring vibrant splashes and gradient patterns merging with random signals and noise. A zoomed-in panda wearing glasses, appearing to look directly at the viewer. The piece is bathed in warm, volumetric lighting against a clear dusk sky background. The reflection in the panda's sunglasses reveals nuclear clouds, adding an element of surrealism."
200
+ negative_prompt = "nsfw, low quality, worst quality, very displeasing, 3d, watermark, signature, ugly, poorly drawn"
201
+
202
+ image = pipe(
203
+ prompt,
204
+ negative_prompt=negative_prompt,
205
+ width=1024,
206
+ height=1024,
207
+ guidance_scale=7,
208
+ num_inference_steps=20
209
+ ).images[0]
210
+
211
+ ```