File size: 1,694 Bytes
ae2cc14 8eb0e39 ae2cc14 8eb0e39 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 |
---
license: creativeml-openrail-m
base_model: amused/amused-512
datasets:
- lambdalabs/pokemon-blip-captions
tags:
- stable-diffusion
- stable-diffusion-diffusers
- text-to-image
- diffusers
- amused
inference: true
---
# aMUSEd finetuning - suvadityamuk/amused-512-pokemon
This pipeline was finetuned from **amused/amused-512** on the **lambdalabs/pokemon-blip-captions** dataset. Below are some example images generated with the finetuned pipeline using the following prompts: ['a pokemon red mammoth with unicorn horns', 'a pokemon blue fish with golden scales', 'a pokemon green goblin with glasses, wearing black pants and red shirt', 'a pokemon golden unicorn with shiny black hair and deep blue horns', 'a pokemon drawing of a dragon with its mouth closed', 'a pokemon red and yellow phoenix with fire on its wings', 'a pokemon purple tree with white leaves and golden nectar flowing', 'a pokemon green caterpillar']:
![val_imgs_grid](./val_imgs_grid.png)
## Pipeline usage
You can use the pipeline like so:
```python
from diffusers import DiffusionPipeline
import torch
pipeline = DiffusionPipeline.from_pretrained("suvadityamuk/amused-512-pokemon", torch_dtype=torch.float16)
prompt = "a pokemon red mammoth with unicorn horns"
image = pipeline(prompt).images[0]
image.save("my_image.png")
```
## Training info
These are the key hyperparameters used during training:
* Train Steps: 750
* Learning rate: 5e-06
* Batch size: 8
* Gradient accumulation steps: 4
* Image resolution: 512
* Mixed-precision: bf16
More information on all the CLI arguments and the environment are available on your [`wandb` run page](https://wandb.ai/ml-colabs/fconn-amused/runs/0cljkrh9).
|