## VQGAN-f16-16384 Model converted to JAX from [boris/vqgan_f16_16384](https://huggingface.co/boris/vqgan_f16_16384). Model finetuned with [taming-transformers](https://github.com/CompVis/taming-transformers): * Training run * [Part 1](https://wandb.ai/wandb/hf-flax-dalle-mini/runs/2021-07-09T15-33-11_dalle_vqgan?workspace=user-borisd13) - started from [vqgan_imagenet_f16_16384 checkpoint](https://heibox.uni-heidelberg.de/d/a7530b09fed84f80a887/) (pretrained on ImageNet) * [Part 2](https://wandb.ai/wandb/hf-flax-dalle-mini/runs/2021-07-09T21-42-07_dalle_vqgan?workspace=user-borisd13) - continuation from Part 1 * Dataset: subset of 2,268,720 images processed once originating from [Conceptual Captions 3M](https://ai.google.com/research/ConceptualCaptions/) and [OpenAI subset of YFCC100M](https://github.com/openai/CLIP/blob/main/data/yfcc100m.md) * Checkpoint uploaded from last artifact version (see training run)