This demo uses [CLIP-Vision-Marian model checkpoint](https://huggingface.co/flax-community/clip-vit-base-patch32_marian-es) to predict caption for a given image in Spanish. Training was done using image encoder and text decoder with approximately 2.5 million image-text pairs taken from the [Conceptual 12M dataset](https://github.com/google-research-datasets/conceptual-12m) with captions translated using [MarianMT English to Spanish](https://huggingface.co/transformers/model_doc/marian.html). For more details, click on `Usage` or `Article` 🤗 below.