clip-italian-demo / readme.md
vinid's picture
adding readme.md
80200b5
|
raw
history blame
883 Bytes
# Italian CLIP
....
# Novel Contributions
The original CLIP model was trained on 400millions text-image pairs; this amount of data is not available for Italian and the only datasets for captioning in the literature are MSCOCO-IT (translated version of MSCOCO) and WIT. To get competitive results we follewed three directions: 1) more data 2) better augmentation and 3) better training.
## More Data
## Better Augmentations
## Better Training
different optimizer and backbone freezing
# Scientific Validity
To better understand how well our clip-italian model works we run an experimental evaluation. Since this is the first clip-based model in Italian, we used the multilingual CLIP model as a comparison baseline.
We selected two different tasks:
+ image-retrieval
+ zero-shot classification
## Image Retrieval
## Zero-shot classification
# Broader Outlook