srisweet commited on
Commit
4df2547
1 Parent(s): c0b17af

Updated Intro lines in the beginning

Browse files
Files changed (1) hide show
  1. introduction.md +3 -1
introduction.md CHANGED
@@ -1,6 +1,8 @@
1
  # CLIP-Italian
2
 
3
- CLIP-Italian is a multimodal model trained on ~1.4 million Italian text-image pairs using Italian Bert model as text encoder and Vision Transformer(ViT) as image encoder.Clip-Italian (Contrastive Language-Image Pre-training in Italian language) is based on OpenAI’s CLIP ([Radford et al., 2021](https://arxiv.org/abs/2103.00020))which is an amazing model that can learn to represent images and text jointly in the same space.
 
 
4
 
5
  In this project, we aim to propose the first CLIP model trained on Italian data, that in this context can be considered a
6
  low resource language. Using a few techniques, we have been able to fine-tune a SOTA Italian CLIP model with **only 1.4 million** training samples. Our Italian CLIP model
 
1
  # CLIP-Italian
2
 
3
+ CLIP-Italian is a multimodal model trained on ~1.4 million Italian text-image pairs using Italian Bert model as text encoder and Vision Transformer(ViT) as image encoder using the JAX/Flax neural network library. The training was carried out during the Hugging Face Community event on Google's TPU machines, sponsored by Google Cloud.
4
+
5
+ Clip-Italian (Contrastive Language-Image Pre-training in Italian language) is based on OpenAI’s CLIP ([Radford et al., 2021](https://arxiv.org/abs/2103.00020))which is an amazing model that can learn to represent images and text jointly in the same space.
6
 
7
  In this project, we aim to propose the first CLIP model trained on Italian data, that in this context can be considered a
8
  low resource language. Using a few techniques, we have been able to fine-tune a SOTA Italian CLIP model with **only 1.4 million** training samples. Our Italian CLIP model