Spaces:
Running
Running
updating the readme
Browse files- introduction.md +11 -0
introduction.md
CHANGED
@@ -14,6 +14,17 @@ We put our **hearts** and **souls** into the project during this week! Not only
|
|
14 |
able to make new friends and and learn a lot from each other to work towards a common goal!
|
15 |
Thank you for this amazing opportunity, we hope you will like the results. :heart:
|
16 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
17 |
# Novel Contributions
|
18 |
|
19 |
The original CLIP model was trained on 400 million image-text pairs; this amount of data is not available for Italian.
|
|
|
14 |
able to make new friends and and learn a lot from each other to work towards a common goal!
|
15 |
Thank you for this amazing opportunity, we hope you will like the results. :heart:
|
16 |
|
17 |
+
# Demo
|
18 |
+
|
19 |
+
In this demo, we present two tasks:
|
20 |
+
|
21 |
+
+ *Text to Image*: This task is essentially an image retrieval task. The user is asked to input a string of text and CLIP is going to
|
22 |
+
compute the similarity between this string of text with respect to a set of images. The webapp is going to display the images that
|
23 |
+
have the highest similarity with the text query.
|
24 |
+
|
25 |
+
+ *Image to Text*: This task is essentially a zero-shot image classification task. The user is asked for an image and for a set of captions/labels and CLIP
|
26 |
+
is going to compute the similarity between the image and each label. The webapp is going to display a probability distribution over the captions.
|
27 |
+
|
28 |
# Novel Contributions
|
29 |
|
30 |
The original CLIP model was trained on 400 million image-text pairs; this amount of data is not available for Italian.
|