Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,42 @@
|
|
1 |
-
---
|
2 |
-
license: mit
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: mit
|
3 |
+
datasets:
|
4 |
+
- imirandam/TROHN-Text
|
5 |
+
---
|
6 |
+
|
7 |
+
# Model Card for CLIP_TROHN-Text
|
8 |
+
|
9 |
+
## Model Description
|
10 |
+
- **Homepage:** https://imirandam.github.io/BiVLC_project_page/
|
11 |
+
- **Repository:** https://github.com/IMirandaM/BiVLC
|
12 |
+
- **Paper:**
|
13 |
+
- **Point of Contact:** [Imanol Miranda](mailto:imanol.miranda@ehu.eus)
|
14 |
+
|
15 |
+
### Model Summary
|
16 |
+
|
17 |
+
CLIP_TROHN-Text is a model presented in the [BiVLC](https://github.com/IMirandaM/BiVLC) paper for experimentation. It has been fine-tuned with OpenCLIP framework using as basis the CLIP ViT-B-32 model pre-trained by 'openai'. Hyperparameters:
|
18 |
+
|
19 |
+
* Learning rate: 1e-6.
|
20 |
+
* Scheduler: Cosine scheduler with 50 warmup steps.
|
21 |
+
* Optimizer: AdamW optimizer with beta1 = 0.9, beta2 = 0.98, eps = 1e-6 and weight decay = 0.1.
|
22 |
+
* Loss function: InfoNCE Loss. The loss is modified to add only negative captions following the idea proposed in NEGCLIP.
|
23 |
+
* Batch size: We define a batch size of 200, and then we add negatives. As it has not hard negative images, it results in 200 images x 400 captions (positive + hard negatives).
|
24 |
+
* Epochs: We fine-tune all models over 10 epochs and we used validation accuracy as the model selection criterion, i.e. we selected the model with the highest accuracy on the corresponding validation set.
|
25 |
+
* Data: It is fine-tuned with [TROHN-Text](https://huggingface.co/datasets/imirandam/TROHN-Text) dataset.
|
26 |
+
|
27 |
+
### Evaluation Data
|
28 |
+
The model is evaluated in [BiVLC](https://huggingface.co/datasets/imirandam/BiVLC).
|
29 |
+
|
30 |
+
### Licensing Information
|
31 |
+
This work is licensed under a MIT License.
|
32 |
+
|
33 |
+
## Citation Information
|
34 |
+
If you find this dataset useful, please consider citing our paper:
|
35 |
+
```
|
36 |
+
@inproceedings{,
|
37 |
+
title={},
|
38 |
+
author={},
|
39 |
+
booktitle={},
|
40 |
+
year={}
|
41 |
+
}
|
42 |
+
```
|