NourFakih's picture
Model save
832248b verified
|
raw
history blame
No virus
5.4 kB
---
license: apache-2.0
base_model: NourFakih/image-captioning-Vit-GPT2-Flickr8k
tags:
- generated_from_trainer
metrics:
- rouge
model-index:
- name: Vit-GPT2-COCO2017Sample-Flickr8k
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# Vit-GPT2-COCO2017Sample-Flickr8k
This model is a fine-tuned version of [NourFakih/image-captioning-Vit-GPT2-Flickr8k](https://huggingface.co/NourFakih/image-captioning-Vit-GPT2-Flickr8k) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2344
- Rouge1: 41.2779
- Rouge2: 15.8081
- Rougel: 37.3177
- Rougelsum: 37.2772
- Gen Len: 11.568
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
| Training Loss | Epoch | Step | Gen Len | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum |
|:-------------:|:-----:|:-----:|:-------:|:---------------:|:-------:|:-------:|:-------:|:---------:|
| 0.2485 | 0.08 | 500 | 11.382 | 0.2394 | 40.2445 | 13.9525 | 36.1329 | 36.1261 |
| 0.2336 | 0.16 | 1000 | 11.346 | 0.2376 | 39.9631 | 14.4981 | 36.1164 | 36.1265 |
| 0.2282 | 0.24 | 1500 | 11.07 | 0.2367 | 40.3107 | 14.533 | 36.489 | 36.5082 |
| 0.2249 | 0.32 | 2000 | 10.87 | 0.2338 | 41.0525 | 15.3076 | 37.0189 | 37.0365 |
| 0.2302 | 0.4 | 2500 | 11.31 | 0.2329 | 40.7052 | 14.8288 | 36.9272 | 36.9197 |
| 0.2255 | 0.48 | 3000 | 11.05 | 0.2321 | 40.6896 | 15.2723 | 36.9654 | 36.9515 |
| 0.2225 | 0.56 | 3500 | 10.946 | 0.2305 | 40.705 | 15.4878 | 37.0456 | 37.0235 |
| 0.2233 | 0.64 | 4000 | 11.25 | 0.2303 | 41.0229 | 15.179 | 37.081 | 37.0924 |
| 0.2177 | 0.72 | 4500 | 11.08 | 0.2307 | 40.0156 | 14.2972 | 36.0288 | 36.043 |
| 0.2159 | 0.8 | 5000 | 11.336 | 0.2298 | 40.4042 | 15.2531 | 36.6967 | 36.7003 |
| 0.2189 | 0.88 | 5500 | 11.39 | 0.2282 | 40.167 | 14.4847 | 36.3855 | 36.3742 |
| 0.2171 | 0.96 | 6000 | 11.002 | 0.2269 | 40.8528 | 15.1811 | 37.0586 | 37.0403 |
| 0.1962 | 1.04 | 6500 | 11.598 | 0.2296 | 40.6676 | 14.9888 | 36.7796 | 36.7703 |
| 0.1835 | 1.12 | 7000 | 11.022 | 0.2311 | 40.6188 | 15.2743 | 36.8519 | 36.8263 |
| 0.1835 | 1.2 | 7500 | 11.248 | 0.2289 | 40.6466 | 15.1727 | 36.6626 | 36.6427 |
| 0.1864 | 1.28 | 8000 | 11.408 | 0.2298 | 40.2408 | 15.0179 | 36.5594 | 36.5756 |
| 0.1838 | 1.36 | 8500 | 11.238 | 0.2295 | 41.0772 | 15.2152 | 37.0647 | 37.0648 |
| 0.1827 | 1.44 | 9000 | 11.28 | 0.2299 | 40.3263 | 14.9976 | 36.6444 | 36.6292 |
| 0.1828 | 1.52 | 9500 | 11.132 | 0.2299 | 40.9308 | 15.181 | 36.9028 | 36.8909 |
| 0.179 | 1.61 | 10000 | 11.164 | 0.2287 | 40.7406 | 15.2746 | 36.85 | 36.8748 |
| 0.1849 | 1.69 | 10500 | 10.988 | 0.2281 | 40.931 | 15.6479 | 37.0222 | 37.0071 |
| 0.1794 | 1.77 | 11000 | 11.218 | 0.2281 | 41.5198 | 15.9659 | 37.3709 | 37.386 |
| 0.1787 | 1.85 | 11500 | 11.274 | 0.2278 | 40.4006 | 14.9496 | 36.4608 | 36.4675 |
| 0.1798 | 1.93 | 12000 | 11.154 | 0.2279 | 41.3118 | 15.4673 | 37.4917 | 37.5101 |
| 0.1803 | 2.01 | 12500 | 11.23 | 0.2282 | 40.5652 | 15.1467 | 36.7946 | 36.7809 |
| 0.1519 | 2.09 | 13000 | 11.498 | 0.2361 | 40.8978 | 15.0865 | 36.7157 | 36.728 |
| 0.1515 | 2.17 | 13500 | 11.37 | 0.2360 | 40.9809 | 15.5877 | 37.0104 | 36.9942 |
| 0.1519 | 2.25 | 14000 | 11.504 | 0.2359 | 40.7947 | 15.3254 | 36.9574 | 36.9431 |
| 0.1543 | 2.33 | 14500 | 0.2346 | 40.7724 | 15.1837 | 36.9003 | 36.848 | 11.586 |
| 0.1548 | 2.41 | 15000 | 0.2355 | 40.7237 | 15.2394 | 37.0767 | 37.0405 | 11.294 |
| 0.1507 | 2.49 | 15500 | 0.2353 | 41.2661 | 15.7703 | 37.3669 | 37.32 | 11.308 |
| 0.1512 | 2.57 | 16000 | 0.2351 | 40.8777 | 15.2821 | 36.9591 | 36.9201 | 11.43 |
| 0.1525 | 2.65 | 16500 | 0.2350 | 40.6184 | 15.1824 | 36.655 | 36.6117 | 11.402 |
| 0.1522 | 2.73 | 17000 | 0.2343 | 41.2818 | 15.7174 | 37.3059 | 37.2695 | 11.502 |
| 0.1544 | 2.81 | 17500 | 0.2349 | 41.0821 | 15.5164 | 37.2206 | 37.1663 | 11.542 |
| 0.1498 | 2.89 | 18000 | 0.2346 | 41.2128 | 15.6698 | 37.2279 | 37.1874 | 11.582 |
| 0.1497 | 2.97 | 18500 | 0.2344 | 41.2779 | 15.8081 | 37.3177 | 37.2772 | 11.568 |
### Framework versions
- Transformers 4.39.3
- Pytorch 2.1.2
- Datasets 2.18.0
- Tokenizers 0.15.2