File size: 5,402 Bytes
875ea0c
 
 
 
 
832248b
 
875ea0c
 
 
 
 
 
 
 
 
 
 
 
832248b
 
 
 
 
 
875ea0c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
832248b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
875ea0c
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
---
license: apache-2.0
base_model: NourFakih/image-captioning-Vit-GPT2-Flickr8k
tags:
- generated_from_trainer
metrics:
- rouge
model-index:
- name: Vit-GPT2-COCO2017Sample-Flickr8k
  results: []
---

<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->

# Vit-GPT2-COCO2017Sample-Flickr8k

This model is a fine-tuned version of [NourFakih/image-captioning-Vit-GPT2-Flickr8k](https://huggingface.co/NourFakih/image-captioning-Vit-GPT2-Flickr8k) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2344
- Rouge1: 41.2779
- Rouge2: 15.8081
- Rougel: 37.3177
- Rougelsum: 37.2772
- Gen Len: 11.568

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0

### Training results

| Training Loss | Epoch | Step  | Gen Len | Validation Loss | Rouge1  | Rouge2  | Rougel  | Rougelsum |
|:-------------:|:-----:|:-----:|:-------:|:---------------:|:-------:|:-------:|:-------:|:---------:|
| 0.2485        | 0.08  | 500   | 11.382  | 0.2394          | 40.2445 | 13.9525 | 36.1329 | 36.1261   |
| 0.2336        | 0.16  | 1000  | 11.346  | 0.2376          | 39.9631 | 14.4981 | 36.1164 | 36.1265   |
| 0.2282        | 0.24  | 1500  | 11.07   | 0.2367          | 40.3107 | 14.533  | 36.489  | 36.5082   |
| 0.2249        | 0.32  | 2000  | 10.87   | 0.2338          | 41.0525 | 15.3076 | 37.0189 | 37.0365   |
| 0.2302        | 0.4   | 2500  | 11.31   | 0.2329          | 40.7052 | 14.8288 | 36.9272 | 36.9197   |
| 0.2255        | 0.48  | 3000  | 11.05   | 0.2321          | 40.6896 | 15.2723 | 36.9654 | 36.9515   |
| 0.2225        | 0.56  | 3500  | 10.946  | 0.2305          | 40.705  | 15.4878 | 37.0456 | 37.0235   |
| 0.2233        | 0.64  | 4000  | 11.25   | 0.2303          | 41.0229 | 15.179  | 37.081  | 37.0924   |
| 0.2177        | 0.72  | 4500  | 11.08   | 0.2307          | 40.0156 | 14.2972 | 36.0288 | 36.043    |
| 0.2159        | 0.8   | 5000  | 11.336  | 0.2298          | 40.4042 | 15.2531 | 36.6967 | 36.7003   |
| 0.2189        | 0.88  | 5500  | 11.39   | 0.2282          | 40.167  | 14.4847 | 36.3855 | 36.3742   |
| 0.2171        | 0.96  | 6000  | 11.002  | 0.2269          | 40.8528 | 15.1811 | 37.0586 | 37.0403   |
| 0.1962        | 1.04  | 6500  | 11.598  | 0.2296          | 40.6676 | 14.9888 | 36.7796 | 36.7703   |
| 0.1835        | 1.12  | 7000  | 11.022  | 0.2311          | 40.6188 | 15.2743 | 36.8519 | 36.8263   |
| 0.1835        | 1.2   | 7500  | 11.248  | 0.2289          | 40.6466 | 15.1727 | 36.6626 | 36.6427   |
| 0.1864        | 1.28  | 8000  | 11.408  | 0.2298          | 40.2408 | 15.0179 | 36.5594 | 36.5756   |
| 0.1838        | 1.36  | 8500  | 11.238  | 0.2295          | 41.0772 | 15.2152 | 37.0647 | 37.0648   |
| 0.1827        | 1.44  | 9000  | 11.28   | 0.2299          | 40.3263 | 14.9976 | 36.6444 | 36.6292   |
| 0.1828        | 1.52  | 9500  | 11.132  | 0.2299          | 40.9308 | 15.181  | 36.9028 | 36.8909   |
| 0.179         | 1.61  | 10000 | 11.164  | 0.2287          | 40.7406 | 15.2746 | 36.85   | 36.8748   |
| 0.1849        | 1.69  | 10500 | 10.988  | 0.2281          | 40.931  | 15.6479 | 37.0222 | 37.0071   |
| 0.1794        | 1.77  | 11000 | 11.218  | 0.2281          | 41.5198 | 15.9659 | 37.3709 | 37.386    |
| 0.1787        | 1.85  | 11500 | 11.274  | 0.2278          | 40.4006 | 14.9496 | 36.4608 | 36.4675   |
| 0.1798        | 1.93  | 12000 | 11.154  | 0.2279          | 41.3118 | 15.4673 | 37.4917 | 37.5101   |
| 0.1803        | 2.01  | 12500 | 11.23   | 0.2282          | 40.5652 | 15.1467 | 36.7946 | 36.7809   |
| 0.1519        | 2.09  | 13000 | 11.498  | 0.2361          | 40.8978 | 15.0865 | 36.7157 | 36.728    |
| 0.1515        | 2.17  | 13500 | 11.37   | 0.2360          | 40.9809 | 15.5877 | 37.0104 | 36.9942   |
| 0.1519        | 2.25  | 14000 | 11.504  | 0.2359          | 40.7947 | 15.3254 | 36.9574 | 36.9431   |
| 0.1543        | 2.33  | 14500 | 0.2346  | 40.7724         | 15.1837 | 36.9003 | 36.848  | 11.586    |
| 0.1548        | 2.41  | 15000 | 0.2355  | 40.7237         | 15.2394 | 37.0767 | 37.0405 | 11.294    |
| 0.1507        | 2.49  | 15500 | 0.2353  | 41.2661         | 15.7703 | 37.3669 | 37.32   | 11.308    |
| 0.1512        | 2.57  | 16000 | 0.2351  | 40.8777         | 15.2821 | 36.9591 | 36.9201 | 11.43     |
| 0.1525        | 2.65  | 16500 | 0.2350  | 40.6184         | 15.1824 | 36.655  | 36.6117 | 11.402    |
| 0.1522        | 2.73  | 17000 | 0.2343  | 41.2818         | 15.7174 | 37.3059 | 37.2695 | 11.502    |
| 0.1544        | 2.81  | 17500 | 0.2349  | 41.0821         | 15.5164 | 37.2206 | 37.1663 | 11.542    |
| 0.1498        | 2.89  | 18000 | 0.2346  | 41.2128         | 15.6698 | 37.2279 | 37.1874 | 11.582    |
| 0.1497        | 2.97  | 18500 | 0.2344  | 41.2779         | 15.8081 | 37.3177 | 37.2772 | 11.568    |


### Framework versions

- Transformers 4.39.3
- Pytorch 2.1.2
- Datasets 2.18.0
- Tokenizers 0.15.2