pminervini commited on
Commit
2c903ef
1 Parent(s): 3a3ddd1

End of training

Browse files
Files changed (29) hide show
  1. README.md +197 -0
  2. adapter_config.json +40 -0
  3. adapter_model.safetensors +3 -0
  4. added_tokens.json +4 -0
  5. runs/Sep16_12-55-27_cdc6798402b0/events.out.tfevents.1726491346.cdc6798402b0.3380.0 +3 -0
  6. runs/Sep16_12-56-57_cdc6798402b0/events.out.tfevents.1726491434.cdc6798402b0.3968.0 +3 -0
  7. runs/Sep16_13-02-05_cdc6798402b0/events.out.tfevents.1726491740.cdc6798402b0.5510.0 +3 -0
  8. runs/Sep16_13-09-15_cdc6798402b0/events.out.tfevents.1726492172.cdc6798402b0.7468.0 +3 -0
  9. runs/Sep16_13-11-01_cdc6798402b0/events.out.tfevents.1726492277.cdc6798402b0.8037.0 +3 -0
  10. runs/Sep16_13-21-53_cdc6798402b0/events.out.tfevents.1726492928.cdc6798402b0.10873.0 +3 -0
  11. runs/Sep16_14-24-59_cdc6798402b0/events.out.tfevents.1726496715.cdc6798402b0.27092.0 +3 -0
  12. runs/Sep16_14-25-51_cdc6798402b0/events.out.tfevents.1726496767.cdc6798402b0.27488.0 +3 -0
  13. runs/Sep16_14-29-41_cdc6798402b0/events.out.tfevents.1726497026.cdc6798402b0.28610.0 +3 -0
  14. runs/Sep16_14-31-56_cdc6798402b0/events.out.tfevents.1726497142.cdc6798402b0.29301.0 +3 -0
  15. runs/Sep16_14-40-43_cdc6798402b0/events.out.tfevents.1726497659.cdc6798402b0.31660.0 +3 -0
  16. runs/Sep16_14-41-45_cdc6798402b0/events.out.tfevents.1726497720.cdc6798402b0.32028.0 +3 -0
  17. runs/Sep16_14-43-48_cdc6798402b0/events.out.tfevents.1726497844.cdc6798402b0.32661.0 +3 -0
  18. runs/Sep16_14-46-43_cdc6798402b0/events.out.tfevents.1726498019.cdc6798402b0.33526.0 +3 -0
  19. runs/Sep16_14-49-47_cdc6798402b0/events.out.tfevents.1726498203.cdc6798402b0.34427.0 +3 -0
  20. runs/Sep16_14-58-46_cdc6798402b0/events.out.tfevents.1726498794.cdc6798402b0.36902.0 +3 -0
  21. runs/Sep16_15-03-16_cdc6798402b0/events.out.tfevents.1726499012.cdc6798402b0.38166.0 +3 -0
  22. runs/Sep16_15-45-17_cdc6798402b0/events.out.tfevents.1726501533.cdc6798402b0.48991.0 +3 -0
  23. runs/Sep16_16-38-01_cdc6798402b0/events.out.tfevents.1726504696.cdc6798402b0.62062.0 +3 -0
  24. runs/Sep16_16-40-32_cdc6798402b0/events.out.tfevents.1726504847.cdc6798402b0.62831.0 +3 -0
  25. special_tokens_map.json +30 -0
  26. tokenizer.json +0 -0
  27. tokenizer.model +3 -0
  28. tokenizer_config.json +61 -0
  29. training_args.bin +3 -0
README.md ADDED
@@ -0,0 +1,197 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: llava-hf/llava-1.5-7b-hf
3
+ library_name: peft
4
+ license: llama2
5
+ metrics:
6
+ - bleu
7
+ - rouge
8
+ - f1
9
+ tags:
10
+ - trl
11
+ - sft
12
+ - generated_from_trainer
13
+ model-index:
14
+ - name: sft-llava-1.5-7b-hf
15
+ results: []
16
+ ---
17
+
18
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
19
+ should probably proofread and complete it, then remove this comment. -->
20
+
21
+ # sft-llava-1.5-7b-hf
22
+
23
+ This model is a fine-tuned version of [llava-hf/llava-1.5-7b-hf](https://huggingface.co/llava-hf/llava-1.5-7b-hf) on an unknown dataset.
24
+ It achieves the following results on the evaluation set:
25
+ - Loss: 2.2066
26
+ - Bleu: 0.1022
27
+ - Rouge1: 0.4068
28
+ - Rouge2: 0.1449
29
+ - Rougel: 0.3039
30
+ - F1: 0.1633
31
+
32
+ ## Model description
33
+
34
+ More information needed
35
+
36
+ ## Intended uses & limitations
37
+
38
+ More information needed
39
+
40
+ ## Training and evaluation data
41
+
42
+ More information needed
43
+
44
+ ## Training procedure
45
+
46
+ ### Training hyperparameters
47
+
48
+ The following hyperparameters were used during training:
49
+ - learning_rate: 5e-05
50
+ - train_batch_size: 1
51
+ - eval_batch_size: 8
52
+ - seed: 42
53
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
54
+ - lr_scheduler_type: linear
55
+ - training_steps: 128
56
+
57
+ ### Training results
58
+
59
+ | Training Loss | Epoch | Step | Validation Loss | Bleu | Rouge1 | Rouge2 | Rougel | F1 |
60
+ |:-------------:|:------:|:----:|:---------------:|:------:|:------:|:------:|:------:|:------:|
61
+ | 2.4895 | 0.0078 | 1 | 2.8980 | 0.0341 | 0.2763 | 0.0784 | 0.1926 | 0.0606 |
62
+ | 2.7403 | 0.0156 | 2 | 2.8865 | 0.0350 | 0.2778 | 0.0797 | 0.1928 | 0.0621 |
63
+ | 2.5288 | 0.0234 | 3 | 2.8686 | 0.0346 | 0.2783 | 0.0798 | 0.1936 | 0.0615 |
64
+ | 2.7888 | 0.0312 | 4 | 2.8437 | 0.0359 | 0.2787 | 0.0814 | 0.1946 | 0.0636 |
65
+ | 2.6592 | 0.0391 | 5 | 2.8111 | 0.0372 | 0.2799 | 0.0820 | 0.1971 | 0.0657 |
66
+ | 2.682 | 0.0469 | 6 | 2.7774 | 0.0374 | 0.2840 | 0.0835 | 0.1998 | 0.0661 |
67
+ | 2.7822 | 0.0547 | 7 | 2.7344 | 0.0410 | 0.2863 | 0.0863 | 0.2037 | 0.0717 |
68
+ | 2.6998 | 0.0625 | 8 | 2.6884 | 0.0425 | 0.2884 | 0.0899 | 0.2054 | 0.0741 |
69
+ | 2.332 | 0.0703 | 9 | 2.6515 | 0.0466 | 0.2929 | 0.0936 | 0.2109 | 0.0804 |
70
+ | 2.6385 | 0.0781 | 10 | 2.6203 | 0.0478 | 0.2953 | 0.0960 | 0.2118 | 0.0822 |
71
+ | 2.8734 | 0.0859 | 11 | 2.5952 | 0.0496 | 0.2994 | 0.0963 | 0.2134 | 0.0850 |
72
+ | 2.5088 | 0.0938 | 12 | 2.5700 | 0.0542 | 0.3012 | 0.1018 | 0.2173 | 0.0919 |
73
+ | 2.7795 | 0.1016 | 13 | 2.5459 | 0.0571 | 0.3050 | 0.1054 | 0.2213 | 0.0961 |
74
+ | 2.4867 | 0.1094 | 14 | 2.5214 | 0.0585 | 0.3083 | 0.1072 | 0.2229 | 0.0984 |
75
+ | 2.1623 | 0.1172 | 15 | 2.5031 | 0.0591 | 0.3085 | 0.1079 | 0.2240 | 0.0992 |
76
+ | 2.1557 | 0.125 | 16 | 2.4844 | 0.0592 | 0.3127 | 0.1089 | 0.2257 | 0.0995 |
77
+ | 2.0768 | 0.1328 | 17 | 2.4667 | 0.0599 | 0.3129 | 0.1094 | 0.2265 | 0.1006 |
78
+ | 2.2171 | 0.1406 | 18 | 2.4522 | 0.0604 | 0.3170 | 0.1107 | 0.2282 | 0.1015 |
79
+ | 2.2241 | 0.1484 | 19 | 2.4378 | 0.0616 | 0.3196 | 0.1130 | 0.2319 | 0.1032 |
80
+ | 2.3845 | 0.1562 | 20 | 2.4224 | 0.0623 | 0.3258 | 0.1144 | 0.2339 | 0.1046 |
81
+ | 2.0968 | 0.1641 | 21 | 2.4077 | 0.0629 | 0.3294 | 0.1154 | 0.2363 | 0.1056 |
82
+ | 2.1321 | 0.1719 | 22 | 2.3935 | 0.0635 | 0.3332 | 0.1169 | 0.2411 | 0.1067 |
83
+ | 2.0084 | 0.1797 | 23 | 2.3798 | 0.0652 | 0.3397 | 0.1212 | 0.2453 | 0.1094 |
84
+ | 2.0457 | 0.1875 | 24 | 2.3698 | 0.0660 | 0.3423 | 0.1219 | 0.2492 | 0.1107 |
85
+ | 2.0829 | 0.1953 | 25 | 2.3562 | 0.0672 | 0.3466 | 0.1221 | 0.2504 | 0.1125 |
86
+ | 2.1274 | 0.2031 | 26 | 2.3441 | 0.0678 | 0.3510 | 0.1233 | 0.2522 | 0.1137 |
87
+ | 2.5369 | 0.2109 | 27 | 2.3315 | 0.0689 | 0.3558 | 0.1245 | 0.2556 | 0.1155 |
88
+ | 2.3216 | 0.2188 | 28 | 2.3180 | 0.0739 | 0.3621 | 0.1289 | 0.2632 | 0.1228 |
89
+ | 2.2894 | 0.2266 | 29 | 2.3073 | 0.0764 | 0.3671 | 0.1312 | 0.2672 | 0.1265 |
90
+ | 2.3701 | 0.2344 | 30 | 2.2968 | 0.0797 | 0.3755 | 0.1337 | 0.2717 | 0.1314 |
91
+ | 1.9144 | 0.2422 | 31 | 2.2904 | 0.0821 | 0.3764 | 0.1362 | 0.2748 | 0.1348 |
92
+ | 2.1149 | 0.25 | 32 | 2.2838 | 0.0840 | 0.3804 | 0.1362 | 0.2776 | 0.1376 |
93
+ | 1.7937 | 0.2578 | 33 | 2.2795 | 0.0859 | 0.3851 | 0.1386 | 0.2833 | 0.1405 |
94
+ | 1.6068 | 0.2656 | 34 | 2.2743 | 0.0869 | 0.3899 | 0.1390 | 0.2856 | 0.1421 |
95
+ | 1.9267 | 0.2734 | 35 | 2.2709 | 0.0884 | 0.3940 | 0.1404 | 0.2889 | 0.1444 |
96
+ | 2.0678 | 0.2812 | 36 | 2.2670 | 0.0905 | 0.3970 | 0.1418 | 0.2918 | 0.1474 |
97
+ | 1.9654 | 0.2891 | 37 | 2.2626 | 0.0936 | 0.4022 | 0.1449 | 0.2977 | 0.1519 |
98
+ | 2.1679 | 0.2969 | 38 | 2.2597 | 0.0952 | 0.4076 | 0.1467 | 0.2992 | 0.1543 |
99
+ | 2.5372 | 0.3047 | 39 | 2.2561 | 0.0972 | 0.4119 | 0.1491 | 0.3031 | 0.1573 |
100
+ | 2.1875 | 0.3125 | 40 | 2.2524 | 0.0952 | 0.4031 | 0.1461 | 0.2962 | 0.1540 |
101
+ | 2.492 | 0.3203 | 41 | 2.2500 | 0.1009 | 0.4168 | 0.1517 | 0.3082 | 0.1624 |
102
+ | 1.9496 | 0.3281 | 42 | 2.2467 | 0.1026 | 0.4203 | 0.1521 | 0.3104 | 0.1650 |
103
+ | 1.9127 | 0.3359 | 43 | 2.2447 | 0.1047 | 0.4221 | 0.1535 | 0.3125 | 0.1678 |
104
+ | 2.2491 | 0.3438 | 44 | 2.2422 | 0.1042 | 0.4228 | 0.1528 | 0.3139 | 0.1672 |
105
+ | 2.1814 | 0.3516 | 45 | 2.2413 | 0.1067 | 0.4251 | 0.1548 | 0.3150 | 0.1706 |
106
+ | 1.7902 | 0.3594 | 46 | 2.2395 | 0.1075 | 0.4268 | 0.1536 | 0.3166 | 0.1718 |
107
+ | 1.982 | 0.3672 | 47 | 2.2390 | 0.1083 | 0.4249 | 0.1542 | 0.3158 | 0.1727 |
108
+ | 1.9774 | 0.375 | 48 | 2.2373 | 0.1092 | 0.4310 | 0.1572 | 0.3209 | 0.1743 |
109
+ | 1.8207 | 0.3828 | 49 | 2.2377 | 0.1092 | 0.4299 | 0.1569 | 0.3214 | 0.1742 |
110
+ | 2.2484 | 0.3906 | 50 | 2.2361 | 0.1102 | 0.4298 | 0.1560 | 0.3184 | 0.1754 |
111
+ | 2.2954 | 0.3984 | 51 | 2.2352 | 0.1098 | 0.4328 | 0.1550 | 0.3201 | 0.1752 |
112
+ | 2.1899 | 0.4062 | 52 | 2.2344 | 0.1115 | 0.4330 | 0.1584 | 0.3215 | 0.1773 |
113
+ | 2.1295 | 0.4141 | 53 | 2.2338 | 0.1120 | 0.4351 | 0.1588 | 0.3233 | 0.1782 |
114
+ | 2.1473 | 0.4219 | 54 | 2.2323 | 0.1120 | 0.4365 | 0.1580 | 0.3254 | 0.1783 |
115
+ | 1.9628 | 0.4297 | 55 | 2.2315 | 0.1130 | 0.4357 | 0.1581 | 0.3260 | 0.1795 |
116
+ | 2.0926 | 0.4375 | 56 | 2.2315 | 0.1143 | 0.4378 | 0.1592 | 0.3261 | 0.1812 |
117
+ | 1.9722 | 0.4453 | 57 | 2.2301 | 0.1140 | 0.4366 | 0.1583 | 0.3250 | 0.1808 |
118
+ | 2.5585 | 0.4531 | 58 | 2.2298 | 0.1140 | 0.4377 | 0.1579 | 0.3270 | 0.1809 |
119
+ | 1.8927 | 0.4609 | 59 | 2.2297 | 0.1145 | 0.4388 | 0.1587 | 0.3266 | 0.1816 |
120
+ | 1.8164 | 0.4688 | 60 | 2.2289 | 0.1140 | 0.4376 | 0.1570 | 0.3267 | 0.1808 |
121
+ | 2.1107 | 0.4766 | 61 | 2.2279 | 0.1139 | 0.4367 | 0.1570 | 0.3276 | 0.1807 |
122
+ | 1.827 | 0.4844 | 62 | 2.2265 | 0.1157 | 0.4388 | 0.1578 | 0.3275 | 0.1831 |
123
+ | 2.0554 | 0.4922 | 63 | 2.2269 | 0.1162 | 0.4359 | 0.1577 | 0.3251 | 0.1835 |
124
+ | 2.2363 | 0.5 | 64 | 2.2254 | 0.1167 | 0.4397 | 0.1587 | 0.3287 | 0.1845 |
125
+ | 2.1859 | 0.5078 | 65 | 2.2258 | 0.1160 | 0.4395 | 0.1562 | 0.3275 | 0.1836 |
126
+ | 2.0456 | 0.5156 | 66 | 2.2248 | 0.1162 | 0.4374 | 0.1580 | 0.3270 | 0.1836 |
127
+ | 2.085 | 0.5234 | 67 | 2.2250 | 0.1164 | 0.4400 | 0.1583 | 0.3279 | 0.1841 |
128
+ | 1.9723 | 0.5312 | 68 | 2.2243 | 0.1152 | 0.4374 | 0.1567 | 0.3241 | 0.1824 |
129
+ | 2.3812 | 0.5391 | 69 | 2.2233 | 0.1161 | 0.4379 | 0.1589 | 0.3265 | 0.1836 |
130
+ | 1.8786 | 0.5469 | 70 | 2.2236 | 0.1101 | 0.4252 | 0.1526 | 0.3184 | 0.1749 |
131
+ | 2.5123 | 0.5547 | 71 | 2.2225 | 0.1110 | 0.4260 | 0.1536 | 0.3187 | 0.1761 |
132
+ | 1.8613 | 0.5625 | 72 | 2.2231 | 0.1169 | 0.4391 | 0.1591 | 0.3283 | 0.1846 |
133
+ | 2.1822 | 0.5703 | 73 | 2.2221 | 0.1094 | 0.4246 | 0.1524 | 0.3175 | 0.1739 |
134
+ | 1.6759 | 0.5781 | 74 | 2.2215 | 0.1174 | 0.4404 | 0.1605 | 0.3296 | 0.1854 |
135
+ | 1.7215 | 0.5859 | 75 | 2.2216 | 0.1175 | 0.4444 | 0.1589 | 0.3291 | 0.1859 |
136
+ | 2.415 | 0.5938 | 76 | 2.2212 | 0.1127 | 0.4323 | 0.1554 | 0.3224 | 0.1787 |
137
+ | 2.366 | 0.6016 | 77 | 2.2210 | 0.1173 | 0.4470 | 0.1598 | 0.3316 | 0.1859 |
138
+ | 2.1624 | 0.6094 | 78 | 2.2200 | 0.1123 | 0.4322 | 0.1553 | 0.3227 | 0.1783 |
139
+ | 2.422 | 0.6172 | 79 | 2.2204 | 0.1117 | 0.4316 | 0.1547 | 0.3216 | 0.1775 |
140
+ | 1.5925 | 0.625 | 80 | 2.2198 | 0.1190 | 0.4465 | 0.1602 | 0.3330 | 0.1879 |
141
+ | 2.0679 | 0.6328 | 81 | 2.2199 | 0.1119 | 0.4329 | 0.1555 | 0.3224 | 0.1779 |
142
+ | 2.0866 | 0.6406 | 82 | 2.2184 | 0.1189 | 0.4467 | 0.1615 | 0.3326 | 0.1878 |
143
+ | 2.3189 | 0.6484 | 83 | 2.2199 | 0.1180 | 0.4458 | 0.1604 | 0.3324 | 0.1866 |
144
+ | 2.0302 | 0.6562 | 84 | 2.2192 | 0.1184 | 0.4460 | 0.1602 | 0.3320 | 0.1871 |
145
+ | 1.7403 | 0.6641 | 85 | 2.2189 | 0.1181 | 0.4455 | 0.1609 | 0.3323 | 0.1867 |
146
+ | 2.1954 | 0.6719 | 86 | 2.2187 | 0.1199 | 0.4440 | 0.1613 | 0.3315 | 0.1889 |
147
+ | 1.997 | 0.6797 | 87 | 2.2199 | 0.1183 | 0.4418 | 0.1588 | 0.3311 | 0.1867 |
148
+ | 2.4519 | 0.6875 | 88 | 2.2186 | 0.1120 | 0.4320 | 0.1553 | 0.3228 | 0.1779 |
149
+ | 2.2017 | 0.6953 | 89 | 2.2179 | 0.1072 | 0.4179 | 0.1513 | 0.3122 | 0.1706 |
150
+ | 2.1003 | 0.7031 | 90 | 2.2180 | 0.1122 | 0.4324 | 0.1541 | 0.3224 | 0.1782 |
151
+ | 1.7368 | 0.7109 | 91 | 2.2168 | 0.1123 | 0.4303 | 0.1541 | 0.3215 | 0.1781 |
152
+ | 1.8353 | 0.7188 | 92 | 2.2168 | 0.1122 | 0.4287 | 0.1543 | 0.3207 | 0.1778 |
153
+ | 2.1268 | 0.7266 | 93 | 2.2160 | 0.1199 | 0.4436 | 0.1622 | 0.3333 | 0.1887 |
154
+ | 1.9001 | 0.7344 | 94 | 2.2151 | 0.1074 | 0.4164 | 0.1523 | 0.3126 | 0.1707 |
155
+ | 1.8191 | 0.7422 | 95 | 2.2145 | 0.1064 | 0.4165 | 0.1495 | 0.3112 | 0.1695 |
156
+ | 2.1177 | 0.75 | 96 | 2.2136 | 0.1064 | 0.4163 | 0.1498 | 0.3114 | 0.1694 |
157
+ | 2.001 | 0.7578 | 97 | 2.2130 | 0.1069 | 0.4186 | 0.1508 | 0.3128 | 0.1703 |
158
+ | 1.954 | 0.7656 | 98 | 2.2137 | 0.1120 | 0.4318 | 0.1534 | 0.3235 | 0.1778 |
159
+ | 2.4814 | 0.7734 | 99 | 2.2128 | 0.1067 | 0.4178 | 0.1500 | 0.3115 | 0.1699 |
160
+ | 1.6829 | 0.7812 | 100 | 2.2121 | 0.1077 | 0.4171 | 0.1528 | 0.3123 | 0.1712 |
161
+ | 2.1517 | 0.7891 | 101 | 2.2124 | 0.1077 | 0.4174 | 0.1529 | 0.3125 | 0.1713 |
162
+ | 2.1862 | 0.7969 | 102 | 2.2117 | 0.1072 | 0.4171 | 0.1513 | 0.3119 | 0.1706 |
163
+ | 2.0831 | 0.8047 | 103 | 2.2109 | 0.1023 | 0.4048 | 0.1442 | 0.3027 | 0.1634 |
164
+ | 1.8499 | 0.8125 | 104 | 2.2098 | 0.1069 | 0.4172 | 0.1511 | 0.3142 | 0.1702 |
165
+ | 2.2972 | 0.8203 | 105 | 2.2105 | 0.1080 | 0.4189 | 0.1524 | 0.3148 | 0.1718 |
166
+ | 1.7076 | 0.8281 | 106 | 2.2099 | 0.1072 | 0.4171 | 0.1504 | 0.3113 | 0.1705 |
167
+ | 2.1683 | 0.8359 | 107 | 2.2090 | 0.1019 | 0.4036 | 0.1454 | 0.3018 | 0.1627 |
168
+ | 1.7442 | 0.8438 | 108 | 2.2085 | 0.1071 | 0.4201 | 0.1516 | 0.3133 | 0.1707 |
169
+ | 1.4781 | 0.8516 | 109 | 2.2078 | 0.1028 | 0.4033 | 0.1466 | 0.3030 | 0.1639 |
170
+ | 2.2545 | 0.8594 | 110 | 2.2093 | 0.1034 | 0.4047 | 0.1467 | 0.3018 | 0.1647 |
171
+ | 2.1487 | 0.8672 | 111 | 2.2077 | 0.1027 | 0.4050 | 0.1466 | 0.3023 | 0.1638 |
172
+ | 2.4301 | 0.875 | 112 | 2.2083 | 0.1025 | 0.4052 | 0.1462 | 0.3055 | 0.1636 |
173
+ | 1.7237 | 0.8828 | 113 | 2.2082 | 0.1028 | 0.4045 | 0.1469 | 0.3021 | 0.1639 |
174
+ | 2.198 | 0.8906 | 114 | 2.2083 | 0.1038 | 0.4066 | 0.1474 | 0.3051 | 0.1654 |
175
+ | 2.2638 | 0.8984 | 115 | 2.2083 | 0.1026 | 0.4037 | 0.1455 | 0.3032 | 0.1636 |
176
+ | 1.7335 | 0.9062 | 116 | 2.2073 | 0.1028 | 0.4040 | 0.1443 | 0.3028 | 0.1639 |
177
+ | 2.2235 | 0.9141 | 117 | 2.2074 | 0.1076 | 0.4170 | 0.1519 | 0.3116 | 0.1711 |
178
+ | 2.0328 | 0.9219 | 118 | 2.2073 | 0.1028 | 0.4043 | 0.1465 | 0.3032 | 0.1639 |
179
+ | 1.7564 | 0.9297 | 119 | 2.2066 | 0.1086 | 0.4184 | 0.1542 | 0.3144 | 0.1725 |
180
+ | 1.9033 | 0.9375 | 120 | 2.2071 | 0.1034 | 0.4042 | 0.1463 | 0.3034 | 0.1647 |
181
+ | 2.2383 | 0.9453 | 121 | 2.2072 | 0.1076 | 0.4170 | 0.1533 | 0.3131 | 0.1711 |
182
+ | 1.9066 | 0.9531 | 122 | 2.2069 | 0.1123 | 0.4314 | 0.1556 | 0.3233 | 0.1782 |
183
+ | 2.0935 | 0.9609 | 123 | 2.2071 | 0.1031 | 0.4047 | 0.1466 | 0.3018 | 0.1643 |
184
+ | 1.9301 | 0.9688 | 124 | 2.2067 | 0.1034 | 0.4047 | 0.1463 | 0.3043 | 0.1647 |
185
+ | 1.8388 | 0.9766 | 125 | 2.2064 | 0.1025 | 0.4038 | 0.1462 | 0.3029 | 0.1634 |
186
+ | 1.93 | 0.9844 | 126 | 2.2073 | 0.1079 | 0.4209 | 0.1510 | 0.3150 | 0.1718 |
187
+ | 1.669 | 0.9922 | 127 | 2.2074 | 0.1029 | 0.4055 | 0.1459 | 0.3033 | 0.1642 |
188
+ | 1.9271 | 1.0 | 128 | 2.2066 | 0.1022 | 0.4068 | 0.1449 | 0.3039 | 0.1633 |
189
+
190
+
191
+ ### Framework versions
192
+
193
+ - PEFT 0.12.0
194
+ - Transformers 4.44.2
195
+ - Pytorch 2.4.0+cu121
196
+ - Datasets 3.0.0
197
+ - Tokenizers 0.19.1
adapter_config.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "llava-hf/llava-1.5-7b-hf",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.05,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 16,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "k_proj",
24
+ "fc1",
25
+ "linear_1",
26
+ "gate_proj",
27
+ "out_proj",
28
+ "fc2",
29
+ "o_proj",
30
+ "up_proj",
31
+ "lm_head",
32
+ "v_proj",
33
+ "q_proj",
34
+ "down_proj",
35
+ "linear_2"
36
+ ],
37
+ "task_type": "CAUSAL_LM",
38
+ "use_dora": false,
39
+ "use_rslora": false
40
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7140fb28366a275498a5574b3b07096313428b367ead3341eda5b3fc9754ac03
3
+ size 454168240
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<image>": 32000,
3
+ "<pad>": 32001
4
+ }
runs/Sep16_12-55-27_cdc6798402b0/events.out.tfevents.1726491346.cdc6798402b0.3380.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0b4e2c16594abc8c0ff70b77618f105127b74e3d6eaa0b2bb7bcd656ce14960
3
+ size 5803
runs/Sep16_12-56-57_cdc6798402b0/events.out.tfevents.1726491434.cdc6798402b0.3968.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9269c2f4d370d7009d40a7f74670ee59e996e682c782ea8ed2ee794a3c9ff43
3
+ size 5831
runs/Sep16_13-02-05_cdc6798402b0/events.out.tfevents.1726491740.cdc6798402b0.5510.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8683648e2e00fdd2e9d62c666342b0fac3c6d2d34f28c167e6d41ecd3e071a7a
3
+ size 5828
runs/Sep16_13-09-15_cdc6798402b0/events.out.tfevents.1726492172.cdc6798402b0.7468.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b7541cda65313d3989885e369cf5025d4cf7cdf977d7d5a729ab697e163cfdb
3
+ size 5828
runs/Sep16_13-11-01_cdc6798402b0/events.out.tfevents.1726492277.cdc6798402b0.8037.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d28fa3d51914a65746e8d9529de3299f9157879e92269e9fd26b74c50c227b31
3
+ size 5828
runs/Sep16_13-21-53_cdc6798402b0/events.out.tfevents.1726492928.cdc6798402b0.10873.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2eea92dde0abe3c5e876a8096448656e30f6cb5f03e7765452b02e2460056c23
3
+ size 5828
runs/Sep16_14-24-59_cdc6798402b0/events.out.tfevents.1726496715.cdc6798402b0.27092.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c46be56b8ba300bf39461762b1dc74ff0a6b0b89bdaa77a873838b28f7091d5c
3
+ size 5828
runs/Sep16_14-25-51_cdc6798402b0/events.out.tfevents.1726496767.cdc6798402b0.27488.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aefae37d09b4bcc4f2afcfc9727a3234d9f999bba7ebea59fa8213a2f12fe5e4
3
+ size 5828
runs/Sep16_14-29-41_cdc6798402b0/events.out.tfevents.1726497026.cdc6798402b0.28610.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e62246ba1bb6415417c752fe6651e2fb32f7840cd1f09e96ea168e9576dfd63
3
+ size 5827
runs/Sep16_14-31-56_cdc6798402b0/events.out.tfevents.1726497142.cdc6798402b0.29301.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e5f404b264c38627a11b499b9fee41653ee97b7153f16394fe4d4833dce1fd9
3
+ size 5827
runs/Sep16_14-40-43_cdc6798402b0/events.out.tfevents.1726497659.cdc6798402b0.31660.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:177db0a11b6767bb3357693061e976dd77fa283b0db6d4eb0c15ce1eb1338b0c
3
+ size 5620
runs/Sep16_14-41-45_cdc6798402b0/events.out.tfevents.1726497720.cdc6798402b0.32028.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f549dd8a10cc6eeb654fcd931f1561f3de0daf6ac589d92912e508a696b24357
3
+ size 5620
runs/Sep16_14-43-48_cdc6798402b0/events.out.tfevents.1726497844.cdc6798402b0.32661.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:668e283839b9e896d6fd420bd56dec3a129fc8eb412fbfda47da618600cacbbd
3
+ size 5827
runs/Sep16_14-46-43_cdc6798402b0/events.out.tfevents.1726498019.cdc6798402b0.33526.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3e671399ce1ddf9290366d4a2bf8177077fa8767bf463a7661ad8185e9b1a43
3
+ size 5827
runs/Sep16_14-49-47_cdc6798402b0/events.out.tfevents.1726498203.cdc6798402b0.34427.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64c043639e4a1ca9f0a37da2f60f5cb2183b1d966e964b9f60371f7366ace15d
3
+ size 5827
runs/Sep16_14-58-46_cdc6798402b0/events.out.tfevents.1726498794.cdc6798402b0.36902.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e41cfbc01f01e2f126433b30f9fb00a4bdb000300bf970ee9588c82c826b48d
3
+ size 5827
runs/Sep16_15-03-16_cdc6798402b0/events.out.tfevents.1726499012.cdc6798402b0.38166.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c076f5255fa1f3f2689920fc45b9ecb21e847e64aa364f1add0322856a02383
3
+ size 5831
runs/Sep16_15-45-17_cdc6798402b0/events.out.tfevents.1726501533.cdc6798402b0.48991.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6836867627eed5b815a63578db4ecbbb5e5501aa2374686f883d11d22fc0e3f
3
+ size 5831
runs/Sep16_16-38-01_cdc6798402b0/events.out.tfevents.1726504696.cdc6798402b0.62062.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52522edc68830df42c3afff1ae877d76dae2e1e5ab242eeaaa4975c1de7eb5c7
3
+ size 5831
runs/Sep16_16-40-32_cdc6798402b0/events.out.tfevents.1726504847.cdc6798402b0.62831.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fbdd7214b86fe5f16033ed3463121e2a6db9d22cd2c7ecfb1a795bd1302ea8f
3
+ size 97128
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "32000": {
31
+ "content": "<image>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ },
38
+ "32001": {
39
+ "content": "<pad>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": true
45
+ }
46
+ },
47
+ "bos_token": "<s>",
48
+ "clean_up_tokenization_spaces": false,
49
+ "eos_token": "</s>",
50
+ "legacy": false,
51
+ "max_len": null,
52
+ "model_max_length": 1000000000000000019884624838656,
53
+ "pad_token": "<pad>",
54
+ "padding_side": "left",
55
+ "processor_class": "LlavaProcessor",
56
+ "sp_model_kwargs": {},
57
+ "tokenizer_class": "LlamaTokenizer",
58
+ "trust_remote_code": false,
59
+ "unk_token": "<unk>",
60
+ "use_default_system_prompt": false
61
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:554247dc1f77a84d085cfee7dbf4131aacb26d12d8e6a160af04773ffc19f521
3
+ size 5496