Kielak2 commited on
Commit
7cdb259
1 Parent(s): 68314d7

End of training

Browse files
README.md CHANGED
@@ -1,5 +1,4 @@
1
  ---
2
- base_model: Kielak2/calculator_model_test
3
  tags:
4
  - generated_from_trainer
5
  model-index:
@@ -12,9 +11,9 @@ should probably proofread and complete it, then remove this comment. -->
12
 
13
  # calculator_model_test
14
 
15
- This model is a fine-tuned version of [Kielak2/calculator_model_test](https://huggingface.co/Kielak2/calculator_model_test) on the None dataset.
16
  It achieves the following results on the evaluation set:
17
- - Loss: 0.2680
18
 
19
  ## Model description
20
 
@@ -46,206 +45,206 @@ The following hyperparameters were used during training:
46
 
47
  | Training Loss | Epoch | Step | Validation Loss |
48
  |:-------------:|:-----:|:----:|:---------------:|
49
- | 0.5001 | 1.0 | 1 | 3.3999 |
50
- | 3.3159 | 2.0 | 2 | 2.5866 |
51
- | 2.3974 | 3.0 | 3 | 2.3937 |
52
- | 2.0971 | 4.0 | 4 | 2.2596 |
53
- | 2.1181 | 5.0 | 5 | 2.1882 |
54
- | 2.0458 | 6.0 | 6 | 1.9196 |
55
- | 1.8142 | 7.0 | 7 | 1.6198 |
56
- | 1.5353 | 8.0 | 8 | 1.3971 |
57
- | 1.3998 | 9.0 | 9 | 1.2959 |
58
- | 1.2674 | 10.0 | 10 | 1.2443 |
59
- | 1.2303 | 11.0 | 11 | 1.2443 |
60
- | 1.242 | 12.0 | 12 | 1.1784 |
61
- | 1.1823 | 13.0 | 13 | 1.1184 |
62
- | 1.1237 | 14.0 | 14 | 1.1045 |
63
- | 1.091 | 15.0 | 15 | 1.0504 |
64
- | 1.0599 | 16.0 | 16 | 1.0647 |
65
- | 1.0767 | 17.0 | 17 | 1.0645 |
66
- | 1.0676 | 18.0 | 18 | 1.0431 |
67
- | 1.035 | 19.0 | 19 | 0.9963 |
68
- | 0.9796 | 20.0 | 20 | 1.0182 |
69
- | 1.0002 | 21.0 | 21 | 1.0211 |
70
- | 1.0016 | 22.0 | 22 | 1.0027 |
71
- | 0.9601 | 23.0 | 23 | 0.9614 |
72
- | 0.9323 | 24.0 | 24 | 0.9258 |
73
- | 0.8823 | 25.0 | 25 | 0.9227 |
74
- | 0.8984 | 26.0 | 26 | 0.9107 |
75
- | 0.9054 | 27.0 | 27 | 0.8881 |
76
- | 0.9126 | 28.0 | 28 | 0.9379 |
77
- | 0.9166 | 29.0 | 29 | 0.9273 |
78
- | 0.9053 | 30.0 | 30 | 0.8998 |
79
- | 0.8671 | 31.0 | 31 | 0.8555 |
80
- | 0.829 | 32.0 | 32 | 0.8551 |
81
- | 0.8544 | 33.0 | 33 | 0.8491 |
82
- | 0.8128 | 34.0 | 34 | 0.8184 |
83
- | 0.7961 | 35.0 | 35 | 0.8312 |
84
- | 0.7854 | 36.0 | 36 | 0.8337 |
85
- | 0.808 | 37.0 | 37 | 0.8184 |
86
- | 0.8211 | 38.0 | 38 | 0.8191 |
87
- | 0.7993 | 39.0 | 39 | 0.7743 |
88
- | 0.7789 | 40.0 | 40 | 0.7454 |
89
- | 0.7924 | 41.0 | 41 | 0.7314 |
90
- | 0.7243 | 42.0 | 42 | 0.8436 |
91
- | 0.7537 | 43.0 | 43 | 0.8050 |
92
- | 0.7622 | 44.0 | 44 | 0.7724 |
93
- | 0.7694 | 45.0 | 45 | 0.7963 |
94
- | 0.7819 | 46.0 | 46 | 0.7872 |
95
- | 0.739 | 47.0 | 47 | 0.8100 |
96
- | 0.7456 | 48.0 | 48 | 0.7989 |
97
- | 0.7214 | 49.0 | 49 | 0.7234 |
98
- | 0.6545 | 50.0 | 50 | 0.6993 |
99
- | 0.6834 | 51.0 | 51 | 0.6556 |
100
- | 0.6664 | 52.0 | 52 | 0.6544 |
101
- | 0.6141 | 53.0 | 53 | 0.6489 |
102
- | 0.5929 | 54.0 | 54 | 0.6268 |
103
- | 0.566 | 55.0 | 55 | 0.6311 |
104
- | 0.6577 | 56.0 | 56 | 0.5828 |
105
- | 0.598 | 57.0 | 57 | 0.6526 |
106
- | 0.6056 | 58.0 | 58 | 0.7250 |
107
- | 0.6204 | 59.0 | 59 | 0.6612 |
108
- | 0.5968 | 60.0 | 60 | 0.5759 |
109
- | 0.5823 | 61.0 | 61 | 0.5836 |
110
- | 0.5986 | 62.0 | 62 | 0.5375 |
111
- | 0.5247 | 63.0 | 63 | 0.5993 |
112
- | 0.5891 | 64.0 | 64 | 0.6175 |
113
- | 0.6142 | 65.0 | 65 | 0.5691 |
114
- | 0.5602 | 66.0 | 66 | 0.5180 |
115
- | 0.5017 | 67.0 | 67 | 0.5726 |
116
- | 0.5304 | 68.0 | 68 | 0.5362 |
117
- | 0.4935 | 69.0 | 69 | 0.5311 |
118
- | 0.5167 | 70.0 | 70 | 0.5698 |
119
- | 0.526 | 71.0 | 71 | 0.5837 |
120
- | 0.5538 | 72.0 | 72 | 0.5436 |
121
- | 0.4825 | 73.0 | 73 | 0.5253 |
122
- | 0.4596 | 74.0 | 74 | 0.5132 |
123
- | 0.4722 | 75.0 | 75 | 0.4970 |
124
- | 0.4662 | 76.0 | 76 | 0.4983 |
125
- | 0.4991 | 77.0 | 77 | 0.4886 |
126
- | 0.4613 | 78.0 | 78 | 0.4791 |
127
- | 0.4589 | 79.0 | 79 | 0.4654 |
128
- | 0.4617 | 80.0 | 80 | 0.4532 |
129
- | 0.4491 | 81.0 | 81 | 0.4617 |
130
- | 0.4471 | 82.0 | 82 | 0.4416 |
131
- | 0.4216 | 83.0 | 83 | 0.4841 |
132
- | 0.4516 | 84.0 | 84 | 0.4817 |
133
- | 0.4372 | 85.0 | 85 | 0.4570 |
134
- | 0.4385 | 86.0 | 86 | 0.4801 |
135
- | 0.4546 | 87.0 | 87 | 0.4929 |
136
- | 0.4381 | 88.0 | 88 | 0.4646 |
137
- | 0.4314 | 89.0 | 89 | 0.4338 |
138
- | 0.3989 | 90.0 | 90 | 0.4458 |
139
- | 0.4442 | 91.0 | 91 | 0.4365 |
140
- | 0.4316 | 92.0 | 92 | 0.4116 |
141
- | 0.4012 | 93.0 | 93 | 0.4157 |
142
- | 0.4116 | 94.0 | 94 | 0.4185 |
143
- | 0.4101 | 95.0 | 95 | 0.4026 |
144
- | 0.3975 | 96.0 | 96 | 0.4144 |
145
- | 0.3985 | 97.0 | 97 | 0.4438 |
146
- | 0.424 | 98.0 | 98 | 0.4383 |
147
- | 0.3901 | 99.0 | 99 | 0.4320 |
148
- | 0.4301 | 100.0 | 100 | 0.4259 |
149
- | 0.428 | 101.0 | 101 | 0.3934 |
150
- | 0.3836 | 102.0 | 102 | 0.4109 |
151
- | 0.3887 | 103.0 | 103 | 0.4203 |
152
- | 0.423 | 104.0 | 104 | 0.3942 |
153
- | 0.3722 | 105.0 | 105 | 0.3682 |
154
- | 0.3909 | 106.0 | 106 | 0.3681 |
155
- | 0.3776 | 107.0 | 107 | 0.3945 |
156
- | 0.392 | 108.0 | 108 | 0.3728 |
157
- | 0.3536 | 109.0 | 109 | 0.3862 |
158
- | 0.4197 | 110.0 | 110 | 0.4024 |
159
- | 0.3988 | 111.0 | 111 | 0.3919 |
160
- | 0.4064 | 112.0 | 112 | 0.4617 |
161
- | 0.4446 | 113.0 | 113 | 0.5055 |
162
- | 0.4482 | 114.0 | 114 | 0.4476 |
163
- | 0.3832 | 115.0 | 115 | 0.3900 |
164
- | 0.3675 | 116.0 | 116 | 0.4018 |
165
- | 0.3782 | 117.0 | 117 | 0.3880 |
166
- | 0.352 | 118.0 | 118 | 0.3790 |
167
- | 0.3458 | 119.0 | 119 | 0.3794 |
168
- | 0.3427 | 120.0 | 120 | 0.3671 |
169
- | 0.3223 | 121.0 | 121 | 0.3703 |
170
- | 0.3161 | 122.0 | 122 | 0.3888 |
171
- | 0.3211 | 123.0 | 123 | 0.4134 |
172
- | 0.3247 | 124.0 | 124 | 0.4017 |
173
- | 0.333 | 125.0 | 125 | 0.3822 |
174
- | 0.3227 | 126.0 | 126 | 0.3792 |
175
- | 0.3264 | 127.0 | 127 | 0.3783 |
176
- | 0.3284 | 128.0 | 128 | 0.3735 |
177
- | 0.3199 | 129.0 | 129 | 0.3614 |
178
- | 0.3344 | 130.0 | 130 | 0.3755 |
179
- | 0.3148 | 131.0 | 131 | 0.3901 |
180
- | 0.3592 | 132.0 | 132 | 0.3819 |
181
- | 0.3358 | 133.0 | 133 | 0.3764 |
182
- | 0.3047 | 134.0 | 134 | 0.3779 |
183
- | 0.3538 | 135.0 | 135 | 0.3580 |
184
- | 0.3257 | 136.0 | 136 | 0.3419 |
185
- | 0.2865 | 137.0 | 137 | 0.3402 |
186
- | 0.3037 | 138.0 | 138 | 0.3470 |
187
- | 0.3098 | 139.0 | 139 | 0.3432 |
188
- | 0.3087 | 140.0 | 140 | 0.3354 |
189
- | 0.2926 | 141.0 | 141 | 0.3469 |
190
- | 0.2811 | 142.0 | 142 | 0.3526 |
191
- | 0.3072 | 143.0 | 143 | 0.3465 |
192
- | 0.3092 | 144.0 | 144 | 0.3487 |
193
- | 0.3048 | 145.0 | 145 | 0.3465 |
194
- | 0.2961 | 146.0 | 146 | 0.3384 |
195
- | 0.3149 | 147.0 | 147 | 0.3383 |
196
- | 0.3147 | 148.0 | 148 | 0.3326 |
197
- | 0.2927 | 149.0 | 149 | 0.3306 |
198
- | 0.2765 | 150.0 | 150 | 0.3331 |
199
- | 0.2755 | 151.0 | 151 | 0.3255 |
200
- | 0.304 | 152.0 | 152 | 0.3027 |
201
- | 0.3011 | 153.0 | 153 | 0.3018 |
202
- | 0.2842 | 154.0 | 154 | 0.3092 |
203
- | 0.2936 | 155.0 | 155 | 0.3037 |
204
- | 0.2852 | 156.0 | 156 | 0.3044 |
205
- | 0.2726 | 157.0 | 157 | 0.3143 |
206
- | 0.2577 | 158.0 | 158 | 0.3330 |
207
- | 0.2904 | 159.0 | 159 | 0.3436 |
208
- | 0.2619 | 160.0 | 160 | 0.3452 |
209
- | 0.276 | 161.0 | 161 | 0.3475 |
210
- | 0.2608 | 162.0 | 162 | 0.3454 |
211
- | 0.2529 | 163.0 | 163 | 0.3336 |
212
- | 0.2685 | 164.0 | 164 | 0.3183 |
213
- | 0.2571 | 165.0 | 165 | 0.3048 |
214
- | 0.2641 | 166.0 | 166 | 0.2957 |
215
- | 0.2549 | 167.0 | 167 | 0.2926 |
216
- | 0.243 | 168.0 | 168 | 0.2904 |
217
- | 0.2574 | 169.0 | 169 | 0.2917 |
218
- | 0.2597 | 170.0 | 170 | 0.2987 |
219
- | 0.2512 | 171.0 | 171 | 0.2979 |
220
- | 0.247 | 172.0 | 172 | 0.2906 |
221
- | 0.2485 | 173.0 | 173 | 0.2851 |
222
- | 0.2512 | 174.0 | 174 | 0.2869 |
223
- | 0.2481 | 175.0 | 175 | 0.2838 |
224
- | 0.268 | 176.0 | 176 | 0.2866 |
225
- | 0.2477 | 177.0 | 177 | 0.2902 |
226
- | 0.2498 | 178.0 | 178 | 0.2963 |
227
- | 0.2535 | 179.0 | 179 | 0.2963 |
228
- | 0.2658 | 180.0 | 180 | 0.2939 |
229
- | 0.2506 | 181.0 | 181 | 0.2943 |
230
- | 0.251 | 182.0 | 182 | 0.2894 |
231
- | 0.2491 | 183.0 | 183 | 0.2818 |
232
- | 0.2484 | 184.0 | 184 | 0.2767 |
233
- | 0.2373 | 185.0 | 185 | 0.2740 |
234
- | 0.2481 | 186.0 | 186 | 0.2718 |
235
- | 0.2438 | 187.0 | 187 | 0.2690 |
236
- | 0.2168 | 188.0 | 188 | 0.2658 |
237
- | 0.237 | 189.0 | 189 | 0.2639 |
238
- | 0.2505 | 190.0 | 190 | 0.2625 |
239
- | 0.2448 | 191.0 | 191 | 0.2622 |
240
- | 0.2366 | 192.0 | 192 | 0.2639 |
241
- | 0.2394 | 193.0 | 193 | 0.2681 |
242
- | 0.2537 | 194.0 | 194 | 0.2727 |
243
- | 0.2259 | 195.0 | 195 | 0.2753 |
244
- | 0.2314 | 196.0 | 196 | 0.2750 |
245
- | 0.2398 | 197.0 | 197 | 0.2730 |
246
- | 0.2515 | 198.0 | 198 | 0.2707 |
247
- | 0.2244 | 199.0 | 199 | 0.2690 |
248
- | 0.2458 | 200.0 | 200 | 0.2680 |
249
 
250
 
251
  ### Framework versions
 
1
  ---
 
2
  tags:
3
  - generated_from_trainer
4
  model-index:
 
11
 
12
  # calculator_model_test
13
 
14
+ This model is a fine-tuned version of [](https://huggingface.co/) on the None dataset.
15
  It achieves the following results on the evaluation set:
16
+ - Loss: 0.0317
17
 
18
  ## Model description
19
 
 
45
 
46
  | Training Loss | Epoch | Step | Validation Loss |
47
  |:-------------:|:-----:|:----:|:---------------:|
48
+ | 3.6147 | 1.0 | 6 | 3.0415 |
49
+ | 2.585 | 2.0 | 12 | 2.0622 |
50
+ | 1.899 | 3.0 | 18 | 1.6836 |
51
+ | 1.6095 | 4.0 | 24 | 1.6248 |
52
+ | 1.5731 | 5.0 | 30 | 1.5728 |
53
+ | 1.5407 | 6.0 | 36 | 1.5594 |
54
+ | 1.5353 | 7.0 | 42 | 1.5163 |
55
+ | 1.4873 | 8.0 | 48 | 1.4470 |
56
+ | 1.4322 | 9.0 | 54 | 1.4274 |
57
+ | 1.3743 | 10.0 | 60 | 1.3462 |
58
+ | 1.3009 | 11.0 | 66 | 1.2124 |
59
+ | 1.1918 | 12.0 | 72 | 1.1226 |
60
+ | 1.1449 | 13.0 | 78 | 1.1215 |
61
+ | 1.0914 | 14.0 | 84 | 1.0471 |
62
+ | 1.0285 | 15.0 | 90 | 0.9795 |
63
+ | 1.0093 | 16.0 | 96 | 1.0062 |
64
+ | 0.9957 | 17.0 | 102 | 0.9296 |
65
+ | 0.944 | 18.0 | 108 | 0.9268 |
66
+ | 0.9194 | 19.0 | 114 | 0.9417 |
67
+ | 0.9014 | 20.0 | 120 | 0.8415 |
68
+ | 0.8789 | 21.0 | 126 | 0.7764 |
69
+ | 0.8155 | 22.0 | 132 | 0.7639 |
70
+ | 0.8593 | 23.0 | 138 | 0.8127 |
71
+ | 0.836 | 24.0 | 144 | 0.7397 |
72
+ | 0.7613 | 25.0 | 150 | 0.7067 |
73
+ | 0.7818 | 26.0 | 156 | 0.7217 |
74
+ | 0.7702 | 27.0 | 162 | 0.7128 |
75
+ | 0.7376 | 28.0 | 168 | 0.7242 |
76
+ | 0.8006 | 29.0 | 174 | 0.7117 |
77
+ | 0.7561 | 30.0 | 180 | 0.6767 |
78
+ | 0.7185 | 31.0 | 186 | 0.6828 |
79
+ | 0.7055 | 32.0 | 192 | 0.6215 |
80
+ | 0.6967 | 33.0 | 198 | 0.6766 |
81
+ | 0.7193 | 34.0 | 204 | 0.6238 |
82
+ | 0.6791 | 35.0 | 210 | 0.5900 |
83
+ | 0.6741 | 36.0 | 216 | 0.6307 |
84
+ | 0.663 | 37.0 | 222 | 0.6012 |
85
+ | 0.6326 | 38.0 | 228 | 0.5944 |
86
+ | 0.6041 | 39.0 | 234 | 0.5459 |
87
+ | 0.617 | 40.0 | 240 | 0.5786 |
88
+ | 0.6369 | 41.0 | 246 | 0.5896 |
89
+ | 0.6243 | 42.0 | 252 | 0.5446 |
90
+ | 0.5921 | 43.0 | 258 | 0.4864 |
91
+ | 0.5529 | 44.0 | 264 | 0.5561 |
92
+ | 0.5757 | 45.0 | 270 | 0.5783 |
93
+ | 0.5919 | 46.0 | 276 | 0.5235 |
94
+ | 0.5509 | 47.0 | 282 | 0.4525 |
95
+ | 0.5229 | 48.0 | 288 | 0.5007 |
96
+ | 0.5871 | 49.0 | 294 | 0.5009 |
97
+ | 0.5793 | 50.0 | 300 | 0.5431 |
98
+ | 0.5922 | 51.0 | 306 | 0.5404 |
99
+ | 0.5539 | 52.0 | 312 | 0.5386 |
100
+ | 0.5785 | 53.0 | 318 | 0.4697 |
101
+ | 0.5528 | 54.0 | 324 | 0.5061 |
102
+ | 0.5047 | 55.0 | 330 | 0.4249 |
103
+ | 0.475 | 56.0 | 336 | 0.4206 |
104
+ | 0.5236 | 57.0 | 342 | 0.5689 |
105
+ | 0.576 | 58.0 | 348 | 0.4258 |
106
+ | 0.4862 | 59.0 | 354 | 0.4070 |
107
+ | 0.4946 | 60.0 | 360 | 0.4136 |
108
+ | 0.4527 | 61.0 | 366 | 0.3848 |
109
+ | 0.4522 | 62.0 | 372 | 0.4288 |
110
+ | 0.5087 | 63.0 | 378 | 0.5660 |
111
+ | 0.5559 | 64.0 | 384 | 0.5371 |
112
+ | 0.5153 | 65.0 | 390 | 0.4595 |
113
+ | 0.4503 | 66.0 | 396 | 0.3648 |
114
+ | 0.4191 | 67.0 | 402 | 0.3787 |
115
+ | 0.4522 | 68.0 | 408 | 0.3469 |
116
+ | 0.4096 | 69.0 | 414 | 0.3622 |
117
+ | 0.4502 | 70.0 | 420 | 0.3613 |
118
+ | 0.4138 | 71.0 | 426 | 0.3700 |
119
+ | 0.3896 | 72.0 | 432 | 0.3920 |
120
+ | 0.4271 | 73.0 | 438 | 0.3354 |
121
+ | 0.4107 | 74.0 | 444 | 0.3193 |
122
+ | 0.391 | 75.0 | 450 | 0.3352 |
123
+ | 0.373 | 76.0 | 456 | 0.3818 |
124
+ | 0.4296 | 77.0 | 462 | 0.3238 |
125
+ | 0.3812 | 78.0 | 468 | 0.3337 |
126
+ | 0.3756 | 79.0 | 474 | 0.3105 |
127
+ | 0.3579 | 80.0 | 480 | 0.3433 |
128
+ | 0.4325 | 81.0 | 486 | 0.3103 |
129
+ | 0.356 | 82.0 | 492 | 0.3060 |
130
+ | 0.3467 | 83.0 | 498 | 0.3780 |
131
+ | 0.3922 | 84.0 | 504 | 0.2863 |
132
+ | 0.3457 | 85.0 | 510 | 0.2865 |
133
+ | 0.3755 | 86.0 | 516 | 0.3041 |
134
+ | 0.3319 | 87.0 | 522 | 0.2777 |
135
+ | 0.3359 | 88.0 | 528 | 0.3803 |
136
+ | 0.4192 | 89.0 | 534 | 0.3473 |
137
+ | 0.3941 | 90.0 | 540 | 0.3745 |
138
+ | 0.3991 | 91.0 | 546 | 0.3331 |
139
+ | 0.3489 | 92.0 | 552 | 0.3579 |
140
+ | 0.3352 | 93.0 | 558 | 0.2947 |
141
+ | 0.3202 | 94.0 | 564 | 0.2416 |
142
+ | 0.3339 | 95.0 | 570 | 0.3635 |
143
+ | 0.4108 | 96.0 | 576 | 0.2779 |
144
+ | 0.3827 | 97.0 | 582 | 0.2846 |
145
+ | 0.3559 | 98.0 | 588 | 0.2754 |
146
+ | 0.2985 | 99.0 | 594 | 0.2107 |
147
+ | 0.264 | 100.0 | 600 | 0.1958 |
148
+ | 0.2807 | 101.0 | 606 | 0.2028 |
149
+ | 0.2861 | 102.0 | 612 | 0.2034 |
150
+ | 0.2661 | 103.0 | 618 | 0.1979 |
151
+ | 0.264 | 104.0 | 624 | 0.2134 |
152
+ | 0.2747 | 105.0 | 630 | 0.1754 |
153
+ | 0.2785 | 106.0 | 636 | 0.2329 |
154
+ | 0.2656 | 107.0 | 642 | 0.1934 |
155
+ | 0.2505 | 108.0 | 648 | 0.2213 |
156
+ | 0.2572 | 109.0 | 654 | 0.2313 |
157
+ | 0.2929 | 110.0 | 660 | 0.2308 |
158
+ | 0.2419 | 111.0 | 666 | 0.1780 |
159
+ | 0.239 | 112.0 | 672 | 0.1694 |
160
+ | 0.2279 | 113.0 | 678 | 0.1580 |
161
+ | 0.2528 | 114.0 | 684 | 0.3002 |
162
+ | 0.3297 | 115.0 | 690 | 0.2676 |
163
+ | 0.3147 | 116.0 | 696 | 0.3287 |
164
+ | 0.2826 | 117.0 | 702 | 0.1475 |
165
+ | 0.2033 | 118.0 | 708 | 0.1359 |
166
+ | 0.1938 | 119.0 | 714 | 0.1592 |
167
+ | 0.2105 | 120.0 | 720 | 0.1696 |
168
+ | 0.2196 | 121.0 | 726 | 0.1532 |
169
+ | 0.2102 | 122.0 | 732 | 0.1157 |
170
+ | 0.2014 | 123.0 | 738 | 0.1835 |
171
+ | 0.2505 | 124.0 | 744 | 0.1851 |
172
+ | 0.2411 | 125.0 | 750 | 0.2881 |
173
+ | 0.2353 | 126.0 | 756 | 0.1911 |
174
+ | 0.2268 | 127.0 | 762 | 0.1874 |
175
+ | 0.2024 | 128.0 | 768 | 0.1613 |
176
+ | 0.2046 | 129.0 | 774 | 0.1938 |
177
+ | 0.199 | 130.0 | 780 | 0.1129 |
178
+ | 0.1703 | 131.0 | 786 | 0.1511 |
179
+ | 0.1924 | 132.0 | 792 | 0.1744 |
180
+ | 0.1854 | 133.0 | 798 | 0.1238 |
181
+ | 0.1632 | 134.0 | 804 | 0.1050 |
182
+ | 0.1589 | 135.0 | 810 | 0.1316 |
183
+ | 0.1787 | 136.0 | 816 | 0.0895 |
184
+ | 0.1658 | 137.0 | 822 | 0.0836 |
185
+ | 0.141 | 138.0 | 828 | 0.1087 |
186
+ | 0.1671 | 139.0 | 834 | 0.1068 |
187
+ | 0.1557 | 140.0 | 840 | 0.0800 |
188
+ | 0.1488 | 141.0 | 846 | 0.1277 |
189
+ | 0.1709 | 142.0 | 852 | 0.1126 |
190
+ | 0.1499 | 143.0 | 858 | 0.0913 |
191
+ | 0.1597 | 144.0 | 864 | 0.0829 |
192
+ | 0.1314 | 145.0 | 870 | 0.0762 |
193
+ | 0.1501 | 146.0 | 876 | 0.0897 |
194
+ | 0.156 | 147.0 | 882 | 0.0902 |
195
+ | 0.1482 | 148.0 | 888 | 0.0903 |
196
+ | 0.1401 | 149.0 | 894 | 0.0749 |
197
+ | 0.1322 | 150.0 | 900 | 0.0781 |
198
+ | 0.1309 | 151.0 | 906 | 0.0719 |
199
+ | 0.1326 | 152.0 | 912 | 0.0691 |
200
+ | 0.1311 | 153.0 | 918 | 0.0701 |
201
+ | 0.1202 | 154.0 | 924 | 0.0742 |
202
+ | 0.1258 | 155.0 | 930 | 0.0728 |
203
+ | 0.1183 | 156.0 | 936 | 0.0566 |
204
+ | 0.1181 | 157.0 | 942 | 0.0541 |
205
+ | 0.1137 | 158.0 | 948 | 0.0662 |
206
+ | 0.1061 | 159.0 | 954 | 0.0662 |
207
+ | 0.1121 | 160.0 | 960 | 0.0628 |
208
+ | 0.1038 | 161.0 | 966 | 0.0609 |
209
+ | 0.1135 | 162.0 | 972 | 0.0728 |
210
+ | 0.1317 | 163.0 | 978 | 0.0785 |
211
+ | 0.1149 | 164.0 | 984 | 0.0753 |
212
+ | 0.1111 | 165.0 | 990 | 0.0647 |
213
+ | 0.0926 | 166.0 | 996 | 0.0592 |
214
+ | 0.0931 | 167.0 | 1002 | 0.0554 |
215
+ | 0.0865 | 168.0 | 1008 | 0.0480 |
216
+ | 0.0881 | 169.0 | 1014 | 0.0498 |
217
+ | 0.0932 | 170.0 | 1020 | 0.0524 |
218
+ | 0.0934 | 171.0 | 1026 | 0.0629 |
219
+ | 0.1054 | 172.0 | 1032 | 0.0561 |
220
+ | 0.0933 | 173.0 | 1038 | 0.0422 |
221
+ | 0.0812 | 174.0 | 1044 | 0.0605 |
222
+ | 0.0953 | 175.0 | 1050 | 0.0485 |
223
+ | 0.0963 | 176.0 | 1056 | 0.0394 |
224
+ | 0.0731 | 177.0 | 1062 | 0.0378 |
225
+ | 0.0758 | 178.0 | 1068 | 0.0394 |
226
+ | 0.0703 | 179.0 | 1074 | 0.0406 |
227
+ | 0.0756 | 180.0 | 1080 | 0.0427 |
228
+ | 0.0812 | 181.0 | 1086 | 0.0538 |
229
+ | 0.0842 | 182.0 | 1092 | 0.0434 |
230
+ | 0.0773 | 183.0 | 1098 | 0.0439 |
231
+ | 0.073 | 184.0 | 1104 | 0.0379 |
232
+ | 0.0707 | 185.0 | 1110 | 0.0422 |
233
+ | 0.0749 | 186.0 | 1116 | 0.0420 |
234
+ | 0.0746 | 187.0 | 1122 | 0.0388 |
235
+ | 0.068 | 188.0 | 1128 | 0.0386 |
236
+ | 0.0654 | 189.0 | 1134 | 0.0378 |
237
+ | 0.0647 | 190.0 | 1140 | 0.0335 |
238
+ | 0.0629 | 191.0 | 1146 | 0.0402 |
239
+ | 0.0642 | 192.0 | 1152 | 0.0344 |
240
+ | 0.063 | 193.0 | 1158 | 0.0374 |
241
+ | 0.0631 | 194.0 | 1164 | 0.0321 |
242
+ | 0.0605 | 195.0 | 1170 | 0.0356 |
243
+ | 0.065 | 196.0 | 1176 | 0.0334 |
244
+ | 0.0591 | 197.0 | 1182 | 0.0321 |
245
+ | 0.0558 | 198.0 | 1188 | 0.0317 |
246
+ | 0.0596 | 199.0 | 1194 | 0.0316 |
247
+ | 0.06 | 200.0 | 1200 | 0.0317 |
248
 
249
 
250
  ### Framework versions
config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_name_or_path": "Kielak2/calculator_model_test",
3
  "architectures": [
4
  "EncoderDecoderModel"
5
  ],
@@ -79,7 +78,7 @@
79
  "typical_p": 1.0,
80
  "use_bfloat16": false,
81
  "use_cache": true,
82
- "vocab_size": 34
83
  },
84
  "decoder_start_token_id": 2,
85
  "encoder": {
@@ -158,7 +157,7 @@
158
  "typical_p": 1.0,
159
  "use_bfloat16": false,
160
  "use_cache": true,
161
- "vocab_size": 34
162
  },
163
  "eos_token_id": 0,
164
  "is_encoder_decoder": true,
 
1
  {
 
2
  "architectures": [
3
  "EncoderDecoderModel"
4
  ],
 
78
  "typical_p": 1.0,
79
  "use_bfloat16": false,
80
  "use_cache": true,
81
+ "vocab_size": 52
82
  },
83
  "decoder_start_token_id": 2,
84
  "encoder": {
 
157
  "typical_p": 1.0,
158
  "use_bfloat16": false,
159
  "use_cache": true,
160
+ "vocab_size": 52
161
  },
162
  "eos_token_id": 0,
163
  "is_encoder_decoder": true,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:97b31f1d11f35ac6b9077eeb87b7858a91521a1949ad3fea8ed167f1dc9df118
3
- size 31168616
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51120f4f2cea38123ca289b764c35b52f8b31bfb6c4f21b3d79654f167659770
3
+ size 31205552
runs/Mar04_10-28-21_c60a5c456cbd/events.out.tfevents.1709548101.c60a5c456cbd.796.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a71d31afeef2bea511c8cbad305b30f5ef69f9b7ba5117dc901c987f77229131
3
+ size 105117
tokenizer.json CHANGED
@@ -112,44 +112,80 @@
112
  "7": 13,
113
  "8": 14,
114
  "9": 15,
115
- "10": 16,
116
- "99": 17,
117
- "11": 18,
118
- "98": 19,
119
- "12": 20,
120
- "97": 21,
121
- "13": 22,
122
- "96": 23,
123
- "14": 24,
124
- "95": 25,
125
- "15": 26,
126
- "94": 27,
127
  "16": 28,
128
  "93": 29,
129
- "17": 30,
130
- "92": 31,
131
- "18": 32,
132
- "91": 33
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
133
  },
134
  "merges": [
135
- "1 0",
136
  "9 9",
137
- "1 1",
138
  "9 8",
139
- "1 2",
140
  "9 7",
141
- "1 3",
142
  "9 6",
143
- "1 4",
144
  "9 5",
145
- "1 5",
146
  "9 4",
 
147
  "1 6",
148
  "9 3",
149
- "1 7",
150
  "9 2",
 
 
151
  "1 8",
152
- "9 1"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
153
  ]
154
  }
155
  }
 
112
  "7": 13,
113
  "8": 14,
114
  "9": 15,
115
+ "99": 16,
116
+ "10": 17,
117
+ "98": 18,
118
+ "11": 19,
119
+ "97": 20,
120
+ "12": 21,
121
+ "96": 22,
122
+ "13": 23,
123
+ "95": 24,
124
+ "14": 25,
125
+ "94": 26,
126
+ "15": 27,
127
  "16": 28,
128
  "93": 29,
129
+ "92": 30,
130
+ "17": 31,
131
+ "91": 32,
132
+ "18": 33,
133
+ "90": 34,
134
+ "19": 35,
135
+ "20": 36,
136
+ "89": 37,
137
+ "88": 38,
138
+ "21": 39,
139
+ "22": 40,
140
+ "87": 41,
141
+ "23": 42,
142
+ "86": 43,
143
+ "24": 44,
144
+ "85": 45,
145
+ "25": 46,
146
+ "84": 47,
147
+ "26": 48,
148
+ "83": 49,
149
+ "27": 50,
150
+ "82": 51
151
  },
152
  "merges": [
 
153
  "9 9",
154
+ "1 0",
155
  "9 8",
156
+ "1 1",
157
  "9 7",
158
+ "1 2",
159
  "9 6",
160
+ "1 3",
161
  "9 5",
162
+ "1 4",
163
  "9 4",
164
+ "1 5",
165
  "1 6",
166
  "9 3",
 
167
  "9 2",
168
+ "1 7",
169
+ "9 1",
170
  "1 8",
171
+ "9 0",
172
+ "1 9",
173
+ "2 0",
174
+ "8 9",
175
+ "8 8",
176
+ "2 1",
177
+ "2 2",
178
+ "8 7",
179
+ "2 3",
180
+ "8 6",
181
+ "2 4",
182
+ "8 5",
183
+ "2 5",
184
+ "8 4",
185
+ "2 6",
186
+ "8 3",
187
+ "2 7",
188
+ "8 2"
189
  ]
190
  }
191
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cebe9638621e159942e15d5eff9381d25a3a2c37f9462c3e4d00b0b340bd9fe6
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7e53b1c89424e1de538277509f2a24e580efb76210b3761752044d4694ee8fe
3
  size 5112