File size: 9,515 Bytes
7d7d3c5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ca78311
 
 
7d7d3c5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7f9472e
fd70f4c
d8d8f6b
8891e0e
a6b950b
cc26306
2043216
3fbe857
b34c0d1
106596b
e89de12
d33d55c
3fcee11
5fd9384
1136126
19a38c3
8e2c6ea
bd6b5a4
ce7cd31
612470a
3fbc867
fbd6e4d
cc96fb4
756171f
48720d4
56ac9ee
edefcac
d1a4321
a28f048
fe0a43a
185285b
354394b
1e2e431
ffafe20
6c4174b
6d6b830
d70bd97
02784b7
c5b339a
5de5af9
dc8f767
ae3365c
a9dcdb7
3eb5e72
b89a327
0338a3e
c4bd8f4
82f6ad4
7a35ff1
430dac5
6bfeead
426396b
ac8738c
e777dfd
c302d40
11d74a7
90e603a
6951ea0
75e3214
535a074
8272d3b
6960e50
b4f4d61
535e015
fca7c4b
a8ac979
c95e72c
1484567
bf69677
f9a0ae1
31e701a
29b0d91
656cba8
528d3b0
aaf25b1
dc6df3d
28a08c7
791fa37
b6266d6
e88bf1d
c11250e
a9a0168
f89e3cb
1250d26
aeda760
8ab728a
d8282f8
033d9b6
cbe1641
f4a730e
42d5238
d295478
78540a9
4938b0a
2d48387
ad19f46
d7e4d25
2d96d98
f56e332
3f1e186
430a8a2
6e5b377
b57e864
dc13a56
48c8c6a
33baa9b
58407b8
214ee03
a34e11a
16a469b
13a05e0
5394b20
be0accf
c534de4
6f222fa
7355947
bac0793
471bf9e
dc460a2
f5f6c80
c817a99
27f839d
44b5ced
49e3e90
6277811
58e674c
b4c208f
48d13ac
924627f
4469759
18c9eec
948850e
f35a666
7a2da74
1d959fd
f2fbbf0
50e6f80
1a0a726
65c920d
e1e8a64
fe7fed8
a5fee77
562be24
3aeb144
030a9db
75bb586
71cf1b9
f42f235
d8f1a63
6e00289
4ff07e5
7f5f1fd
ea38c90
2cdd5b6
ff3e5c8
48188c9
b75fa45
bdaf93d
2b2b5ff
e1372ee
f710dae
4d053b3
d0c23c5
1bace52
8f06b5b
7564e1f
726be00
e7de98e
ce91200
b0c1e66
4c2d311
1e6a7aa
ad7aa98
c11ae0b
137420f
2e7f44a
61f7465
5e760dc
5ae85b3
f637419
d58aa87
a86d43a
1f960f2
03573d8
2e9d4bf
570c169
936fdcb
5c15acc
a72c548
74bd706
4afe696
f9b30f1
52f98c8
957064d
bac091f
2cda385
53cee62
2cbfda0
6da8263
ca78311
7d7d3c5
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
---
license: apache-2.0
base_model: t5-small
tags:
- generated_from_keras_callback
model-index:
- name: pijarcandra22/NMTIndoBaliT5
  results: []
---

<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->

# pijarcandra22/NMTIndoBaliT5

This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.1551
- Validation Loss: 2.1825
- Epoch: 200

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 1e-04, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01}
- training_precision: float32

### Training results

| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 3.2881     | 2.6852          | 0     |
| 2.7514     | 2.4004          | 1     |
| 2.5012     | 2.2171          | 2     |
| 2.3252     | 2.0959          | 3     |
| 2.1930     | 1.9901          | 4     |
| 2.0837     | 1.9130          | 5     |
| 1.9912     | 1.8452          | 6     |
| 1.9107     | 1.7974          | 7     |
| 1.8459     | 1.7521          | 8     |
| 1.7902     | 1.7165          | 9     |
| 1.7321     | 1.6842          | 10    |
| 1.6811     | 1.6400          | 11    |
| 1.6374     | 1.6230          | 12    |
| 1.5973     | 1.5960          | 13    |
| 1.5588     | 1.5765          | 14    |
| 1.5244     | 1.5589          | 15    |
| 1.4933     | 1.5370          | 16    |
| 1.4588     | 1.5300          | 17    |
| 1.4325     | 1.5107          | 18    |
| 1.4054     | 1.4970          | 19    |
| 1.3730     | 1.4839          | 20    |
| 1.3475     | 1.4789          | 21    |
| 1.3231     | 1.4616          | 22    |
| 1.3035     | 1.4568          | 23    |
| 1.2768     | 1.4489          | 24    |
| 1.2587     | 1.4396          | 25    |
| 1.2380     | 1.4364          | 26    |
| 1.2208     | 1.4273          | 27    |
| 1.2026     | 1.4228          | 28    |
| 1.1755     | 1.4141          | 29    |
| 1.1614     | 1.4062          | 30    |
| 1.1460     | 1.4060          | 31    |
| 1.1289     | 1.3934          | 32    |
| 1.1134     | 1.4007          | 33    |
| 1.0965     | 1.3927          | 34    |
| 1.0818     | 1.3874          | 35    |
| 1.0661     | 1.3921          | 36    |
| 1.0482     | 1.3795          | 37    |
| 1.0345     | 1.3853          | 38    |
| 1.0195     | 1.3835          | 39    |
| 1.0074     | 1.3772          | 40    |
| 0.9890     | 1.3851          | 41    |
| 0.9833     | 1.3724          | 42    |
| 0.9667     | 1.3740          | 43    |
| 0.9561     | 1.3752          | 44    |
| 0.9429     | 1.3673          | 45    |
| 0.9301     | 1.3828          | 46    |
| 0.9141     | 1.3806          | 47    |
| 0.9050     | 1.3772          | 48    |
| 0.8952     | 1.3812          | 49    |
| 0.8809     | 1.3718          | 50    |
| 0.8725     | 1.3825          | 51    |
| 0.8601     | 1.3842          | 52    |
| 0.8488     | 1.3827          | 53    |
| 0.8375     | 1.3920          | 54    |
| 0.8257     | 1.3936          | 55    |
| 0.8184     | 1.3842          | 56    |
| 0.8081     | 1.3846          | 57    |
| 0.7986     | 1.3860          | 58    |
| 0.7883     | 1.3943          | 59    |
| 0.7787     | 1.4004          | 60    |
| 0.7666     | 1.4071          | 61    |
| 0.7554     | 1.4079          | 62    |
| 0.7470     | 1.4038          | 63    |
| 0.7366     | 1.4141          | 64    |
| 0.7279     | 1.4135          | 65    |
| 0.7250     | 1.4111          | 66    |
| 0.7128     | 1.4196          | 67    |
| 0.7042     | 1.4182          | 68    |
| 0.6946     | 1.4378          | 69    |
| 0.6851     | 1.4350          | 70    |
| 0.6764     | 1.4403          | 71    |
| 0.6695     | 1.4474          | 72    |
| 0.6606     | 1.4454          | 73    |
| 0.6565     | 1.4516          | 74    |
| 0.6450     | 1.4595          | 75    |
| 0.6347     | 1.4700          | 76    |
| 0.6287     | 1.4746          | 77    |
| 0.6183     | 1.4813          | 78    |
| 0.6143     | 1.4785          | 79    |
| 0.6053     | 1.4848          | 80    |
| 0.5994     | 1.4777          | 81    |
| 0.5903     | 1.4962          | 82    |
| 0.5828     | 1.5102          | 83    |
| 0.5760     | 1.4957          | 84    |
| 0.5696     | 1.5121          | 85    |
| 0.5637     | 1.5168          | 86    |
| 0.5578     | 1.5183          | 87    |
| 0.5499     | 1.5184          | 88    |
| 0.5396     | 1.5433          | 89    |
| 0.5345     | 1.5411          | 90    |
| 0.5268     | 1.5338          | 91    |
| 0.5220     | 1.5556          | 92    |
| 0.5184     | 1.5489          | 93    |
| 0.5122     | 1.5635          | 94    |
| 0.5014     | 1.5674          | 95    |
| 0.4921     | 1.5773          | 96    |
| 0.4925     | 1.5773          | 97    |
| 0.4821     | 1.5938          | 98    |
| 0.4769     | 1.6013          | 99    |
| 0.4723     | 1.5979          | 100   |
| 0.4692     | 1.6131          | 101   |
| 0.4603     | 1.6247          | 102   |
| 0.4553     | 1.6276          | 103   |
| 0.4476     | 1.6376          | 104   |
| 0.4401     | 1.6390          | 105   |
| 0.4384     | 1.6442          | 106   |
| 0.4305     | 1.6548          | 107   |
| 0.4263     | 1.6617          | 108   |
| 0.4232     | 1.6523          | 109   |
| 0.4185     | 1.6561          | 110   |
| 0.4129     | 1.6779          | 111   |
| 0.4036     | 1.6897          | 112   |
| 0.4005     | 1.6873          | 113   |
| 0.3948     | 1.6987          | 114   |
| 0.3892     | 1.7120          | 115   |
| 0.3859     | 1.7049          | 116   |
| 0.3795     | 1.7241          | 117   |
| 0.3802     | 1.7273          | 118   |
| 0.3731     | 1.7387          | 119   |
| 0.3672     | 1.7447          | 120   |
| 0.3629     | 1.7513          | 121   |
| 0.3607     | 1.7515          | 122   |
| 0.3543     | 1.7585          | 123   |
| 0.3504     | 1.7601          | 124   |
| 0.3477     | 1.7657          | 125   |
| 0.3453     | 1.7733          | 126   |
| 0.3448     | 1.7718          | 127   |
| 0.3390     | 1.7971          | 128   |
| 0.3352     | 1.7929          | 129   |
| 0.3273     | 1.7988          | 130   |
| 0.3250     | 1.8192          | 131   |
| 0.3222     | 1.8220          | 132   |
| 0.3173     | 1.8289          | 133   |
| 0.3171     | 1.8261          | 134   |
| 0.3124     | 1.8415          | 135   |
| 0.3040     | 1.8379          | 136   |
| 0.3040     | 1.8533          | 137   |
| 0.3030     | 1.8511          | 138   |
| 0.2970     | 1.8537          | 139   |
| 0.2938     | 1.8697          | 140   |
| 0.2929     | 1.8730          | 141   |
| 0.2892     | 1.8632          | 142   |
| 0.2816     | 1.8796          | 143   |
| 0.2812     | 1.8870          | 144   |
| 0.2761     | 1.8891          | 145   |
| 0.2731     | 1.9134          | 146   |
| 0.2698     | 1.9100          | 147   |
| 0.2671     | 1.9207          | 148   |
| 0.2639     | 1.9196          | 149   |
| 0.2621     | 1.9130          | 150   |
| 0.2589     | 1.9273          | 151   |
| 0.2558     | 1.9336          | 152   |
| 0.2545     | 1.9355          | 153   |
| 0.2487     | 1.9551          | 154   |
| 0.2493     | 1.9573          | 155   |
| 0.2449     | 1.9552          | 156   |
| 0.2421     | 1.9591          | 157   |
| 0.2405     | 1.9556          | 158   |
| 0.2367     | 1.9807          | 159   |
| 0.2342     | 1.9859          | 160   |
| 0.2316     | 1.9803          | 161   |
| 0.2281     | 1.9853          | 162   |
| 0.2269     | 1.9970          | 163   |
| 0.2250     | 2.0120          | 164   |
| 0.2236     | 2.0107          | 165   |
| 0.2194     | 2.0208          | 166   |
| 0.2183     | 2.0198          | 167   |
| 0.2168     | 2.0265          | 168   |
| 0.2172     | 2.0278          | 169   |
| 0.2117     | 2.0380          | 170   |
| 0.2078     | 2.0448          | 171   |
| 0.2091     | 2.0415          | 172   |
| 0.2065     | 2.0459          | 173   |
| 0.2027     | 2.0597          | 174   |
| 0.1995     | 2.0659          | 175   |
| 0.1980     | 2.0811          | 176   |
| 0.1971     | 2.0704          | 177   |
| 0.1932     | 2.0785          | 178   |
| 0.1892     | 2.0783          | 179   |
| 0.1924     | 2.0742          | 180   |
| 0.1872     | 2.0979          | 181   |
| 0.1858     | 2.0958          | 182   |
| 0.1853     | 2.1005          | 183   |
| 0.1834     | 2.1166          | 184   |
| 0.1810     | 2.1027          | 185   |
| 0.1789     | 2.1151          | 186   |
| 0.1768     | 2.1302          | 187   |
| 0.1768     | 2.1200          | 188   |
| 0.1766     | 2.1399          | 189   |
| 0.1732     | 2.1196          | 190   |
| 0.1719     | 2.1362          | 191   |
| 0.1697     | 2.1447          | 192   |
| 0.1684     | 2.1464          | 193   |
| 0.1699     | 2.1442          | 194   |
| 0.1657     | 2.1492          | 195   |
| 0.1607     | 2.1644          | 196   |
| 0.1603     | 2.1667          | 197   |
| 0.1580     | 2.1715          | 198   |
| 0.1588     | 2.1818          | 199   |
| 0.1551     | 2.1825          | 200   |


### Framework versions

- Transformers 4.38.2
- TensorFlow 2.15.0
- Datasets 2.18.0
- Tokenizers 0.15.2