File size: 5,004 Bytes
7d7d3c5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f4a730e
 
 
7d7d3c5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7f9472e
fd70f4c
d8d8f6b
8891e0e
a6b950b
cc26306
2043216
3fbe857
b34c0d1
106596b
e89de12
d33d55c
3fcee11
5fd9384
1136126
19a38c3
8e2c6ea
bd6b5a4
ce7cd31
612470a
3fbc867
fbd6e4d
cc96fb4
756171f
48720d4
56ac9ee
edefcac
d1a4321
a28f048
fe0a43a
185285b
354394b
1e2e431
ffafe20
6c4174b
6d6b830
d70bd97
02784b7
c5b339a
5de5af9
dc8f767
ae3365c
a9dcdb7
3eb5e72
b89a327
0338a3e
c4bd8f4
82f6ad4
7a35ff1
430dac5
6bfeead
426396b
ac8738c
e777dfd
c302d40
11d74a7
90e603a
6951ea0
75e3214
535a074
8272d3b
6960e50
b4f4d61
535e015
fca7c4b
a8ac979
c95e72c
1484567
bf69677
f9a0ae1
31e701a
29b0d91
656cba8
528d3b0
aaf25b1
dc6df3d
28a08c7
791fa37
b6266d6
e88bf1d
c11250e
a9a0168
f89e3cb
1250d26
aeda760
8ab728a
d8282f8
033d9b6
cbe1641
f4a730e
7d7d3c5
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
---
license: apache-2.0
base_model: t5-small
tags:
- generated_from_keras_callback
model-index:
- name: pijarcandra22/NMTIndoBaliT5
  results: []
---

<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->

# pijarcandra22/NMTIndoBaliT5

This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.5345
- Validation Loss: 1.5411
- Epoch: 90

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 1e-04, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01}
- training_precision: float32

### Training results

| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 3.2881     | 2.6852          | 0     |
| 2.7514     | 2.4004          | 1     |
| 2.5012     | 2.2171          | 2     |
| 2.3252     | 2.0959          | 3     |
| 2.1930     | 1.9901          | 4     |
| 2.0837     | 1.9130          | 5     |
| 1.9912     | 1.8452          | 6     |
| 1.9107     | 1.7974          | 7     |
| 1.8459     | 1.7521          | 8     |
| 1.7902     | 1.7165          | 9     |
| 1.7321     | 1.6842          | 10    |
| 1.6811     | 1.6400          | 11    |
| 1.6374     | 1.6230          | 12    |
| 1.5973     | 1.5960          | 13    |
| 1.5588     | 1.5765          | 14    |
| 1.5244     | 1.5589          | 15    |
| 1.4933     | 1.5370          | 16    |
| 1.4588     | 1.5300          | 17    |
| 1.4325     | 1.5107          | 18    |
| 1.4054     | 1.4970          | 19    |
| 1.3730     | 1.4839          | 20    |
| 1.3475     | 1.4789          | 21    |
| 1.3231     | 1.4616          | 22    |
| 1.3035     | 1.4568          | 23    |
| 1.2768     | 1.4489          | 24    |
| 1.2587     | 1.4396          | 25    |
| 1.2380     | 1.4364          | 26    |
| 1.2208     | 1.4273          | 27    |
| 1.2026     | 1.4228          | 28    |
| 1.1755     | 1.4141          | 29    |
| 1.1614     | 1.4062          | 30    |
| 1.1460     | 1.4060          | 31    |
| 1.1289     | 1.3934          | 32    |
| 1.1134     | 1.4007          | 33    |
| 1.0965     | 1.3927          | 34    |
| 1.0818     | 1.3874          | 35    |
| 1.0661     | 1.3921          | 36    |
| 1.0482     | 1.3795          | 37    |
| 1.0345     | 1.3853          | 38    |
| 1.0195     | 1.3835          | 39    |
| 1.0074     | 1.3772          | 40    |
| 0.9890     | 1.3851          | 41    |
| 0.9833     | 1.3724          | 42    |
| 0.9667     | 1.3740          | 43    |
| 0.9561     | 1.3752          | 44    |
| 0.9429     | 1.3673          | 45    |
| 0.9301     | 1.3828          | 46    |
| 0.9141     | 1.3806          | 47    |
| 0.9050     | 1.3772          | 48    |
| 0.8952     | 1.3812          | 49    |
| 0.8809     | 1.3718          | 50    |
| 0.8725     | 1.3825          | 51    |
| 0.8601     | 1.3842          | 52    |
| 0.8488     | 1.3827          | 53    |
| 0.8375     | 1.3920          | 54    |
| 0.8257     | 1.3936          | 55    |
| 0.8184     | 1.3842          | 56    |
| 0.8081     | 1.3846          | 57    |
| 0.7986     | 1.3860          | 58    |
| 0.7883     | 1.3943          | 59    |
| 0.7787     | 1.4004          | 60    |
| 0.7666     | 1.4071          | 61    |
| 0.7554     | 1.4079          | 62    |
| 0.7470     | 1.4038          | 63    |
| 0.7366     | 1.4141          | 64    |
| 0.7279     | 1.4135          | 65    |
| 0.7250     | 1.4111          | 66    |
| 0.7128     | 1.4196          | 67    |
| 0.7042     | 1.4182          | 68    |
| 0.6946     | 1.4378          | 69    |
| 0.6851     | 1.4350          | 70    |
| 0.6764     | 1.4403          | 71    |
| 0.6695     | 1.4474          | 72    |
| 0.6606     | 1.4454          | 73    |
| 0.6565     | 1.4516          | 74    |
| 0.6450     | 1.4595          | 75    |
| 0.6347     | 1.4700          | 76    |
| 0.6287     | 1.4746          | 77    |
| 0.6183     | 1.4813          | 78    |
| 0.6143     | 1.4785          | 79    |
| 0.6053     | 1.4848          | 80    |
| 0.5994     | 1.4777          | 81    |
| 0.5903     | 1.4962          | 82    |
| 0.5828     | 1.5102          | 83    |
| 0.5760     | 1.4957          | 84    |
| 0.5696     | 1.5121          | 85    |
| 0.5637     | 1.5168          | 86    |
| 0.5578     | 1.5183          | 87    |
| 0.5499     | 1.5184          | 88    |
| 0.5396     | 1.5433          | 89    |
| 0.5345     | 1.5411          | 90    |


### Framework versions

- Transformers 4.38.2
- TensorFlow 2.15.0
- Datasets 2.18.0
- Tokenizers 0.15.2