File size: 2,544 Bytes
7d7d3c5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fe0a43a
 
 
7d7d3c5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7f9472e
fd70f4c
d8d8f6b
8891e0e
a6b950b
cc26306
2043216
3fbe857
b34c0d1
106596b
e89de12
d33d55c
3fcee11
5fd9384
1136126
19a38c3
8e2c6ea
bd6b5a4
ce7cd31
612470a
3fbc867
fbd6e4d
cc96fb4
756171f
48720d4
56ac9ee
edefcac
d1a4321
a28f048
fe0a43a
7d7d3c5
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
---
license: apache-2.0
base_model: t5-small
tags:
- generated_from_keras_callback
model-index:
- name: pijarcandra22/NMTIndoBaliT5
  results: []
---

<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->

# pijarcandra22/NMTIndoBaliT5

This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 1.1614
- Validation Loss: 1.4062
- Epoch: 30

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 1e-04, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01}
- training_precision: float32

### Training results

| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 3.2881     | 2.6852          | 0     |
| 2.7514     | 2.4004          | 1     |
| 2.5012     | 2.2171          | 2     |
| 2.3252     | 2.0959          | 3     |
| 2.1930     | 1.9901          | 4     |
| 2.0837     | 1.9130          | 5     |
| 1.9912     | 1.8452          | 6     |
| 1.9107     | 1.7974          | 7     |
| 1.8459     | 1.7521          | 8     |
| 1.7902     | 1.7165          | 9     |
| 1.7321     | 1.6842          | 10    |
| 1.6811     | 1.6400          | 11    |
| 1.6374     | 1.6230          | 12    |
| 1.5973     | 1.5960          | 13    |
| 1.5588     | 1.5765          | 14    |
| 1.5244     | 1.5589          | 15    |
| 1.4933     | 1.5370          | 16    |
| 1.4588     | 1.5300          | 17    |
| 1.4325     | 1.5107          | 18    |
| 1.4054     | 1.4970          | 19    |
| 1.3730     | 1.4839          | 20    |
| 1.3475     | 1.4789          | 21    |
| 1.3231     | 1.4616          | 22    |
| 1.3035     | 1.4568          | 23    |
| 1.2768     | 1.4489          | 24    |
| 1.2587     | 1.4396          | 25    |
| 1.2380     | 1.4364          | 26    |
| 1.2208     | 1.4273          | 27    |
| 1.2026     | 1.4228          | 28    |
| 1.1755     | 1.4141          | 29    |
| 1.1614     | 1.4062          | 30    |


### Framework versions

- Transformers 4.38.2
- TensorFlow 2.15.0
- Datasets 2.18.0
- Tokenizers 0.15.2