File size: 5,003 Bytes
98d5b3a
 
cc208fa
98d5b3a
 
 
 
 
 
 
 
 
 
 
 
cc208fa
98d5b3a
fd8da39
 
 
d28d1dc
fd8da39
 
 
 
98d5b3a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cc208fa
a2de875
 
 
a22db26
 
 
7226aa1
 
 
5665c3f
 
 
5feadbf
 
 
d28d1dc
 
 
4abc512
 
 
34632d6
 
 
fd8da39
 
 
98d5b3a
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
---
license: apache-2.0
base_model: pakawadeep/mt5-base-finetuned-ctfl
tags:
- generated_from_keras_callback
model-index:
- name: pakawadeep/mt5-base-finetuned-ctfl
  results: []
---

<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->

# pakawadeep/mt5-base-finetuned-ctfl

This model is a fine-tuned version of [pakawadeep/mt5-base-finetuned-ctfl](https://huggingface.co/pakawadeep/mt5-base-finetuned-ctfl) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.3091
- Validation Loss: 1.1147
- Train Rouge1: 8.9816
- Train Rouge2: 1.1881
- Train Rougel: 8.8048
- Train Rougelsum: 8.7871
- Train Gen Len: 11.9604
- Epoch: 27

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 2e-05, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01}
- training_precision: float32

### Training results

| Train Loss | Validation Loss | Train Rouge1 | Train Rouge2 | Train Rougel | Train Rougelsum | Train Gen Len | Epoch |
|:----------:|:---------------:|:------------:|:------------:|:------------:|:---------------:|:-------------:|:-----:|
| 1.1067     | 1.0353          | 7.4965       | 1.6832       | 7.4257       | 7.3904          | 11.8762       | 0     |
| 0.9573     | 1.0010          | 7.9915       | 1.6832       | 7.9208       | 7.7793          | 11.9109       | 1     |
| 0.8858     | 1.0002          | 8.4866       | 2.1782       | 8.2744       | 8.2744          | 11.9158       | 2     |
| 0.8402     | 0.9827          | 8.4866       | 2.1782       | 8.2744       | 8.2744          | 11.9554       | 3     |
| 0.7900     | 0.9961          | 8.4866       | 2.1782       | 8.2744       | 8.2744          | 11.9158       | 4     |
| 0.7646     | 0.9898          | 8.4866       | 2.1782       | 8.2744       | 8.2744          | 11.9505       | 5     |
| 0.7190     | 0.9805          | 8.4866       | 2.1782       | 8.2744       | 8.2744          | 11.9208       | 6     |
| 0.7021     | 0.9683          | 8.4866       | 2.1782       | 8.2744       | 8.2744          | 11.9455       | 7     |
| 0.6613     | 0.9732          | 8.9816       | 2.1782       | 8.7694       | 8.8755          | 11.9703       | 8     |
| 0.6416     | 0.9807          | 8.4866       | 2.1782       | 8.2744       | 8.2744          | 11.9505       | 9     |
| 0.6139     | 0.9771          | 8.4866       | 2.1782       | 8.2744       | 8.2744          | 11.9307       | 10    |
| 0.5864     | 0.9723          | 8.4866       | 2.1782       | 8.2744       | 8.2744          | 11.9505       | 11    |
| 0.5844     | 0.9919          | 8.4866       | 2.1782       | 8.2744       | 8.2744          | 11.9653       | 12    |
| 0.5679     | 1.0097          | 8.4866       | 2.1782       | 8.2744       | 8.2744          | 11.9307       | 13    |
| 0.5329     | 0.9947          | 7.9915       | 1.1881       | 7.8501       | 7.7793          | 11.9554       | 14    |
| 0.5173     | 0.9877          | 8.2037       | 1.6832       | 8.0622       | 8.0269          | 11.9505       | 15    |
| 0.4823     | 0.9955          | 7.7793       | 1.1881       | 7.5318       | 7.5318          | 11.9109       | 16    |
| 0.4626     | 1.0106          | 7.9915       | 1.1881       | 7.8501       | 7.7793          | 11.9703       | 17    |
| 0.4497     | 1.0056          | 7.7793       | 1.1881       | 7.5318       | 7.5318          | 11.9109       | 18    |
| 0.4276     | 1.0341          | 7.7793       | 1.1881       | 7.5318       | 7.5318          | 11.8911       | 19    |
| 0.4056     | 1.0482          | 7.7793       | 1.1881       | 7.5318       | 7.5318          | 11.8960       | 20    |
| 0.4003     | 1.0365          | 8.2390       | 1.1881       | 7.9915       | 8.1683          | 11.9356       | 21    |
| 0.3774     | 1.0646          | 8.2390       | 1.1881       | 7.9915       | 8.1683          | 11.9158       | 22    |
| 0.3668     | 1.0713          | 8.2390       | 1.1881       | 7.9915       | 8.1683          | 11.9158       | 23    |
| 0.3539     | 1.0748          | 8.2390       | 1.1881       | 7.9915       | 8.1683          | 11.9257       | 24    |
| 0.3355     | 1.0859          | 8.0387       | 0.8911       | 7.9208       | 7.9208          | 11.8663       | 25    |
| 0.3247     | 1.0929          | 8.7694       | 1.1881       | 8.5573       | 8.5573          | 11.9356       | 26    |
| 0.3091     | 1.1147          | 8.9816       | 1.1881       | 8.8048       | 8.7871          | 11.9604       | 27    |


### Framework versions

- Transformers 4.38.2
- TensorFlow 2.15.0
- Datasets 2.18.0
- Tokenizers 0.15.2