File size: 2,981 Bytes
fef289f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1f724fd
 
 
fef289f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c895142
9c9931b
d7c2f6c
26ceaba
d6d8f42
1185ce9
a9cadca
a9d0e82
0486a4e
e7fba8a
4c94305
827f942
2e1dda9
f01a740
51ef68f
73fcf00
75430f1
637ad66
47c2f3c
08a2504
d33fe70
5110552
f06a1ca
97fa167
f1dc6f2
2df5ea5
89a2060
a135491
16afdb3
cce11f0
29bc798
9554628
7ed9137
e89aa60
d2ce971
5209f03
e3d8c6e
6f9ebdc
eb5f966
54f0f49
1f724fd
fef289f
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
---
base_model: Eliac11/tinkNLP
tags:
- generated_from_keras_callback
model-index:
- name: Eliac11/FitModelKTP
  results: []
---

<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->

# Eliac11/FitModelKTP

This model is a fine-tuned version of [Eliac11/tinkNLP](https://huggingface.co/Eliac11/tinkNLP) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.2798
- Validation Loss: 3.0329
- Epoch: 41

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 2e-05, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.005}
- training_precision: float32

### Training results

| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 3.9625     | 4.0902          | 0     |
| 3.7233     | 3.9629          | 1     |
| 3.5004     | 3.8605          | 2     |
| 3.2832     | 3.7674          | 3     |
| 3.0715     | 3.7121          | 4     |
| 2.8942     | 3.6216          | 5     |
| 2.7196     | 3.5805          | 6     |
| 2.5480     | 3.4985          | 7     |
| 2.3730     | 3.4634          | 8     |
| 2.2230     | 3.4004          | 9     |
| 2.0769     | 3.3615          | 10    |
| 1.9424     | 3.3254          | 11    |
| 1.8186     | 3.2929          | 12    |
| 1.7042     | 3.2540          | 13    |
| 1.5870     | 3.2088          | 14    |
| 1.4708     | 3.1860          | 15    |
| 1.3852     | 3.1750          | 16    |
| 1.2858     | 3.1422          | 17    |
| 1.2008     | 3.1238          | 18    |
| 1.1220     | 3.0919          | 19    |
| 1.0540     | 3.0795          | 20    |
| 0.9794     | 3.0708          | 21    |
| 0.9126     | 3.0515          | 22    |
| 0.8535     | 3.0567          | 23    |
| 0.8003     | 3.0344          | 24    |
| 0.7417     | 3.0280          | 25    |
| 0.6913     | 3.0222          | 26    |
| 0.6467     | 3.0232          | 27    |
| 0.6076     | 3.0171          | 28    |
| 0.5714     | 3.0096          | 29    |
| 0.5296     | 3.0148          | 30    |
| 0.4989     | 3.0026          | 31    |
| 0.4666     | 3.0084          | 32    |
| 0.4366     | 3.0061          | 33    |
| 0.4123     | 3.0066          | 34    |
| 0.3904     | 3.0154          | 35    |
| 0.3676     | 3.0024          | 36    |
| 0.3467     | 3.0194          | 37    |
| 0.3249     | 3.0249          | 38    |
| 0.3095     | 3.0149          | 39    |
| 0.2947     | 3.0215          | 40    |
| 0.2798     | 3.0329          | 41    |


### Framework versions

- Transformers 4.38.2
- TensorFlow 2.15.0
- Datasets 2.18.0
- Tokenizers 0.15.2