File size: 2,981 Bytes
fef289f 1f724fd fef289f c895142 9c9931b d7c2f6c 26ceaba d6d8f42 1185ce9 a9cadca a9d0e82 0486a4e e7fba8a 4c94305 827f942 2e1dda9 f01a740 51ef68f 73fcf00 75430f1 637ad66 47c2f3c 08a2504 d33fe70 5110552 f06a1ca 97fa167 f1dc6f2 2df5ea5 89a2060 a135491 16afdb3 cce11f0 29bc798 9554628 7ed9137 e89aa60 d2ce971 5209f03 e3d8c6e 6f9ebdc eb5f966 54f0f49 1f724fd fef289f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 |
---
base_model: Eliac11/tinkNLP
tags:
- generated_from_keras_callback
model-index:
- name: Eliac11/FitModelKTP
results: []
---
<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->
# Eliac11/FitModelKTP
This model is a fine-tuned version of [Eliac11/tinkNLP](https://huggingface.co/Eliac11/tinkNLP) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.2798
- Validation Loss: 3.0329
- Epoch: 41
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 2e-05, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.005}
- training_precision: float32
### Training results
| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 3.9625 | 4.0902 | 0 |
| 3.7233 | 3.9629 | 1 |
| 3.5004 | 3.8605 | 2 |
| 3.2832 | 3.7674 | 3 |
| 3.0715 | 3.7121 | 4 |
| 2.8942 | 3.6216 | 5 |
| 2.7196 | 3.5805 | 6 |
| 2.5480 | 3.4985 | 7 |
| 2.3730 | 3.4634 | 8 |
| 2.2230 | 3.4004 | 9 |
| 2.0769 | 3.3615 | 10 |
| 1.9424 | 3.3254 | 11 |
| 1.8186 | 3.2929 | 12 |
| 1.7042 | 3.2540 | 13 |
| 1.5870 | 3.2088 | 14 |
| 1.4708 | 3.1860 | 15 |
| 1.3852 | 3.1750 | 16 |
| 1.2858 | 3.1422 | 17 |
| 1.2008 | 3.1238 | 18 |
| 1.1220 | 3.0919 | 19 |
| 1.0540 | 3.0795 | 20 |
| 0.9794 | 3.0708 | 21 |
| 0.9126 | 3.0515 | 22 |
| 0.8535 | 3.0567 | 23 |
| 0.8003 | 3.0344 | 24 |
| 0.7417 | 3.0280 | 25 |
| 0.6913 | 3.0222 | 26 |
| 0.6467 | 3.0232 | 27 |
| 0.6076 | 3.0171 | 28 |
| 0.5714 | 3.0096 | 29 |
| 0.5296 | 3.0148 | 30 |
| 0.4989 | 3.0026 | 31 |
| 0.4666 | 3.0084 | 32 |
| 0.4366 | 3.0061 | 33 |
| 0.4123 | 3.0066 | 34 |
| 0.3904 | 3.0154 | 35 |
| 0.3676 | 3.0024 | 36 |
| 0.3467 | 3.0194 | 37 |
| 0.3249 | 3.0249 | 38 |
| 0.3095 | 3.0149 | 39 |
| 0.2947 | 3.0215 | 40 |
| 0.2798 | 3.0329 | 41 |
### Framework versions
- Transformers 4.38.2
- TensorFlow 2.15.0
- Datasets 2.18.0
- Tokenizers 0.15.2
|