File size: 5,719 Bytes
b0117c3 ba58351 b0117c3 b754b6e b0117c3 f2f2b8d fb70783 6326f56 a04e06d 86e00f3 4fbcdf4 c123b62 b46fccd 537c3e4 3c83d38 ed88514 9b2e5d1 120b37a c174fea 029a0c8 c638760 9301a78 f00a780 138d97b d159ff5 fcd40cb 8733b66 2c83dab 0999306 f6577f5 e175563 fb8c9b2 cac0f79 0795ef0 c28886a 10c1f9c 7e3096e 4f112cc 7f11759 601ba53 6859b05 29d2ca4 fe8e8a5 14dd4b6 84de246 d5717e3 fb89b3b 9d4cd13 ffba5cd d38a425 9e85803 0e4809e 7318341 e1899d2 1985447 7812050 3195389 b6da3d2 47b8f2f 9818f05 91ca895 ee2e18e fd3116c bba29a8 e6dcd7d 32017ae 848481a 55b9b29 aa27402 c7af71d 1092d8c 00a9e04 62c206b 64c076f 3900d86 5c5a92a 67b1b82 450bcb6 027e351 9e9a92b 4e904b3 7c46023 693f498 fc327f4 c15184d 7b02d84 92e2b68 95ffee9 a016bb0 81cd69c abd45a0 b82154e 3b763c5 78898a6 c050ce5 9e9e787 b25dffd 3363c0d f181e42 269f3bf 6a9a9bd 81e40bc b3191a8 b754b6e b0117c3 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 |
---
license: apache-2.0
base_model: Helsinki-NLP/opus-mt-id-en
tags:
- generated_from_keras_callback
model-index:
- name: aditnnda/machine_translation_informal2formal
results: []
---
<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->
# aditnnda/machine_translation_informal2formal
This model is a fine-tuned version of [Helsinki-NLP/opus-mt-id-en](https://huggingface.co/Helsinki-NLP/opus-mt-id-en) on [STIF Indonesia](haryoaw/stif-indonesia) dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.0077
- Validation Loss: 1.2870
- Epoch: 99
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 5e-05, 'decay_steps': 6000, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01}
- training_precision: float32
### Training results
| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 3.4298 | 2.4070 | 0 |
| 2.1508 | 1.8031 | 1 |
| 1.6301 | 1.5249 | 2 |
| 1.3013 | 1.3417 | 3 |
| 1.0752 | 1.2465 | 4 |
| 0.9119 | 1.1651 | 5 |
| 0.7778 | 1.1213 | 6 |
| 0.6763 | 1.0813 | 7 |
| 0.5907 | 1.0542 | 8 |
| 0.5162 | 1.0289 | 9 |
| 0.4573 | 1.0265 | 10 |
| 0.4057 | 1.0115 | 11 |
| 0.3645 | 1.0096 | 12 |
| 0.3227 | 1.0037 | 13 |
| 0.2864 | 1.0016 | 14 |
| 0.2598 | 1.0121 | 15 |
| 0.2291 | 1.0079 | 16 |
| 0.2069 | 1.0199 | 17 |
| 0.1876 | 1.0247 | 18 |
| 0.1717 | 1.0199 | 19 |
| 0.1544 | 1.0283 | 20 |
| 0.1393 | 1.0416 | 21 |
| 0.1285 | 1.0370 | 22 |
| 0.1171 | 1.0430 | 23 |
| 0.1069 | 1.0593 | 24 |
| 0.0990 | 1.0670 | 25 |
| 0.0915 | 1.0655 | 26 |
| 0.0827 | 1.0818 | 27 |
| 0.0781 | 1.0903 | 28 |
| 0.0729 | 1.0998 | 29 |
| 0.0678 | 1.0932 | 30 |
| 0.0639 | 1.1051 | 31 |
| 0.0592 | 1.1125 | 32 |
| 0.0556 | 1.1240 | 33 |
| 0.0509 | 1.1177 | 34 |
| 0.0512 | 1.1355 | 35 |
| 0.0438 | 1.1405 | 36 |
| 0.0453 | 1.1322 | 37 |
| 0.0443 | 1.1419 | 38 |
| 0.0407 | 1.1419 | 39 |
| 0.0397 | 1.1495 | 40 |
| 0.0386 | 1.1609 | 41 |
| 0.0346 | 1.1619 | 42 |
| 0.0351 | 1.1638 | 43 |
| 0.0344 | 1.1711 | 44 |
| 0.0302 | 1.1782 | 45 |
| 0.0470 | 1.1836 | 46 |
| 0.0330 | 1.1913 | 47 |
| 0.0284 | 1.1963 | 48 |
| 0.0268 | 1.1964 | 49 |
| 0.0255 | 1.2017 | 50 |
| 0.0236 | 1.2092 | 51 |
| 0.0241 | 1.2104 | 52 |
| 0.0234 | 1.2170 | 53 |
| 0.0216 | 1.2192 | 54 |
| 0.0209 | 1.2317 | 55 |
| 0.0205 | 1.2289 | 56 |
| 0.0193 | 1.2363 | 57 |
| 0.0191 | 1.2295 | 58 |
| 0.0184 | 1.2306 | 59 |
| 0.0185 | 1.2352 | 60 |
| 0.0184 | 1.2415 | 61 |
| 0.0174 | 1.2389 | 62 |
| 0.0166 | 1.2392 | 63 |
| 0.0167 | 1.2469 | 64 |
| 0.0166 | 1.2457 | 65 |
| 0.0147 | 1.2456 | 66 |
| 0.0146 | 1.2511 | 67 |
| 0.0147 | 1.2552 | 68 |
| 0.0147 | 1.2493 | 69 |
| 0.0133 | 1.2532 | 70 |
| 0.0135 | 1.2561 | 71 |
| 0.0136 | 1.2609 | 72 |
| 0.0130 | 1.2602 | 73 |
| 0.0119 | 1.2629 | 74 |
| 0.0123 | 1.2667 | 75 |
| 0.0114 | 1.2675 | 76 |
| 0.0122 | 1.2673 | 77 |
| 0.0111 | 1.2649 | 78 |
| 0.0099 | 1.2722 | 79 |
| 0.0109 | 1.2693 | 80 |
| 0.0101 | 1.2727 | 81 |
| 0.0101 | 1.2746 | 82 |
| 0.0096 | 1.2739 | 83 |
| 0.0103 | 1.2734 | 84 |
| 0.0096 | 1.2805 | 85 |
| 0.0093 | 1.2799 | 86 |
| 0.0097 | 1.2823 | 87 |
| 0.0093 | 1.2826 | 88 |
| 0.0095 | 1.2808 | 89 |
| 0.0091 | 1.2875 | 90 |
| 0.0081 | 1.2849 | 91 |
| 0.0084 | 1.2849 | 92 |
| 0.0083 | 1.2838 | 93 |
| 0.0089 | 1.2866 | 94 |
| 0.0084 | 1.2851 | 95 |
| 0.0082 | 1.2870 | 96 |
| 0.0078 | 1.2871 | 97 |
| 0.0078 | 1.2872 | 98 |
| 0.0077 | 1.2870 | 99 |
### Framework versions
- Transformers 4.35.2
- TensorFlow 2.14.0
- Datasets 2.15.0
- Tokenizers 0.15.0
|