File size: 9,515 Bytes
7d7d3c5 ca78311 7d7d3c5 7f9472e fd70f4c d8d8f6b 8891e0e a6b950b cc26306 2043216 3fbe857 b34c0d1 106596b e89de12 d33d55c 3fcee11 5fd9384 1136126 19a38c3 8e2c6ea bd6b5a4 ce7cd31 612470a 3fbc867 fbd6e4d cc96fb4 756171f 48720d4 56ac9ee edefcac d1a4321 a28f048 fe0a43a 185285b 354394b 1e2e431 ffafe20 6c4174b 6d6b830 d70bd97 02784b7 c5b339a 5de5af9 dc8f767 ae3365c a9dcdb7 3eb5e72 b89a327 0338a3e c4bd8f4 82f6ad4 7a35ff1 430dac5 6bfeead 426396b ac8738c e777dfd c302d40 11d74a7 90e603a 6951ea0 75e3214 535a074 8272d3b 6960e50 b4f4d61 535e015 fca7c4b a8ac979 c95e72c 1484567 bf69677 f9a0ae1 31e701a 29b0d91 656cba8 528d3b0 aaf25b1 dc6df3d 28a08c7 791fa37 b6266d6 e88bf1d c11250e a9a0168 f89e3cb 1250d26 aeda760 8ab728a d8282f8 033d9b6 cbe1641 f4a730e 42d5238 d295478 78540a9 4938b0a 2d48387 ad19f46 d7e4d25 2d96d98 f56e332 3f1e186 430a8a2 6e5b377 b57e864 dc13a56 48c8c6a 33baa9b 58407b8 214ee03 a34e11a 16a469b 13a05e0 5394b20 be0accf c534de4 6f222fa 7355947 bac0793 471bf9e dc460a2 f5f6c80 c817a99 27f839d 44b5ced 49e3e90 6277811 58e674c b4c208f 48d13ac 924627f 4469759 18c9eec 948850e f35a666 7a2da74 1d959fd f2fbbf0 50e6f80 1a0a726 65c920d e1e8a64 fe7fed8 a5fee77 562be24 3aeb144 030a9db 75bb586 71cf1b9 f42f235 d8f1a63 6e00289 4ff07e5 7f5f1fd ea38c90 2cdd5b6 ff3e5c8 48188c9 b75fa45 bdaf93d 2b2b5ff e1372ee f710dae 4d053b3 d0c23c5 1bace52 8f06b5b 7564e1f 726be00 e7de98e ce91200 b0c1e66 4c2d311 1e6a7aa ad7aa98 c11ae0b 137420f 2e7f44a 61f7465 5e760dc 5ae85b3 f637419 d58aa87 a86d43a 1f960f2 03573d8 2e9d4bf 570c169 936fdcb 5c15acc a72c548 74bd706 4afe696 f9b30f1 52f98c8 957064d bac091f 2cda385 53cee62 2cbfda0 6da8263 ca78311 7d7d3c5 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 |
---
license: apache-2.0
base_model: t5-small
tags:
- generated_from_keras_callback
model-index:
- name: pijarcandra22/NMTIndoBaliT5
results: []
---
<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->
# pijarcandra22/NMTIndoBaliT5
This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.1551
- Validation Loss: 2.1825
- Epoch: 200
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 1e-04, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01}
- training_precision: float32
### Training results
| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 3.2881 | 2.6852 | 0 |
| 2.7514 | 2.4004 | 1 |
| 2.5012 | 2.2171 | 2 |
| 2.3252 | 2.0959 | 3 |
| 2.1930 | 1.9901 | 4 |
| 2.0837 | 1.9130 | 5 |
| 1.9912 | 1.8452 | 6 |
| 1.9107 | 1.7974 | 7 |
| 1.8459 | 1.7521 | 8 |
| 1.7902 | 1.7165 | 9 |
| 1.7321 | 1.6842 | 10 |
| 1.6811 | 1.6400 | 11 |
| 1.6374 | 1.6230 | 12 |
| 1.5973 | 1.5960 | 13 |
| 1.5588 | 1.5765 | 14 |
| 1.5244 | 1.5589 | 15 |
| 1.4933 | 1.5370 | 16 |
| 1.4588 | 1.5300 | 17 |
| 1.4325 | 1.5107 | 18 |
| 1.4054 | 1.4970 | 19 |
| 1.3730 | 1.4839 | 20 |
| 1.3475 | 1.4789 | 21 |
| 1.3231 | 1.4616 | 22 |
| 1.3035 | 1.4568 | 23 |
| 1.2768 | 1.4489 | 24 |
| 1.2587 | 1.4396 | 25 |
| 1.2380 | 1.4364 | 26 |
| 1.2208 | 1.4273 | 27 |
| 1.2026 | 1.4228 | 28 |
| 1.1755 | 1.4141 | 29 |
| 1.1614 | 1.4062 | 30 |
| 1.1460 | 1.4060 | 31 |
| 1.1289 | 1.3934 | 32 |
| 1.1134 | 1.4007 | 33 |
| 1.0965 | 1.3927 | 34 |
| 1.0818 | 1.3874 | 35 |
| 1.0661 | 1.3921 | 36 |
| 1.0482 | 1.3795 | 37 |
| 1.0345 | 1.3853 | 38 |
| 1.0195 | 1.3835 | 39 |
| 1.0074 | 1.3772 | 40 |
| 0.9890 | 1.3851 | 41 |
| 0.9833 | 1.3724 | 42 |
| 0.9667 | 1.3740 | 43 |
| 0.9561 | 1.3752 | 44 |
| 0.9429 | 1.3673 | 45 |
| 0.9301 | 1.3828 | 46 |
| 0.9141 | 1.3806 | 47 |
| 0.9050 | 1.3772 | 48 |
| 0.8952 | 1.3812 | 49 |
| 0.8809 | 1.3718 | 50 |
| 0.8725 | 1.3825 | 51 |
| 0.8601 | 1.3842 | 52 |
| 0.8488 | 1.3827 | 53 |
| 0.8375 | 1.3920 | 54 |
| 0.8257 | 1.3936 | 55 |
| 0.8184 | 1.3842 | 56 |
| 0.8081 | 1.3846 | 57 |
| 0.7986 | 1.3860 | 58 |
| 0.7883 | 1.3943 | 59 |
| 0.7787 | 1.4004 | 60 |
| 0.7666 | 1.4071 | 61 |
| 0.7554 | 1.4079 | 62 |
| 0.7470 | 1.4038 | 63 |
| 0.7366 | 1.4141 | 64 |
| 0.7279 | 1.4135 | 65 |
| 0.7250 | 1.4111 | 66 |
| 0.7128 | 1.4196 | 67 |
| 0.7042 | 1.4182 | 68 |
| 0.6946 | 1.4378 | 69 |
| 0.6851 | 1.4350 | 70 |
| 0.6764 | 1.4403 | 71 |
| 0.6695 | 1.4474 | 72 |
| 0.6606 | 1.4454 | 73 |
| 0.6565 | 1.4516 | 74 |
| 0.6450 | 1.4595 | 75 |
| 0.6347 | 1.4700 | 76 |
| 0.6287 | 1.4746 | 77 |
| 0.6183 | 1.4813 | 78 |
| 0.6143 | 1.4785 | 79 |
| 0.6053 | 1.4848 | 80 |
| 0.5994 | 1.4777 | 81 |
| 0.5903 | 1.4962 | 82 |
| 0.5828 | 1.5102 | 83 |
| 0.5760 | 1.4957 | 84 |
| 0.5696 | 1.5121 | 85 |
| 0.5637 | 1.5168 | 86 |
| 0.5578 | 1.5183 | 87 |
| 0.5499 | 1.5184 | 88 |
| 0.5396 | 1.5433 | 89 |
| 0.5345 | 1.5411 | 90 |
| 0.5268 | 1.5338 | 91 |
| 0.5220 | 1.5556 | 92 |
| 0.5184 | 1.5489 | 93 |
| 0.5122 | 1.5635 | 94 |
| 0.5014 | 1.5674 | 95 |
| 0.4921 | 1.5773 | 96 |
| 0.4925 | 1.5773 | 97 |
| 0.4821 | 1.5938 | 98 |
| 0.4769 | 1.6013 | 99 |
| 0.4723 | 1.5979 | 100 |
| 0.4692 | 1.6131 | 101 |
| 0.4603 | 1.6247 | 102 |
| 0.4553 | 1.6276 | 103 |
| 0.4476 | 1.6376 | 104 |
| 0.4401 | 1.6390 | 105 |
| 0.4384 | 1.6442 | 106 |
| 0.4305 | 1.6548 | 107 |
| 0.4263 | 1.6617 | 108 |
| 0.4232 | 1.6523 | 109 |
| 0.4185 | 1.6561 | 110 |
| 0.4129 | 1.6779 | 111 |
| 0.4036 | 1.6897 | 112 |
| 0.4005 | 1.6873 | 113 |
| 0.3948 | 1.6987 | 114 |
| 0.3892 | 1.7120 | 115 |
| 0.3859 | 1.7049 | 116 |
| 0.3795 | 1.7241 | 117 |
| 0.3802 | 1.7273 | 118 |
| 0.3731 | 1.7387 | 119 |
| 0.3672 | 1.7447 | 120 |
| 0.3629 | 1.7513 | 121 |
| 0.3607 | 1.7515 | 122 |
| 0.3543 | 1.7585 | 123 |
| 0.3504 | 1.7601 | 124 |
| 0.3477 | 1.7657 | 125 |
| 0.3453 | 1.7733 | 126 |
| 0.3448 | 1.7718 | 127 |
| 0.3390 | 1.7971 | 128 |
| 0.3352 | 1.7929 | 129 |
| 0.3273 | 1.7988 | 130 |
| 0.3250 | 1.8192 | 131 |
| 0.3222 | 1.8220 | 132 |
| 0.3173 | 1.8289 | 133 |
| 0.3171 | 1.8261 | 134 |
| 0.3124 | 1.8415 | 135 |
| 0.3040 | 1.8379 | 136 |
| 0.3040 | 1.8533 | 137 |
| 0.3030 | 1.8511 | 138 |
| 0.2970 | 1.8537 | 139 |
| 0.2938 | 1.8697 | 140 |
| 0.2929 | 1.8730 | 141 |
| 0.2892 | 1.8632 | 142 |
| 0.2816 | 1.8796 | 143 |
| 0.2812 | 1.8870 | 144 |
| 0.2761 | 1.8891 | 145 |
| 0.2731 | 1.9134 | 146 |
| 0.2698 | 1.9100 | 147 |
| 0.2671 | 1.9207 | 148 |
| 0.2639 | 1.9196 | 149 |
| 0.2621 | 1.9130 | 150 |
| 0.2589 | 1.9273 | 151 |
| 0.2558 | 1.9336 | 152 |
| 0.2545 | 1.9355 | 153 |
| 0.2487 | 1.9551 | 154 |
| 0.2493 | 1.9573 | 155 |
| 0.2449 | 1.9552 | 156 |
| 0.2421 | 1.9591 | 157 |
| 0.2405 | 1.9556 | 158 |
| 0.2367 | 1.9807 | 159 |
| 0.2342 | 1.9859 | 160 |
| 0.2316 | 1.9803 | 161 |
| 0.2281 | 1.9853 | 162 |
| 0.2269 | 1.9970 | 163 |
| 0.2250 | 2.0120 | 164 |
| 0.2236 | 2.0107 | 165 |
| 0.2194 | 2.0208 | 166 |
| 0.2183 | 2.0198 | 167 |
| 0.2168 | 2.0265 | 168 |
| 0.2172 | 2.0278 | 169 |
| 0.2117 | 2.0380 | 170 |
| 0.2078 | 2.0448 | 171 |
| 0.2091 | 2.0415 | 172 |
| 0.2065 | 2.0459 | 173 |
| 0.2027 | 2.0597 | 174 |
| 0.1995 | 2.0659 | 175 |
| 0.1980 | 2.0811 | 176 |
| 0.1971 | 2.0704 | 177 |
| 0.1932 | 2.0785 | 178 |
| 0.1892 | 2.0783 | 179 |
| 0.1924 | 2.0742 | 180 |
| 0.1872 | 2.0979 | 181 |
| 0.1858 | 2.0958 | 182 |
| 0.1853 | 2.1005 | 183 |
| 0.1834 | 2.1166 | 184 |
| 0.1810 | 2.1027 | 185 |
| 0.1789 | 2.1151 | 186 |
| 0.1768 | 2.1302 | 187 |
| 0.1768 | 2.1200 | 188 |
| 0.1766 | 2.1399 | 189 |
| 0.1732 | 2.1196 | 190 |
| 0.1719 | 2.1362 | 191 |
| 0.1697 | 2.1447 | 192 |
| 0.1684 | 2.1464 | 193 |
| 0.1699 | 2.1442 | 194 |
| 0.1657 | 2.1492 | 195 |
| 0.1607 | 2.1644 | 196 |
| 0.1603 | 2.1667 | 197 |
| 0.1580 | 2.1715 | 198 |
| 0.1588 | 2.1818 | 199 |
| 0.1551 | 2.1825 | 200 |
### Framework versions
- Transformers 4.38.2
- TensorFlow 2.15.0
- Datasets 2.18.0
- Tokenizers 0.15.2
|