File size: 7,099 Bytes
6ea6f2f d0e2028 6ea6f2f d0e2028 6ea6f2f b105040 6ea6f2f 25f6133 ff30341 e3c8509 6c2f3ab 40e6615 2b5f59d fc35459 219142e 1e0ffcd 41213fc 7eb507c 64e514f 3960e9d 4f16080 2e90df9 d9969b1 f81a95b ff8cd4c cddaf36 fc802f5 15ecfd9 d59c414 83ece39 7ce78d0 bbcad9f 9a7cbc9 9ddda77 1c55bc3 a59c794 1a87b7a 4996b6b 50f881f 81b0c44 87b35eb 55b0826 92276a2 6650e0d ef5782b 82ed31f e6cfcf0 a8d1188 32f0564 a3c3029 fcb117f 5724768 d8612b9 c15007f 474fe74 e0273ea 6d9bbb4 b455273 4d5d1f8 6ea49d8 0dbe9ea 1ac703e 744ee66 26c902d 256b3b7 d91b2fe b4151c0 19954c2 23c5ef1 3009268 8ee8b16 df17981 c9ad1dc ce8fb32 f54cc54 6876116 12822af 18d5d30 49ab585 b5a3e77 e43545e 1687d9c 4cdd67c c5bc86a 8506a7b 69c04fa 2e1efbd 2915355 304b3e6 b7756e7 9633c87 f99d809 c091d00 7c62719 d65149d 986f7c9 08138b9 768a390 e769d70 eed8ff6 7260197 1603513 23d9133 aa18475 f8c4a4c dd2b096 217724b 2219111 3df2d23 0657b19 71a6535 77559ce 35407c3 ae83898 777419e 3834157 04d5213 e2ae233 5ca2798 199cfa8 55a3aab 5f97547 77a4a71 ef22e50 4dad7f6 8de672c 7d974f2 a6cceb7 e9f82d4 7717317 bb1b96e 07b230c 08504f0 7211864 c5e5910 8b8b4d9 a518839 4299435 93babf6 a31cd0e b105040 6ea6f2f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 |
---
license: apache-2.0
base_model: bedus-creation/mBart-small-dataset-ii-eng-lim-003
tags:
- generated_from_keras_callback
model-index:
- name: bedus-creation/mBart-small-dataset-ii-eng-lim-003
results: []
---
<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->
# bedus-creation/mBart-small-dataset-ii-eng-lim-003
This model is a fine-tuned version of [bedus-creation/mBart-small-dataset-ii-eng-lim-003](https://huggingface.co/bedus-creation/mBart-small-dataset-ii-eng-lim-003) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.1069
- Validation Loss: 0.3933
- Epoch: 137
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 1e-04, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01}
- training_precision: float32
### Training results
| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 0.2093 | 0.2072 | 0 |
| 0.2068 | 0.2056 | 1 |
| 0.2062 | 0.2023 | 2 |
| 0.2045 | 0.2054 | 3 |
| 0.2027 | 0.2188 | 4 |
| 0.2019 | 0.2067 | 5 |
| 0.1997 | 0.2056 | 6 |
| 0.1991 | 0.2074 | 7 |
| 0.1978 | 0.2024 | 8 |
| 0.1962 | 0.2067 | 9 |
| 0.1955 | 0.2074 | 10 |
| 0.1945 | 0.2089 | 11 |
| 0.1928 | 0.2168 | 12 |
| 0.1907 | 0.2201 | 13 |
| 0.1900 | 0.2102 | 14 |
| 0.1888 | 0.2130 | 15 |
| 0.1882 | 0.2211 | 16 |
| 0.1870 | 0.2117 | 17 |
| 0.1857 | 0.2134 | 18 |
| 0.1838 | 0.2147 | 19 |
| 0.1824 | 0.2187 | 20 |
| 0.1812 | 0.2224 | 21 |
| 0.1813 | 0.2249 | 22 |
| 0.1798 | 0.2200 | 23 |
| 0.1787 | 0.2273 | 24 |
| 0.1772 | 0.2263 | 25 |
| 0.1780 | 0.2273 | 26 |
| 0.1764 | 0.2270 | 27 |
| 0.1754 | 0.2245 | 28 |
| 0.1738 | 0.2260 | 29 |
| 0.1730 | 0.2327 | 30 |
| 0.1720 | 0.2300 | 31 |
| 0.1702 | 0.2347 | 32 |
| 0.1698 | 0.2396 | 33 |
| 0.1689 | 0.2340 | 34 |
| 0.1693 | 0.2345 | 35 |
| 0.1661 | 0.2424 | 36 |
| 0.1663 | 0.2388 | 37 |
| 0.1658 | 0.2436 | 38 |
| 0.1654 | 0.2506 | 39 |
| 0.1639 | 0.2406 | 40 |
| 0.1635 | 0.2524 | 41 |
| 0.1619 | 0.2379 | 42 |
| 0.1609 | 0.2449 | 43 |
| 0.1602 | 0.2466 | 44 |
| 0.1602 | 0.2537 | 45 |
| 0.1586 | 0.2457 | 46 |
| 0.1576 | 0.2589 | 47 |
| 0.1573 | 0.2547 | 48 |
| 0.1566 | 0.2532 | 49 |
| 0.1546 | 0.2565 | 50 |
| 0.1540 | 0.2544 | 51 |
| 0.1545 | 0.2637 | 52 |
| 0.1515 | 0.2580 | 53 |
| 0.1520 | 0.2654 | 54 |
| 0.1524 | 0.2650 | 55 |
| 0.1513 | 0.2701 | 56 |
| 0.1500 | 0.2767 | 57 |
| 0.1492 | 0.2646 | 58 |
| 0.1483 | 0.2696 | 59 |
| 0.1480 | 0.2729 | 60 |
| 0.1475 | 0.2709 | 61 |
| 0.1458 | 0.2757 | 62 |
| 0.1460 | 0.2778 | 63 |
| 0.1446 | 0.2775 | 64 |
| 0.1440 | 0.2727 | 65 |
| 0.1438 | 0.2862 | 66 |
| 0.1444 | 0.2719 | 67 |
| 0.1423 | 0.2827 | 68 |
| 0.1418 | 0.2830 | 69 |
| 0.1402 | 0.2787 | 70 |
| 0.1404 | 0.2799 | 71 |
| 0.1388 | 0.2857 | 72 |
| 0.1392 | 0.2889 | 73 |
| 0.1398 | 0.2868 | 74 |
| 0.1389 | 0.2920 | 75 |
| 0.1359 | 0.3010 | 76 |
| 0.1369 | 0.2873 | 77 |
| 0.1366 | 0.2921 | 78 |
| 0.1358 | 0.2895 | 79 |
| 0.1343 | 0.3071 | 80 |
| 0.1344 | 0.2981 | 81 |
| 0.1341 | 0.3033 | 82 |
| 0.1328 | 0.3008 | 83 |
| 0.1332 | 0.2933 | 84 |
| 0.1317 | 0.3155 | 85 |
| 0.1310 | 0.3091 | 86 |
| 0.1307 | 0.3205 | 87 |
| 0.1295 | 0.3142 | 88 |
| 0.1295 | 0.3141 | 89 |
| 0.1299 | 0.3103 | 90 |
| 0.1282 | 0.3209 | 91 |
| 0.1284 | 0.3167 | 92 |
| 0.1272 | 0.3242 | 93 |
| 0.1270 | 0.3159 | 94 |
| 0.1245 | 0.3275 | 95 |
| 0.1244 | 0.3218 | 96 |
| 0.1248 | 0.3270 | 97 |
| 0.1241 | 0.3354 | 98 |
| 0.1231 | 0.3430 | 99 |
| 0.1233 | 0.3318 | 100 |
| 0.1222 | 0.3387 | 101 |
| 0.1225 | 0.3367 | 102 |
| 0.1221 | 0.3501 | 103 |
| 0.1214 | 0.3370 | 104 |
| 0.1207 | 0.3391 | 105 |
| 0.1197 | 0.3436 | 106 |
| 0.1193 | 0.3388 | 107 |
| 0.1208 | 0.3383 | 108 |
| 0.1186 | 0.3526 | 109 |
| 0.1177 | 0.3471 | 110 |
| 0.1179 | 0.3490 | 111 |
| 0.1179 | 0.3498 | 112 |
| 0.1177 | 0.3379 | 113 |
| 0.1169 | 0.3518 | 114 |
| 0.1165 | 0.3590 | 115 |
| 0.1161 | 0.3550 | 116 |
| 0.1159 | 0.3545 | 117 |
| 0.1150 | 0.3562 | 118 |
| 0.1123 | 0.3641 | 119 |
| 0.1137 | 0.3658 | 120 |
| 0.1153 | 0.3613 | 121 |
| 0.1130 | 0.3767 | 122 |
| 0.1129 | 0.3812 | 123 |
| 0.1127 | 0.3696 | 124 |
| 0.1118 | 0.3704 | 125 |
| 0.1116 | 0.3689 | 126 |
| 0.1107 | 0.3776 | 127 |
| 0.1103 | 0.3775 | 128 |
| 0.1108 | 0.3803 | 129 |
| 0.1097 | 0.3877 | 130 |
| 0.1093 | 0.3860 | 131 |
| 0.1080 | 0.3919 | 132 |
| 0.1082 | 0.3886 | 133 |
| 0.1091 | 0.3888 | 134 |
| 0.1071 | 0.3931 | 135 |
| 0.1072 | 0.3925 | 136 |
| 0.1069 | 0.3933 | 137 |
### Framework versions
- Transformers 4.33.3
- TensorFlow 2.13.0
- Datasets 2.14.5
- Tokenizers 0.13.3
|