bedus-creation
commited on
Commit
•
dfcc5e1
1
Parent(s):
6f913a8
Training in progress epoch 0
Browse files- README.md +6 -143
- config.json +1 -1
- tf_model.h5 +1 -1
- tokenizer.json +0 -0
- tokenizer_config.json +4 -0
README.md
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
-
base_model: t5-base
|
4 |
tags:
|
5 |
- generated_from_keras_callback
|
6 |
model-index:
|
@@ -13,11 +13,11 @@ probably proofread and complete it, then remove this comment. -->
|
|
13 |
|
14 |
# bedus-creation/eng-limbu-t5-base-all-001
|
15 |
|
16 |
-
This model is a fine-tuned version of [t5-base](https://huggingface.co/t5-base) on an unknown dataset.
|
17 |
It achieves the following results on the evaluation set:
|
18 |
-
- Train Loss:
|
19 |
-
- Validation Loss:
|
20 |
-
- Epoch:
|
21 |
|
22 |
## Model description
|
23 |
|
@@ -43,144 +43,7 @@ The following hyperparameters were used during training:
|
|
43 |
|
44 |
| Train Loss | Validation Loss | Epoch |
|
45 |
|:----------:|:---------------:|:-----:|
|
46 |
-
| 7.
|
47 |
-
| 6.4232 | 6.1990 | 1 |
|
48 |
-
| 6.1568 | 6.0153 | 2 |
|
49 |
-
| 6.0061 | 5.8803 | 3 |
|
50 |
-
| 5.9020 | 5.7640 | 4 |
|
51 |
-
| 5.8048 | 5.6789 | 5 |
|
52 |
-
| 5.6626 | 5.5948 | 6 |
|
53 |
-
| 5.5848 | 5.5166 | 7 |
|
54 |
-
| 5.5104 | 5.4389 | 8 |
|
55 |
-
| 5.4123 | 5.3547 | 9 |
|
56 |
-
| 5.3234 | 5.2776 | 10 |
|
57 |
-
| 5.2340 | 5.2203 | 11 |
|
58 |
-
| 5.1654 | 5.1697 | 12 |
|
59 |
-
| 5.0841 | 5.1080 | 13 |
|
60 |
-
| 4.9937 | 5.0664 | 14 |
|
61 |
-
| 4.9113 | 5.0153 | 15 |
|
62 |
-
| 4.8582 | 4.9646 | 16 |
|
63 |
-
| 4.7942 | 4.9417 | 17 |
|
64 |
-
| 4.7292 | 4.8847 | 18 |
|
65 |
-
| 4.6360 | 4.8478 | 19 |
|
66 |
-
| 4.5783 | 4.8167 | 20 |
|
67 |
-
| 4.5391 | 4.7565 | 21 |
|
68 |
-
| 4.4614 | 4.7051 | 22 |
|
69 |
-
| 4.3773 | 4.6806 | 23 |
|
70 |
-
| 4.3157 | 4.6535 | 24 |
|
71 |
-
| 4.2730 | 4.6082 | 25 |
|
72 |
-
| 4.2060 | 4.5799 | 26 |
|
73 |
-
| 4.1279 | 4.5355 | 27 |
|
74 |
-
| 4.0707 | 4.5202 | 28 |
|
75 |
-
| 4.0216 | 4.4787 | 29 |
|
76 |
-
| 3.9908 | 4.4401 | 30 |
|
77 |
-
| 3.9034 | 4.4155 | 31 |
|
78 |
-
| 3.8432 | 4.3823 | 32 |
|
79 |
-
| 3.7864 | 4.3547 | 33 |
|
80 |
-
| 3.7511 | 4.3313 | 34 |
|
81 |
-
| 3.6824 | 4.3112 | 35 |
|
82 |
-
| 3.6172 | 4.2764 | 36 |
|
83 |
-
| 3.5942 | 4.2378 | 37 |
|
84 |
-
| 3.5092 | 4.1943 | 38 |
|
85 |
-
| 3.4827 | 4.1674 | 39 |
|
86 |
-
| 3.4252 | 4.1349 | 40 |
|
87 |
-
| 3.3630 | 4.1109 | 41 |
|
88 |
-
| 3.3343 | 4.0672 | 42 |
|
89 |
-
| 3.2801 | 4.0650 | 43 |
|
90 |
-
| 3.2159 | 4.0524 | 44 |
|
91 |
-
| 3.1785 | 4.0022 | 45 |
|
92 |
-
| 3.1228 | 3.9873 | 46 |
|
93 |
-
| 3.0704 | 3.9762 | 47 |
|
94 |
-
| 3.0201 | 3.9244 | 48 |
|
95 |
-
| 2.9825 | 3.9083 | 49 |
|
96 |
-
| 2.9366 | 3.8723 | 50 |
|
97 |
-
| 2.8811 | 3.8877 | 51 |
|
98 |
-
| 2.8314 | 3.8137 | 52 |
|
99 |
-
| 2.7829 | 3.7879 | 53 |
|
100 |
-
| 2.7540 | 3.7691 | 54 |
|
101 |
-
| 2.7024 | 3.7628 | 55 |
|
102 |
-
| 2.6563 | 3.7334 | 56 |
|
103 |
-
| 2.6096 | 3.7073 | 57 |
|
104 |
-
| 2.5802 | 3.6936 | 58 |
|
105 |
-
| 2.5430 | 3.6533 | 59 |
|
106 |
-
| 2.5094 | 3.6374 | 60 |
|
107 |
-
| 2.4740 | 3.6153 | 61 |
|
108 |
-
| 2.4471 | 3.5830 | 62 |
|
109 |
-
| 2.3820 | 3.5756 | 63 |
|
110 |
-
| 2.3559 | 3.5195 | 64 |
|
111 |
-
| 2.2974 | 3.5223 | 65 |
|
112 |
-
| 2.2807 | 3.4774 | 66 |
|
113 |
-
| 2.2312 | 3.4700 | 67 |
|
114 |
-
| 2.2084 | 3.4522 | 68 |
|
115 |
-
| 2.1697 | 3.4271 | 69 |
|
116 |
-
| 2.1373 | 3.4175 | 70 |
|
117 |
-
| 2.0972 | 3.3715 | 71 |
|
118 |
-
| 2.0672 | 3.3515 | 72 |
|
119 |
-
| 2.0228 | 3.3328 | 73 |
|
120 |
-
| 1.9925 | 3.3276 | 74 |
|
121 |
-
| 1.9623 | 3.2728 | 75 |
|
122 |
-
| 1.9449 | 3.2553 | 76 |
|
123 |
-
| 1.9007 | 3.2412 | 77 |
|
124 |
-
| 1.8839 | 3.2152 | 78 |
|
125 |
-
| 1.8487 | 3.1977 | 79 |
|
126 |
-
| 1.8286 | 3.1870 | 80 |
|
127 |
-
| 1.8008 | 3.1476 | 81 |
|
128 |
-
| 1.7678 | 3.1412 | 82 |
|
129 |
-
| 1.7306 | 3.1152 | 83 |
|
130 |
-
| 1.7038 | 3.1064 | 84 |
|
131 |
-
| 1.6817 | 3.0788 | 85 |
|
132 |
-
| 1.6639 | 3.0666 | 86 |
|
133 |
-
| 1.6280 | 3.0470 | 87 |
|
134 |
-
| 1.5991 | 3.0384 | 88 |
|
135 |
-
| 1.5706 | 3.0114 | 89 |
|
136 |
-
| 1.5425 | 2.9920 | 90 |
|
137 |
-
| 1.5268 | 2.9801 | 91 |
|
138 |
-
| 1.5053 | 2.9711 | 92 |
|
139 |
-
| 1.4853 | 2.9577 | 93 |
|
140 |
-
| 1.4577 | 2.9318 | 94 |
|
141 |
-
| 1.4283 | 2.9221 | 95 |
|
142 |
-
| 1.4214 | 2.9050 | 96 |
|
143 |
-
| 1.3879 | 2.8948 | 97 |
|
144 |
-
| 1.3845 | 2.8711 | 98 |
|
145 |
-
| 1.3533 | 2.8578 | 99 |
|
146 |
-
| 1.3173 | 2.8623 | 100 |
|
147 |
-
| 1.3351 | 2.8313 | 101 |
|
148 |
-
| 1.2870 | 2.8154 | 102 |
|
149 |
-
| 1.2675 | 2.8059 | 103 |
|
150 |
-
| 1.2685 | 2.8006 | 104 |
|
151 |
-
| 1.2394 | 2.7965 | 105 |
|
152 |
-
| 1.2306 | 2.7801 | 106 |
|
153 |
-
| 1.2128 | 2.7736 | 107 |
|
154 |
-
| 1.1816 | 2.7579 | 108 |
|
155 |
-
| 1.1840 | 2.7607 | 109 |
|
156 |
-
| 1.1553 | 2.7330 | 110 |
|
157 |
-
| 1.1401 | 2.7383 | 111 |
|
158 |
-
| 1.1129 | 2.7264 | 112 |
|
159 |
-
| 1.1071 | 2.7153 | 113 |
|
160 |
-
| 1.1001 | 2.7166 | 114 |
|
161 |
-
| 1.0730 | 2.7045 | 115 |
|
162 |
-
| 1.0775 | 2.6979 | 116 |
|
163 |
-
| 1.0503 | 2.6844 | 117 |
|
164 |
-
| 1.0358 | 2.6796 | 118 |
|
165 |
-
| 1.0197 | 2.6764 | 119 |
|
166 |
-
| 1.0154 | 2.6817 | 120 |
|
167 |
-
| 1.0029 | 2.6551 | 121 |
|
168 |
-
| 0.9715 | 2.6581 | 122 |
|
169 |
-
| 0.9633 | 2.6540 | 123 |
|
170 |
-
| 0.9544 | 2.6502 | 124 |
|
171 |
-
| 0.9475 | 2.6607 | 125 |
|
172 |
-
| 0.9349 | 2.6438 | 126 |
|
173 |
-
| 0.9166 | 2.6376 | 127 |
|
174 |
-
| 0.9094 | 2.6522 | 128 |
|
175 |
-
| 0.8969 | 2.6271 | 129 |
|
176 |
-
| 0.9016 | 2.6322 | 130 |
|
177 |
-
| 0.8777 | 2.6246 | 131 |
|
178 |
-
| 0.8714 | 2.6267 | 132 |
|
179 |
-
| 0.8560 | 2.6213 | 133 |
|
180 |
-
| 0.8525 | 2.6302 | 134 |
|
181 |
-
| 0.8437 | 2.6147 | 135 |
|
182 |
-
| 0.8297 | 2.6206 | 136 |
|
183 |
-
| 0.8226 | 2.6325 | 137 |
|
184 |
|
185 |
|
186 |
### Framework versions
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
+
base_model: bedus-creation/eng-limbu-t5-base-all-001
|
4 |
tags:
|
5 |
- generated_from_keras_callback
|
6 |
model-index:
|
|
|
13 |
|
14 |
# bedus-creation/eng-limbu-t5-base-all-001
|
15 |
|
16 |
+
This model is a fine-tuned version of [bedus-creation/eng-limbu-t5-base-all-001](https://huggingface.co/bedus-creation/eng-limbu-t5-base-all-001) on an unknown dataset.
|
17 |
It achieves the following results on the evaluation set:
|
18 |
+
- Train Loss: 7.0062
|
19 |
+
- Validation Loss: 6.1115
|
20 |
+
- Epoch: 0
|
21 |
|
22 |
## Model description
|
23 |
|
|
|
43 |
|
44 |
| Train Loss | Validation Loss | Epoch |
|
45 |
|:----------:|:---------------:|:-----:|
|
46 |
+
| 7.0062 | 6.1115 | 0 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
47 |
|
48 |
|
49 |
### Framework versions
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "t5-base",
|
3 |
"architectures": [
|
4 |
"T5ForConditionalGeneration"
|
5 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "bedus-creation/eng-limbu-t5-base-all-001",
|
3 |
"architectures": [
|
4 |
"T5ForConditionalGeneration"
|
5 |
],
|
tf_model.h5
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1089544048
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7282f3a27d16093cbcbc478f850aec7a9f61de4fa96b596a2ee0599dadce2285
|
3 |
size 1089544048
|
tokenizer.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
CHANGED
@@ -104,8 +104,12 @@
|
|
104 |
"clean_up_tokenization_spaces": true,
|
105 |
"eos_token": "</s>",
|
106 |
"extra_ids": 100,
|
|
|
107 |
"model_max_length": 512,
|
108 |
"pad_token": "<pad>",
|
|
|
109 |
"tokenizer_class": "T5Tokenizer",
|
|
|
|
|
110 |
"unk_token": "<unk>"
|
111 |
}
|
|
|
104 |
"clean_up_tokenization_spaces": true,
|
105 |
"eos_token": "</s>",
|
106 |
"extra_ids": 100,
|
107 |
+
"max_length": 128,
|
108 |
"model_max_length": 512,
|
109 |
"pad_token": "<pad>",
|
110 |
+
"stride": 0,
|
111 |
"tokenizer_class": "T5Tokenizer",
|
112 |
+
"truncation_side": "right",
|
113 |
+
"truncation_strategy": "longest_first",
|
114 |
"unk_token": "<unk>"
|
115 |
}
|