kaen2891 commited on
Commit
28da9a1
1 Parent(s): 20ad293

Model save

Browse files
Files changed (3) hide show
  1. README.md +159 -0
  2. generation_config.json +240 -0
  3. model.safetensors +1 -1
README.md ADDED
@@ -0,0 +1,159 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: openai/whisper-tiny
4
+ tags:
5
+ - generated_from_trainer
6
+ model-index:
7
+ - name: whisper-tiny-kor-430k-hf-ep100
8
+ results: []
9
+ ---
10
+
11
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
+ should probably proofread and complete it, then remove this comment. -->
13
+
14
+ # whisper-tiny-kor-430k-hf-ep100
15
+
16
+ This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the None dataset.
17
+ It achieves the following results on the evaluation set:
18
+ - Loss: 0.4802
19
+ - Cer: 7.0810
20
+
21
+ ## Model description
22
+
23
+ More information needed
24
+
25
+ ## Intended uses & limitations
26
+
27
+ More information needed
28
+
29
+ ## Training and evaluation data
30
+
31
+ More information needed
32
+
33
+ ## Training procedure
34
+
35
+ ### Training hyperparameters
36
+
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 5e-05
39
+ - train_batch_size: 128
40
+ - eval_batch_size: 32
41
+ - seed: 42
42
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
+ - lr_scheduler_type: linear
44
+ - lr_scheduler_warmup_steps: 500
45
+ - num_epochs: 100
46
+ - mixed_precision_training: Native AMP
47
+
48
+ ### Training results
49
+
50
+ | Training Loss | Epoch | Step | Validation Loss | Cer |
51
+ |:-------------:|:-----:|:------:|:---------------:|:-------:|
52
+ | 0.0276 | 1.0 | 2969 | 0.1394 | 10.5081 |
53
+ | 0.0115 | 2.0 | 5938 | 0.1684 | 9.0941 |
54
+ | 0.0082 | 3.0 | 8907 | 0.1932 | 8.4145 |
55
+ | 0.0063 | 4.0 | 11876 | 0.2197 | 7.8659 |
56
+ | 0.0052 | 5.0 | 14845 | 0.2516 | 7.7668 |
57
+ | 0.0047 | 6.0 | 17814 | 0.2314 | 8.0565 |
58
+ | 0.004 | 7.0 | 20783 | 0.2270 | 8.3521 |
59
+ | 0.0037 | 8.0 | 23752 | 0.2540 | 7.8283 |
60
+ | 0.0033 | 9.0 | 26721 | 0.2586 | 7.9098 |
61
+ | 0.0028 | 10.0 | 29690 | 0.2891 | 7.4437 |
62
+ | 0.0029 | 11.0 | 32659 | 0.2796 | 7.7976 |
63
+ | 0.0025 | 12.0 | 35628 | 0.2630 | 8.2731 |
64
+ | 0.0025 | 13.0 | 38597 | 0.2955 | 7.8518 |
65
+ | 0.0025 | 14.0 | 41566 | 0.2812 | 7.4797 |
66
+ | 0.002 | 15.0 | 44535 | 0.2859 | 7.9954 |
67
+ | 0.0023 | 16.0 | 47504 | 0.3172 | 7.2374 |
68
+ | 0.002 | 17.0 | 50473 | 0.3382 | 7.5966 |
69
+ | 0.0018 | 18.0 | 53442 | 0.3320 | 7.6383 |
70
+ | 0.0018 | 19.0 | 56411 | 0.3197 | 7.6900 |
71
+ | 0.0015 | 20.0 | 59380 | 0.3305 | 8.3678 |
72
+ | 0.0016 | 21.0 | 62349 | 0.3409 | 7.5117 |
73
+ | 0.0015 | 22.0 | 65318 | 0.3382 | 7.8556 |
74
+ | 0.0016 | 23.0 | 68287 | 0.3282 | 7.5863 |
75
+ | 0.0015 | 24.0 | 71256 | 0.3220 | 8.2449 |
76
+ | 0.0013 | 25.0 | 74225 | 0.3272 | 7.7731 |
77
+ | 0.0015 | 26.0 | 77194 | 0.3557 | 7.8019 |
78
+ | 0.0014 | 27.0 | 80163 | 0.3807 | 7.3311 |
79
+ | 0.0012 | 28.0 | 83132 | 0.3398 | 7.8117 |
80
+ | 0.0013 | 29.0 | 86101 | 0.3892 | 7.6089 |
81
+ | 0.001 | 30.0 | 89070 | 0.3876 | 7.7875 |
82
+ | 0.0011 | 31.0 | 92039 | 0.3942 | 7.3922 |
83
+ | 0.0012 | 32.0 | 95008 | 0.3836 | 8.0308 |
84
+ | 0.0011 | 33.0 | 97977 | 0.3745 | 7.9775 |
85
+ | 0.001 | 34.0 | 100946 | 0.3605 | 8.0117 |
86
+ | 0.001 | 35.0 | 103915 | 0.3615 | 7.4853 |
87
+ | 0.001 | 36.0 | 106884 | 0.3563 | 7.5916 |
88
+ | 0.0009 | 37.0 | 109853 | 0.3469 | 7.4750 |
89
+ | 0.0009 | 38.0 | 112822 | 0.3940 | 7.5919 |
90
+ | 0.0009 | 39.0 | 115791 | 0.3771 | 7.5443 |
91
+ | 0.0009 | 40.0 | 118760 | 0.3392 | 7.6593 |
92
+ | 0.0009 | 41.0 | 121729 | 0.3498 | 7.6393 |
93
+ | 0.0009 | 42.0 | 124698 | 0.3705 | 7.4474 |
94
+ | 0.0008 | 43.0 | 127667 | 0.3758 | 7.2274 |
95
+ | 0.0008 | 44.0 | 130636 | 0.3944 | 7.6919 |
96
+ | 0.0009 | 45.0 | 133605 | 0.3885 | 7.5565 |
97
+ | 0.0008 | 46.0 | 136574 | 0.3830 | 7.4628 |
98
+ | 0.0008 | 47.0 | 139543 | 0.3972 | 7.8546 |
99
+ | 0.0008 | 48.0 | 142512 | 0.3875 | 7.4916 |
100
+ | 0.0007 | 49.0 | 145481 | 0.3438 | 7.2606 |
101
+ | 0.0007 | 50.0 | 148450 | 0.3540 | 7.1581 |
102
+ | 0.0008 | 51.0 | 151419 | 0.3768 | 7.1712 |
103
+ | 0.0007 | 52.0 | 154388 | 0.4050 | 7.2286 |
104
+ | 0.0007 | 53.0 | 157357 | 0.3785 | 7.4637 |
105
+ | 0.0008 | 54.0 | 160326 | 0.4145 | 7.4800 |
106
+ | 0.0008 | 55.0 | 163295 | 0.4042 | 7.3791 |
107
+ | 0.0006 | 56.0 | 166264 | 0.3885 | 7.6994 |
108
+ | 0.0006 | 57.0 | 169233 | 0.4153 | 7.5440 |
109
+ | 0.0006 | 58.0 | 172202 | 0.4111 | 7.3408 |
110
+ | 0.0006 | 59.0 | 175171 | 0.4147 | 7.2872 |
111
+ | 0.0006 | 60.0 | 178140 | 0.4209 | 7.6270 |
112
+ | 0.0006 | 61.0 | 181109 | 0.4041 | 7.4258 |
113
+ | 0.0006 | 62.0 | 184078 | 0.4032 | 7.5324 |
114
+ | 0.0006 | 63.0 | 187047 | 0.4214 | 7.3687 |
115
+ | 0.0005 | 64.0 | 190016 | 0.3991 | 7.2750 |
116
+ | 0.0005 | 65.0 | 192985 | 0.3885 | 7.1731 |
117
+ | 0.0006 | 66.0 | 195954 | 0.4087 | 7.5063 |
118
+ | 0.0005 | 67.0 | 198923 | 0.3760 | 7.4913 |
119
+ | 0.0005 | 68.0 | 201892 | 0.3929 | 7.3314 |
120
+ | 0.0005 | 69.0 | 204861 | 0.4044 | 7.5173 |
121
+ | 0.0005 | 70.0 | 207830 | 0.4075 | 7.2712 |
122
+ | 0.0005 | 71.0 | 210799 | 0.4170 | 7.2415 |
123
+ | 0.0005 | 72.0 | 213768 | 0.4148 | 7.1142 |
124
+ | 0.0005 | 73.0 | 216737 | 0.4271 | 7.3020 |
125
+ | 0.0005 | 74.0 | 219706 | 0.4281 | 7.1863 |
126
+ | 0.0004 | 75.0 | 222675 | 0.4202 | 7.1543 |
127
+ | 0.0005 | 76.0 | 225644 | 0.4320 | 7.2910 |
128
+ | 0.0005 | 77.0 | 228613 | 0.4328 | 7.3995 |
129
+ | 0.0005 | 78.0 | 231582 | 0.4304 | 7.2255 |
130
+ | 0.0005 | 79.0 | 234551 | 0.4537 | 7.0023 |
131
+ | 0.0005 | 80.0 | 237520 | 0.4544 | 7.2048 |
132
+ | 0.0004 | 81.0 | 240489 | 0.4485 | 7.2167 |
133
+ | 0.0005 | 82.0 | 243458 | 0.4564 | 7.1794 |
134
+ | 0.0004 | 83.0 | 246427 | 0.4608 | 7.2145 |
135
+ | 0.0004 | 84.0 | 249396 | 0.4724 | 7.2098 |
136
+ | 0.0004 | 85.0 | 252365 | 0.4726 | 7.1424 |
137
+ | 0.0004 | 86.0 | 255334 | 0.4754 | 7.2832 |
138
+ | 0.0005 | 87.0 | 258303 | 0.4765 | 7.1709 |
139
+ | 0.0004 | 88.0 | 261272 | 0.4610 | 7.1358 |
140
+ | 0.0004 | 89.0 | 264241 | 0.4697 | 7.0797 |
141
+ | 0.0004 | 90.0 | 267210 | 0.4717 | 7.0913 |
142
+ | 0.0004 | 91.0 | 270179 | 0.4756 | 7.1017 |
143
+ | 0.0004 | 92.0 | 273148 | 0.4766 | 7.2089 |
144
+ | 0.0004 | 93.0 | 276117 | 0.4763 | 7.1057 |
145
+ | 0.0004 | 94.0 | 279086 | 0.4764 | 7.1101 |
146
+ | 0.0004 | 95.0 | 282055 | 0.4759 | 7.2170 |
147
+ | 0.0004 | 96.0 | 285024 | 0.4772 | 7.1104 |
148
+ | 0.0004 | 97.0 | 287993 | 0.4781 | 7.0819 |
149
+ | 0.0004 | 98.0 | 290962 | 0.4798 | 7.0897 |
150
+ | 0.0004 | 99.0 | 293931 | 0.4800 | 7.0872 |
151
+ | 0.0004 | 100.0 | 296900 | 0.4802 | 7.0810 |
152
+
153
+
154
+ ### Framework versions
155
+
156
+ - Transformers 4.36.2
157
+ - Pytorch 2.0.1+cu117
158
+ - Datasets 2.16.1
159
+ - Tokenizers 0.15.0
generation_config.json ADDED
@@ -0,0 +1,240 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alignment_heads": [
3
+ [
4
+ 2,
5
+ 2
6
+ ],
7
+ [
8
+ 3,
9
+ 0
10
+ ],
11
+ [
12
+ 3,
13
+ 2
14
+ ],
15
+ [
16
+ 3,
17
+ 3
18
+ ],
19
+ [
20
+ 3,
21
+ 4
22
+ ],
23
+ [
24
+ 3,
25
+ 5
26
+ ]
27
+ ],
28
+ "begin_suppress_tokens": [
29
+ 220,
30
+ 50257
31
+ ],
32
+ "bos_token_id": 50257,
33
+ "decoder_start_token_id": 50258,
34
+ "eos_token_id": 50257,
35
+ "is_multilingual": true,
36
+ "lang_to_id": {
37
+ "<|af|>": 50327,
38
+ "<|am|>": 50334,
39
+ "<|ar|>": 50272,
40
+ "<|as|>": 50350,
41
+ "<|az|>": 50304,
42
+ "<|ba|>": 50355,
43
+ "<|be|>": 50330,
44
+ "<|bg|>": 50292,
45
+ "<|bn|>": 50302,
46
+ "<|bo|>": 50347,
47
+ "<|br|>": 50309,
48
+ "<|bs|>": 50315,
49
+ "<|ca|>": 50270,
50
+ "<|cs|>": 50283,
51
+ "<|cy|>": 50297,
52
+ "<|da|>": 50285,
53
+ "<|de|>": 50261,
54
+ "<|el|>": 50281,
55
+ "<|en|>": 50259,
56
+ "<|es|>": 50262,
57
+ "<|et|>": 50307,
58
+ "<|eu|>": 50310,
59
+ "<|fa|>": 50300,
60
+ "<|fi|>": 50277,
61
+ "<|fo|>": 50338,
62
+ "<|fr|>": 50265,
63
+ "<|gl|>": 50319,
64
+ "<|gu|>": 50333,
65
+ "<|haw|>": 50352,
66
+ "<|ha|>": 50354,
67
+ "<|he|>": 50279,
68
+ "<|hi|>": 50276,
69
+ "<|hr|>": 50291,
70
+ "<|ht|>": 50339,
71
+ "<|hu|>": 50286,
72
+ "<|hy|>": 50312,
73
+ "<|id|>": 50275,
74
+ "<|is|>": 50311,
75
+ "<|it|>": 50274,
76
+ "<|ja|>": 50266,
77
+ "<|jw|>": 50356,
78
+ "<|ka|>": 50329,
79
+ "<|kk|>": 50316,
80
+ "<|km|>": 50323,
81
+ "<|kn|>": 50306,
82
+ "<|ko|>": 50264,
83
+ "<|la|>": 50294,
84
+ "<|lb|>": 50345,
85
+ "<|ln|>": 50353,
86
+ "<|lo|>": 50336,
87
+ "<|lt|>": 50293,
88
+ "<|lv|>": 50301,
89
+ "<|mg|>": 50349,
90
+ "<|mi|>": 50295,
91
+ "<|mk|>": 50308,
92
+ "<|ml|>": 50296,
93
+ "<|mn|>": 50314,
94
+ "<|mr|>": 50320,
95
+ "<|ms|>": 50282,
96
+ "<|mt|>": 50343,
97
+ "<|my|>": 50346,
98
+ "<|ne|>": 50313,
99
+ "<|nl|>": 50271,
100
+ "<|nn|>": 50342,
101
+ "<|no|>": 50288,
102
+ "<|oc|>": 50328,
103
+ "<|pa|>": 50321,
104
+ "<|pl|>": 50269,
105
+ "<|ps|>": 50340,
106
+ "<|pt|>": 50267,
107
+ "<|ro|>": 50284,
108
+ "<|ru|>": 50263,
109
+ "<|sa|>": 50344,
110
+ "<|sd|>": 50332,
111
+ "<|si|>": 50322,
112
+ "<|sk|>": 50298,
113
+ "<|sl|>": 50305,
114
+ "<|sn|>": 50324,
115
+ "<|so|>": 50326,
116
+ "<|sq|>": 50317,
117
+ "<|sr|>": 50303,
118
+ "<|su|>": 50357,
119
+ "<|sv|>": 50273,
120
+ "<|sw|>": 50318,
121
+ "<|ta|>": 50287,
122
+ "<|te|>": 50299,
123
+ "<|tg|>": 50331,
124
+ "<|th|>": 50289,
125
+ "<|tk|>": 50341,
126
+ "<|tl|>": 50348,
127
+ "<|tr|>": 50268,
128
+ "<|tt|>": 50351,
129
+ "<|uk|>": 50280,
130
+ "<|ur|>": 50290,
131
+ "<|uz|>": 50337,
132
+ "<|vi|>": 50278,
133
+ "<|yi|>": 50335,
134
+ "<|yo|>": 50325,
135
+ "<|zh|>": 50260
136
+ },
137
+ "language": "korean",
138
+ "max_initial_timestamp_index": 50,
139
+ "max_length": 448,
140
+ "no_timestamps_token_id": 50363,
141
+ "pad_token_id": 50257,
142
+ "prev_sot_token_id": 50361,
143
+ "return_timestamps": false,
144
+ "suppress_tokens": [
145
+ 1,
146
+ 2,
147
+ 7,
148
+ 8,
149
+ 9,
150
+ 10,
151
+ 14,
152
+ 25,
153
+ 26,
154
+ 27,
155
+ 28,
156
+ 29,
157
+ 31,
158
+ 58,
159
+ 59,
160
+ 60,
161
+ 61,
162
+ 62,
163
+ 63,
164
+ 90,
165
+ 91,
166
+ 92,
167
+ 93,
168
+ 359,
169
+ 503,
170
+ 522,
171
+ 542,
172
+ 873,
173
+ 893,
174
+ 902,
175
+ 918,
176
+ 922,
177
+ 931,
178
+ 1350,
179
+ 1853,
180
+ 1982,
181
+ 2460,
182
+ 2627,
183
+ 3246,
184
+ 3253,
185
+ 3268,
186
+ 3536,
187
+ 3846,
188
+ 3961,
189
+ 4183,
190
+ 4667,
191
+ 6585,
192
+ 6647,
193
+ 7273,
194
+ 9061,
195
+ 9383,
196
+ 10428,
197
+ 10929,
198
+ 11938,
199
+ 12033,
200
+ 12331,
201
+ 12562,
202
+ 13793,
203
+ 14157,
204
+ 14635,
205
+ 15265,
206
+ 15618,
207
+ 16553,
208
+ 16604,
209
+ 18362,
210
+ 18956,
211
+ 20075,
212
+ 21675,
213
+ 22520,
214
+ 26130,
215
+ 26161,
216
+ 26435,
217
+ 28279,
218
+ 29464,
219
+ 31650,
220
+ 32302,
221
+ 32470,
222
+ 36865,
223
+ 42863,
224
+ 47425,
225
+ 49870,
226
+ 50254,
227
+ 50258,
228
+ 50358,
229
+ 50359,
230
+ 50360,
231
+ 50361,
232
+ 50362
233
+ ],
234
+ "task": "transcribe",
235
+ "task_to_id": {
236
+ "transcribe": 50359,
237
+ "translate": 50358
238
+ },
239
+ "transformers_version": "4.36.2"
240
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0de3f6f227597b3ec76d44da257619fd0124734ac58dfe91b5b982443edcbfb9
3
  size 151061672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc3bb6de279b4a9cb6f6028437a5148f4fa413cf79c37f02ee1b7069f2936559
3
  size 151061672