End of training
Browse files- README.md +5 -12
- config.json +1 -3
- generation_config.json +2 -10
- model-00001-of-00002.safetensors +3 -0
- model-00002-of-00002.safetensors +3 -0
- model.safetensors.index.json +0 -0
- training_args.bin +2 -2
README.md
CHANGED
@@ -17,7 +17,7 @@ should probably proofread and complete it, then remove this comment. -->
|
|
17 |
|
18 |
# Whisper Large v3 Trained on Hindi
|
19 |
|
20 |
-
This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on the
|
21 |
|
22 |
## Model description
|
23 |
|
@@ -37,24 +37,17 @@ More information needed
|
|
37 |
|
38 |
The following hyperparameters were used during training:
|
39 |
- learning_rate: 1e-05
|
40 |
-
- train_batch_size:
|
41 |
-
- eval_batch_size:
|
42 |
- seed: 42
|
43 |
-
- distributed_type: multi-GPU
|
44 |
-
- num_devices: 4
|
45 |
- gradient_accumulation_steps: 16
|
46 |
- total_train_batch_size: 64
|
47 |
-
- total_eval_batch_size: 64
|
48 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
49 |
- lr_scheduler_type: linear
|
50 |
-
- lr_scheduler_warmup_steps:
|
51 |
-
- training_steps:
|
52 |
- mixed_precision_training: Native AMP
|
53 |
|
54 |
-
### Training results
|
55 |
-
|
56 |
-
|
57 |
-
|
58 |
### Framework versions
|
59 |
|
60 |
- Transformers 4.41.1
|
|
|
17 |
|
18 |
# Whisper Large v3 Trained on Hindi
|
19 |
|
20 |
+
This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on the Common Voice 17.0 dataset.
|
21 |
|
22 |
## Model description
|
23 |
|
|
|
37 |
|
38 |
The following hyperparameters were used during training:
|
39 |
- learning_rate: 1e-05
|
40 |
+
- train_batch_size: 4
|
41 |
+
- eval_batch_size: 64
|
42 |
- seed: 42
|
|
|
|
|
43 |
- gradient_accumulation_steps: 16
|
44 |
- total_train_batch_size: 64
|
|
|
45 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
46 |
- lr_scheduler_type: linear
|
47 |
+
- lr_scheduler_warmup_steps: 500
|
48 |
+
- training_steps: 2500
|
49 |
- mixed_precision_training: Native AMP
|
50 |
|
|
|
|
|
|
|
|
|
51 |
### Framework versions
|
52 |
|
53 |
- Transformers 4.41.1
|
config.json
CHANGED
@@ -25,7 +25,6 @@
|
|
25 |
"encoder_layerdrop": 0.0,
|
26 |
"encoder_layers": 32,
|
27 |
"eos_token_id": 50257,
|
28 |
-
"forced_decoder_ids": null,
|
29 |
"init_std": 0.02,
|
30 |
"is_encoder_decoder": true,
|
31 |
"mask_feature_length": 10,
|
@@ -43,8 +42,7 @@
|
|
43 |
"num_mel_bins": 128,
|
44 |
"pad_token_id": 50256,
|
45 |
"scale_embedding": false,
|
46 |
-
"
|
47 |
-
"torch_dtype": "float16",
|
48 |
"transformers_version": "4.41.1",
|
49 |
"use_cache": true,
|
50 |
"use_weighted_layer_sum": false,
|
|
|
25 |
"encoder_layerdrop": 0.0,
|
26 |
"encoder_layers": 32,
|
27 |
"eos_token_id": 50257,
|
|
|
28 |
"init_std": 0.02,
|
29 |
"is_encoder_decoder": true,
|
30 |
"mask_feature_length": 10,
|
|
|
42 |
"num_mel_bins": 128,
|
43 |
"pad_token_id": 50256,
|
44 |
"scale_embedding": false,
|
45 |
+
"torch_dtype": "float32",
|
|
|
46 |
"transformers_version": "4.41.1",
|
47 |
"use_cache": true,
|
48 |
"use_weighted_layer_sum": false,
|
generation_config.json
CHANGED
@@ -48,16 +48,6 @@
|
|
48 |
"bos_token_id": 50257,
|
49 |
"decoder_start_token_id": 50258,
|
50 |
"eos_token_id": 50257,
|
51 |
-
"forced_decoder_ids": [
|
52 |
-
[
|
53 |
-
1,
|
54 |
-
null
|
55 |
-
],
|
56 |
-
[
|
57 |
-
2,
|
58 |
-
50360
|
59 |
-
]
|
60 |
-
],
|
61 |
"is_multilingual": true,
|
62 |
"lang_to_id": {
|
63 |
"<|af|>": 50327,
|
@@ -161,6 +151,7 @@
|
|
161 |
"<|yue|>": 50358,
|
162 |
"<|zh|>": 50260
|
163 |
},
|
|
|
164 |
"max_initial_timestamp_index": 50,
|
165 |
"max_length": 448,
|
166 |
"no_timestamps_token_id": 50364,
|
@@ -257,6 +248,7 @@
|
|
257 |
50362,
|
258 |
50363
|
259 |
],
|
|
|
260 |
"task_to_id": {
|
261 |
"transcribe": 50360,
|
262 |
"translate": 50359
|
|
|
48 |
"bos_token_id": 50257,
|
49 |
"decoder_start_token_id": 50258,
|
50 |
"eos_token_id": 50257,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
51 |
"is_multilingual": true,
|
52 |
"lang_to_id": {
|
53 |
"<|af|>": 50327,
|
|
|
151 |
"<|yue|>": 50358,
|
152 |
"<|zh|>": 50260
|
153 |
},
|
154 |
+
"language": "hindi",
|
155 |
"max_initial_timestamp_index": 50,
|
156 |
"max_length": 448,
|
157 |
"no_timestamps_token_id": 50364,
|
|
|
248 |
50362,
|
249 |
50363
|
250 |
],
|
251 |
+
"task": "transcribe",
|
252 |
"task_to_id": {
|
253 |
"transcribe": 50360,
|
254 |
"translate": 50359
|
model-00001-of-00002.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:08e0005225b3dbaf55dd13ac62926cc7e02c1025d66fa375e6fb305ff79cd4f9
|
3 |
+
size 4993448880
|
model-00002-of-00002.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:630ca774672856d2e0e39a702e590f635a1cfc5726a64b6578ab46dd367369a9
|
3 |
+
size 1180663192
|
model.safetensors.index.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c16c49102d4d640aad5c87fd2ada677c9a727a75db44136bf2fca0c34d6f3240
|
3 |
+
size 4783
|