End of training
Browse files
README.md
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
language:
|
3 |
- hi
|
4 |
license: apache-2.0
|
5 |
-
base_model:
|
6 |
tags:
|
7 |
- generated_from_trainer
|
8 |
datasets:
|
@@ -17,7 +17,7 @@ should probably proofread and complete it, then remove this comment. -->
|
|
17 |
|
18 |
# Whisper Large v3 Trained on Hindi
|
19 |
|
20 |
-
This model is a fine-tuned version of [
|
21 |
|
22 |
## Model description
|
23 |
|
@@ -37,17 +37,24 @@ More information needed
|
|
37 |
|
38 |
The following hyperparameters were used during training:
|
39 |
- learning_rate: 1e-05
|
40 |
-
- train_batch_size:
|
41 |
-
- eval_batch_size:
|
42 |
- seed: 42
|
|
|
|
|
43 |
- gradient_accumulation_steps: 16
|
44 |
- total_train_batch_size: 64
|
|
|
45 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
46 |
- lr_scheduler_type: linear
|
47 |
- lr_scheduler_warmup_steps: 500
|
48 |
-
- training_steps:
|
49 |
- mixed_precision_training: Native AMP
|
50 |
|
|
|
|
|
|
|
|
|
51 |
### Framework versions
|
52 |
|
53 |
- Transformers 4.41.1
|
|
|
2 |
language:
|
3 |
- hi
|
4 |
license: apache-2.0
|
5 |
+
base_model: quinnb/whisper-Large-v3-hindi
|
6 |
tags:
|
7 |
- generated_from_trainer
|
8 |
datasets:
|
|
|
17 |
|
18 |
# Whisper Large v3 Trained on Hindi
|
19 |
|
20 |
+
This model is a fine-tuned version of [quinnb/whisper-Large-v3-hindi](https://huggingface.co/quinnb/whisper-Large-v3-hindi) on the Common Voice 17.0 dataset.
|
21 |
|
22 |
## Model description
|
23 |
|
|
|
37 |
|
38 |
The following hyperparameters were used during training:
|
39 |
- learning_rate: 1e-05
|
40 |
+
- train_batch_size: 1
|
41 |
+
- eval_batch_size: 16
|
42 |
- seed: 42
|
43 |
+
- distributed_type: multi-GPU
|
44 |
+
- num_devices: 4
|
45 |
- gradient_accumulation_steps: 16
|
46 |
- total_train_batch_size: 64
|
47 |
+
- total_eval_batch_size: 64
|
48 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
49 |
- lr_scheduler_type: linear
|
50 |
- lr_scheduler_warmup_steps: 500
|
51 |
+
- training_steps: 2000
|
52 |
- mixed_precision_training: Native AMP
|
53 |
|
54 |
+
### Training results
|
55 |
+
|
56 |
+
|
57 |
+
|
58 |
### Framework versions
|
59 |
|
60 |
- Transformers 4.41.1
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"activation_dropout": 0.0,
|
4 |
"activation_function": "gelu",
|
5 |
"apply_spec_augment": false,
|
@@ -42,7 +42,7 @@
|
|
42 |
"num_mel_bins": 128,
|
43 |
"pad_token_id": 50256,
|
44 |
"scale_embedding": false,
|
45 |
-
"torch_dtype": "
|
46 |
"transformers_version": "4.41.1",
|
47 |
"use_cache": true,
|
48 |
"use_weighted_layer_sum": false,
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "quinnb/whisper-Large-v3-hindi",
|
3 |
"activation_dropout": 0.0,
|
4 |
"activation_function": "gelu",
|
5 |
"apply_spec_augment": false,
|
|
|
42 |
"num_mel_bins": 128,
|
43 |
"pad_token_id": 50256,
|
44 |
"scale_embedding": false,
|
45 |
+
"torch_dtype": "float16",
|
46 |
"transformers_version": "4.41.1",
|
47 |
"use_cache": true,
|
48 |
"use_weighted_layer_sum": false,
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3219908024
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d1ff3694f9fa53a69eba7760ad5fbad10174a1c715a461eaee7b275df55e6f3e
|
3 |
size 3219908024
|
runs/May31_01-34-01_bhrathgpt-v1/events.out.tfevents.1717119265.bhrathgpt-v1.1840323.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:378ae266fc3eb9126b0c42b496a98a9dd88ee58415d98a07d6981475e4ac71ff
|
3 |
+
size 5616
|
runs/May31_01-38-23_bhrathgpt-v1/events.out.tfevents.1717119524.bhrathgpt-v1.1841996.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:845b205994f49a9b6e665ca08a08f27b9172c8e99d514ce4af4cf57cca933744
|
3 |
+
size 22832
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3be1cc8c0ac887ec1ab6d0469bb80e0bf977a2e7ee3554593ef4f6ad601191b3
|
3 |
+
size 5615
|