quinnb commited on
Commit
a62b122
·
verified ·
1 Parent(s): ec0efea

End of training

Browse files
README.md CHANGED
@@ -2,7 +2,7 @@
2
  language:
3
  - hi
4
  license: apache-2.0
5
- base_model: openai/whisper-large-v3
6
  tags:
7
  - generated_from_trainer
8
  datasets:
@@ -17,7 +17,7 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # Whisper Large v3 Trained on Hindi
19
 
20
- This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on the Common Voice 17.0 dataset.
21
 
22
  ## Model description
23
 
@@ -37,17 +37,24 @@ More information needed
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 1e-05
40
- - train_batch_size: 4
41
- - eval_batch_size: 64
42
  - seed: 42
 
 
43
  - gradient_accumulation_steps: 16
44
  - total_train_batch_size: 64
 
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
  - lr_scheduler_warmup_steps: 500
48
- - training_steps: 5000
49
  - mixed_precision_training: Native AMP
50
 
 
 
 
 
51
  ### Framework versions
52
 
53
  - Transformers 4.41.1
 
2
  language:
3
  - hi
4
  license: apache-2.0
5
+ base_model: quinnb/whisper-Large-v3-hindi
6
  tags:
7
  - generated_from_trainer
8
  datasets:
 
17
 
18
  # Whisper Large v3 Trained on Hindi
19
 
20
+ This model is a fine-tuned version of [quinnb/whisper-Large-v3-hindi](https://huggingface.co/quinnb/whisper-Large-v3-hindi) on the Common Voice 17.0 dataset.
21
 
22
  ## Model description
23
 
 
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 1e-05
40
+ - train_batch_size: 1
41
+ - eval_batch_size: 16
42
  - seed: 42
43
+ - distributed_type: multi-GPU
44
+ - num_devices: 4
45
  - gradient_accumulation_steps: 16
46
  - total_train_batch_size: 64
47
+ - total_eval_batch_size: 64
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
  - lr_scheduler_warmup_steps: 500
51
+ - training_steps: 2000
52
  - mixed_precision_training: Native AMP
53
 
54
+ ### Training results
55
+
56
+
57
+
58
  ### Framework versions
59
 
60
  - Transformers 4.41.1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "openai/whisper-large-v3",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
@@ -42,7 +42,7 @@
42
  "num_mel_bins": 128,
43
  "pad_token_id": 50256,
44
  "scale_embedding": false,
45
- "torch_dtype": "float32",
46
  "transformers_version": "4.41.1",
47
  "use_cache": true,
48
  "use_weighted_layer_sum": false,
 
1
  {
2
+ "_name_or_path": "quinnb/whisper-Large-v3-hindi",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
 
42
  "num_mel_bins": 128,
43
  "pad_token_id": 50256,
44
  "scale_embedding": false,
45
+ "torch_dtype": "float16",
46
  "transformers_version": "4.41.1",
47
  "use_cache": true,
48
  "use_weighted_layer_sum": false,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a02fde508dec44d0cbb27ad745f11c77a2b742eaecb29d3ff60bc4191c846f9b
3
  size 3219908024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1ff3694f9fa53a69eba7760ad5fbad10174a1c715a461eaee7b275df55e6f3e
3
  size 3219908024
runs/May31_01-34-01_bhrathgpt-v1/events.out.tfevents.1717119265.bhrathgpt-v1.1840323.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:378ae266fc3eb9126b0c42b496a98a9dd88ee58415d98a07d6981475e4ac71ff
3
+ size 5616
runs/May31_01-38-23_bhrathgpt-v1/events.out.tfevents.1717119524.bhrathgpt-v1.1841996.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:845b205994f49a9b6e665ca08a08f27b9172c8e99d514ce4af4cf57cca933744
3
+ size 22832
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34facea37d4705d15a4fdc1db99d081ec602439fc3e74151978af6c6ac17b08e
3
- size 4783
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3be1cc8c0ac887ec1ab6d0469bb80e0bf977a2e7ee3554593ef4f6ad601191b3
3
+ size 5615