vasista22 commited on
Commit
edb3e49
1 Parent(s): 7f8d15f

first commit

Browse files
Files changed (2) hide show
  1. README.md +56 -0
  2. pytorch_model.bin +1 -1
README.md ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - te
4
+ license: apache-2.0
5
+ tags:
6
+ - whisper-event
7
+ - generated_from_trainer
8
+ metrics:
9
+ - wer
10
+ model-index:
11
+ - name: Whisper Telugu Tiny - Vasista Sai Lodagala
12
+ results:
13
+ - task:
14
+ type: automatic-speech-recognition
15
+ name: Automatic Speech Recognition
16
+ dataset:
17
+ name: google/fleurs
18
+ type: google/fleurs
19
+ config: te_in
20
+ split: test
21
+ metrics:
22
+ - type: wer
23
+ value: 20.0
24
+ name: WER
25
+ ---
26
+
27
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
28
+ should probably proofread and complete it, then remove this comment. -->
29
+
30
+ # Whisper Telugu Tiny
31
+
32
+ This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the Telugu data available from multiple publicly available ASR corpuses.
33
+ It has been fine-tuned as a part of the Whisper fine-tuning sprint.
34
+
35
+ ## Training and evaluation data at Speech Lab, IITM
36
+
37
+ Training Data: CSTD IIIT-H ASR Corpus, ULCA ASR Corpus, Shrutilipi ASR Corpus, Microsoft Research Telugu Corpus (Train+Dev), Babel ASR Corpus, Google/Fleurs (Train+Dev) set.
38
+ Evaluation Data: Babel Test, Microsoft Research Telugu Corpus Test, Google/Fleurs Test set, OpenSLR.
39
+
40
+ ## Training procedure
41
+
42
+ ### Training hyperparameters
43
+
44
+ The following hyperparameters were used during training:
45
+ - learning_rate: 5e-05
46
+ - train_batch_size: 88
47
+ - eval_batch_size: 88
48
+ - seed: 22
49
+ - optimizer: adamw_bnb_8bit
50
+ - lr_scheduler_type: linear
51
+ - lr_scheduler_warmup_steps: 15000
52
+ - training_steps: 14652 (terminated upon convergence. Initially set to 85952 steps)
53
+ - mixed_precision_training: True
54
+
55
+ ## Acknowledgement
56
+ This work was done at Speech Lab, IITM. The compute resources for this work were funded by "Bhashini: National Language translation Mission" project of the Ministry of Electronics and Information Technology (MeitY), Government of India.
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6108827bfa1a2d77e686ca1cb2ac9cda9a1c3579277730814bc7d2c11b915dd5
3
  size 151097331
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8daec184d3b36598f8c21c0832ca8c783e4fac17589f53f6f33be87a5c4e9bd5
3
  size 151097331