Deepakr07 commited on
Commit
66fd18c
·
1 Parent(s): 3d4c35f
Files changed (1) hide show
  1. README.md +13 -13
README.md CHANGED
@@ -1,24 +1,25 @@
1
  ---
2
  language:
3
- - en
4
  license: apache-2.0
5
  base_model: openai/whisper-tiny
6
  tags:
7
- - generated_from_trainer
8
  metrics:
9
- - wer
10
  model-index:
11
- - name: Deepakr07
12
- results: []
13
  ---
14
 
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
  should probably proofread and complete it, then remove this comment. -->
17
 
18
- # whispertiny-shreyas
19
 
20
  This model is a fine-tuned version of [Whisper Tiny](https://huggingface.co/openai/whisper-tiny) on the [AI4Bharat-svarah](https://github.com/AI4Bharat/Svarah) dataset.
21
  It achieves the following results on the evaluation set:
 
22
  - Loss: 0.5414
23
  - Wer: 22.8322
24
 
@@ -30,14 +31,14 @@ The models were trained on either English-only data or multilingual data. The En
30
 
31
  Whisper checkpoints come in five configurations of varying model sizes. The smallest four are trained on either English-only or multilingual data. The largest checkpoints are multilingual only. All ten of the pre-trained checkpoints are available on the Hugging Face Hub. The checkpoints are summarised in the following table with links to the models on the Hub:
32
 
33
-
34
  ## Training procedure
35
 
36
- Refer to [Sanchit's blog](https://huggingface.co/blog/fine-tune-whisper) and make changes according to the dependencies' version you have.
37
 
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
 
41
  - learning_rate: 1e-05
42
  - train_batch_size: 8
43
  - eval_batch_size: 8
@@ -52,11 +53,10 @@ The following hyperparameters were used during training:
52
 
53
  ### Training results
54
 
55
- | Training Loss | Epoch | Step | Validation Loss | Wer |
56
- |:-------------:|:------:|:----:|:---------------:|:-------:|
57
- | 0.2412 | 2.6702 | 1000 | 0.5319 | 22.8914 |
58
- | 0.1071 | 5.3405 | 2000 | 0.5414 | 22.8322 |
59
-
60
 
61
  ### Framework versions
62
 
 
1
  ---
2
  language:
3
+ - en
4
  license: apache-2.0
5
  base_model: openai/whisper-tiny
6
  tags:
7
+ - generated_from_trainer
8
  metrics:
9
+ - wer
10
  model-index:
11
+ - name: Deepakr07
12
+ results: []
13
  ---
14
 
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
  should probably proofread and complete it, then remove this comment. -->
17
 
18
+ # Deepakr07-whisper_finetune
19
 
20
  This model is a fine-tuned version of [Whisper Tiny](https://huggingface.co/openai/whisper-tiny) on the [AI4Bharat-svarah](https://github.com/AI4Bharat/Svarah) dataset.
21
  It achieves the following results on the evaluation set:
22
+
23
  - Loss: 0.5414
24
  - Wer: 22.8322
25
 
 
31
 
32
  Whisper checkpoints come in five configurations of varying model sizes. The smallest four are trained on either English-only or multilingual data. The largest checkpoints are multilingual only. All ten of the pre-trained checkpoints are available on the Hugging Face Hub. The checkpoints are summarised in the following table with links to the models on the Hub:
33
 
 
34
  ## Training procedure
35
 
36
+ Refer to [Sanchit's blog](https://huggingface.co/blog/fine-tune-whisper) and make changes according to the dependencies' version you have.
37
 
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
+
42
  - learning_rate: 1e-05
43
  - train_batch_size: 8
44
  - eval_batch_size: 8
 
53
 
54
  ### Training results
55
 
56
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
57
+ | :-----------: | :----: | :--: | :-------------: | :-----: |
58
+ | 0.2412 | 2.6702 | 1000 | 0.5319 | 22.8914 |
59
+ | 0.1071 | 5.3405 | 2000 | 0.5414 | 22.8322 |
 
60
 
61
  ### Framework versions
62