Scrya commited on
Commit
d65e4a0
1 Parent(s): da4149b

update model card README.md

Browse files
README.md ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - yue
4
+ license: apache-2.0
5
+ tags:
6
+ - whisper-event
7
+ - generated_from_trainer
8
+ datasets:
9
+ - mozilla-foundation/common_voice_11_0
10
+ model-index:
11
+ - name: Whisper Large V2 - Cantonese - Augmented
12
+ results: []
13
+ ---
14
+
15
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
+
18
+ # Whisper Large V2 - Cantonese - Augmented
19
+
20
+ This model is a fine-tuned version of [openai/whisper-large-v2](https://huggingface.co/openai/whisper-large-v2) on the mozilla-foundation/common_voice_11_0 dataset.
21
+ It achieves the following results on the evaluation set:
22
+ - Loss: 0.1828
23
+ - Cer: 6.2133
24
+
25
+ ## Model description
26
+
27
+ More information needed
28
+
29
+ ## Intended uses & limitations
30
+
31
+ More information needed
32
+
33
+ ## Training and evaluation data
34
+
35
+ More information needed
36
+
37
+ ## Training procedure
38
+
39
+ ### Training hyperparameters
40
+
41
+ The following hyperparameters were used during training:
42
+ - learning_rate: 1e-05
43
+ - train_batch_size: 8
44
+ - eval_batch_size: 4
45
+ - seed: 42
46
+ - gradient_accumulation_steps: 4
47
+ - total_train_batch_size: 32
48
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
+ - lr_scheduler_type: linear
50
+ - lr_scheduler_warmup_steps: 100
51
+ - training_steps: 1000
52
+ - mixed_precision_training: Native AMP
53
+
54
+ ### Training results
55
+
56
+ | Training Loss | Epoch | Step | Validation Loss | Cer |
57
+ |:-------------:|:-----:|:----:|:---------------:|:------:|
58
+ | 0.1126 | 1.21 | 200 | 0.1666 | 7.3103 |
59
+ | 0.0467 | 2.42 | 400 | 0.1610 | 6.9419 |
60
+ | 0.0217 | 3.63 | 600 | 0.1621 | 6.3874 |
61
+ | 0.008 | 4.85 | 800 | 0.1699 | 6.3064 |
62
+ | 0.0023 | 6.06 | 1000 | 0.1828 | 6.2133 |
63
+
64
+
65
+ ### Framework versions
66
+
67
+ - Transformers 4.26.0.dev0
68
+ - Pytorch 1.13.1+cu117
69
+ - Datasets 2.8.1.dev0
70
+ - Tokenizers 0.13.2
fine-tune-whisper-non-streaming-id.ipynb CHANGED
@@ -47786,8 +47786,38 @@
47786
  "Saving model checkpoint to ./\n",
47787
  "Configuration saved in ./config.json\n",
47788
  "Model weights saved in ./pytorch_model.bin\n",
47789
- "Feature extractor saved in ./preprocessor_config.json\n"
 
 
47790
  ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47791
  }
47792
  ],
47793
  "source": [
 
47786
  "Saving model checkpoint to ./\n",
47787
  "Configuration saved in ./config.json\n",
47788
  "Model weights saved in ./pytorch_model.bin\n",
47789
+ "Feature extractor saved in ./preprocessor_config.json\n",
47790
+ "Several commits (2) will be pushed upstream.\n",
47791
+ "The progress bars may be unreliable.\n"
47792
  ]
47793
+ },
47794
+ {
47795
+ "data": {
47796
+ "application/vnd.jupyter.widget-view+json": {
47797
+ "model_id": "0e2d9a13a7f143fcb814130c727cebe0",
47798
+ "version_major": 2,
47799
+ "version_minor": 0
47800
+ },
47801
+ "text/plain": [
47802
+ "Upload file pytorch_model.bin: 0%| | 1.00/5.75G [00:00<?, ?B/s]"
47803
+ ]
47804
+ },
47805
+ "metadata": {},
47806
+ "output_type": "display_data"
47807
+ },
47808
+ {
47809
+ "data": {
47810
+ "application/vnd.jupyter.widget-view+json": {
47811
+ "model_id": "7daa9c192f004d8391c9c8fe96f3a757",
47812
+ "version_major": 2,
47813
+ "version_minor": 0
47814
+ },
47815
+ "text/plain": [
47816
+ "Upload file runs/Dec19_12-09-59_129-146-74-165/events.out.tfevents.1671451847.129-146-74-165.127332.2: 0%| …"
47817
+ ]
47818
+ },
47819
+ "metadata": {},
47820
+ "output_type": "display_data"
47821
  }
47822
  ],
47823
  "source": [