DewiBrynJones commited on
Commit
5dfd7cd
1 Parent(s): 9a56b7c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -12
README.md CHANGED
@@ -6,29 +6,33 @@ metrics:
6
  model-index:
7
  - name: whisper-tiny-ft-cy
8
  results: []
 
 
 
 
 
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
- # whisper-tiny-ft-cy
15
 
16
- This model was trained from scratch on an unknown dataset.
17
- It achieves the following results on the evaluation set:
18
- - Loss: 0.7176
19
- - Wer: 53.1135
20
-
21
- ## Model description
22
-
23
- More information needed
24
 
25
  ## Intended uses & limitations
26
 
27
- More information needed
 
28
 
29
  ## Training and evaluation data
30
 
31
- More information needed
 
 
 
32
 
33
  ## Training procedure
34
 
@@ -61,4 +65,4 @@ The following hyperparameters were used during training:
61
  - Transformers 4.37.2
62
  - Pytorch 2.2.0+cu121
63
  - Datasets 2.16.1
64
- - Tokenizers 0.15.1
 
6
  model-index:
7
  - name: whisper-tiny-ft-cy
8
  results: []
9
+ license: apache-2.0
10
+ language:
11
+ - cy
12
+ - en
13
+ pipeline_tag: automatic-speech-recognition
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
  should probably proofread and complete it, then remove this comment. -->
18
 
19
+ # whisper-tiny-ft-cy-en
20
 
21
+ This model is a fine-tune of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) using custom splits from
22
+ Common Voice 16.1 Welsh and English datasets as well as normalized verbatim transcriptions from
23
+ [techiaith/banc-trawsgrifiadau-bangor](https://huggingface.co/datasets/techiaith/banc-trawsgrifiadau-bangor)
 
 
 
 
 
24
 
25
  ## Intended uses & limitations
26
 
27
+ Due to its small size, this model is intended to be used as the basis for offline speech recognition on devices such as
28
+ Android phones.
29
 
30
  ## Training and evaluation data
31
 
32
+ It achieves the following results on the evaluation set:
33
+
34
+ - Loss: 0.7176
35
+ - Wer: 53.1135
36
 
37
  ## Training procedure
38
 
 
65
  - Transformers 4.37.2
66
  - Pytorch 2.2.0+cu121
67
  - Datasets 2.16.1
68
+ - Tokenizers 0.15.1