imvladikon commited on
Commit
c9ee198
1 Parent(s): c67c171

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -5
README.md CHANGED
@@ -16,25 +16,37 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # wav2vec2-xls-r-300m-hebrew
18
 
19
- This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the private datasets in 2 stages - firstly was fine-tuned on a small dataset with good samples and it achieves the following results on the evaluation set with the dataset:
20
 
 
21
 
22
  | split |size(gb) | n_samples | duration(hrs)| |
23
  |---|---|---|---|---|
24
  |train|4.19| 20306 | 28 | |
25
  |dev |1.05| 5076 | 7 | |
26
 
 
 
 
 
 
 
 
 
 
 
 
27
  - Loss: 0.5438
28
  - WER: 0.1773
29
 
30
- and on a large dataset
31
  - WER: 0.3811
32
 
33
- Then the obtained model was fine-tuned on a large dataset with the small good dataset, with various samples from different sources, and with an unlabeled dataset that was weakly labeled using a previously trained model.
34
- on a small dataset from previous step achieves
35
  - WER: 0.1697
36
 
37
- on a whole dataset
38
  - Loss: 0.4502
39
  - WER: 0.2318
40
 
 
16
 
17
  # wav2vec2-xls-r-300m-hebrew
18
 
19
+ This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the private datasets in 2 stages - firstly was fine-tuned on a small dataset with good samples Then the obtained model was fine-tuned on a large dataset with the small good dataset, with various samples from different sources, and with an unlabeled dataset that was weakly labeled using a previously trained model.
20
 
21
+ Small dataset:
22
 
23
  | split |size(gb) | n_samples | duration(hrs)| |
24
  |---|---|---|---|---|
25
  |train|4.19| 20306 | 28 | |
26
  |dev |1.05| 5076 | 7 | |
27
 
28
+ Large dataset:
29
+
30
+ | split |size(gb) | n_samples | duration(hrs)| |
31
+ |---|---|---|---|---|
32
+ |train|12.3| 90777 | 69 | |
33
+ |dev |1.05| 20246 | 14* | |
34
+ (*weakly labeled data wasn't used in validation set)
35
+
36
+ After firts training it achieves:
37
+
38
+ on small dataset
39
  - Loss: 0.5438
40
  - WER: 0.1773
41
 
42
+ on large dataset
43
  - WER: 0.3811
44
 
45
+ after second training:
46
+ on small dataset
47
  - WER: 0.1697
48
 
49
+ on large dataset
50
  - Loss: 0.4502
51
  - WER: 0.2318
52