arampacha commited on
Commit
4b34232
1 Parent(s): 8941767
Files changed (3) hide show
  1. README.md +4 -4
  2. config.json +2 -2
  3. pytorch_model.bin +1 -1
README.md CHANGED
@@ -21,12 +21,12 @@ model-index:
21
  metrics:
22
  - name: Test WER
23
  type: wer
24
- value: 34.38
25
  ---
26
 
27
  # Wav2Vec2-Large-XLSR-53-Ukrainian
28
 
29
- Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on Ukrainian using the [Common Voice](https://huggingface.co/datasets/common_voice) dataset.
30
 
31
  When using this model, make sure that your speech input is sampled at 16kHz.
32
 
@@ -118,10 +118,10 @@ print("WER: {:2f}".format(100 * wer.compute(predictions=result["pred_strings"],
118
 
119
  ```
120
 
121
- **Test Result**: 34.38
122
 
123
  ## Training
124
 
125
- The Common Voice `train`, `validation`.
126
 
127
  The script used for training will be available [here](https://github.com/arampacha/hf-sprint-xlsr) soon.
 
21
  metrics:
22
  - name: Test WER
23
  type: wer
24
+ value: 31.30
25
  ---
26
 
27
  # Wav2Vec2-Large-XLSR-53-Ukrainian
28
 
29
+ Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on Ukrainian using the [Common Voice](https://huggingface.co/datasets/common_voice) and sample of [M-AILABS Ukrainian Corpus](https://www.caito.de/2019/01/the-m-ailabs-speech-dataset/) datasets.
30
 
31
  When using this model, make sure that your speech input is sampled at 16kHz.
32
 
 
118
 
119
  ```
120
 
121
+ **Test Result**: 31.30
122
 
123
  ## Training
124
 
125
+ The Common Voice `train`, `validation` and sample of the M-AILABS Ukrainian corpus.
126
 
127
  The script used for training will be available [here](https://github.com/arampacha/hf-sprint-xlsr) soon.
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "facebook/wav2vec2-large-xlsr-53",
3
  "activation_dropout": 0.1,
4
  "apply_spec_augment": true,
5
  "architectures": [
@@ -42,7 +42,7 @@
42
  "feat_extract_activation": "gelu",
43
  "feat_extract_dropout": 0.0,
44
  "feat_extract_norm": "layer",
45
- "feat_proj_dropout": 0.008,
46
  "final_dropout": 0.0,
47
  "gradient_checkpointing": true,
48
  "hidden_act": "gelu",
 
1
  {
2
+ "_name_or_path": "arampacha/wav2vec2-large-xlsr-ukrainian",
3
  "activation_dropout": 0.1,
4
  "apply_spec_augment": true,
5
  "architectures": [
 
42
  "feat_extract_activation": "gelu",
43
  "feat_extract_dropout": 0.0,
44
  "feat_extract_norm": "layer",
45
+ "feat_proj_dropout": 0.004,
46
  "final_dropout": 0.0,
47
  "gradient_checkpointing": true,
48
  "hidden_act": "gelu",
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7fb8c6fd0cf3288fa3cbe26a6906bb29a5a5eea3b237bff64b406614bfb7ed04
3
  size 1262118359
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:797a5b497ee33bf44024d2596e6bebf6591ba15e6a3d18df7e147bb1ed92d1be
3
  size 1262118359