j0eBee commited on
Commit
a13bd88
1 Parent(s): 0fddae3

Model save

Browse files
README.md CHANGED
@@ -1,15 +1,10 @@
1
  ---
2
- language:
3
- - tw
4
  license: apache-2.0
5
  base_model: facebook/wav2vec2-xls-r-300m
6
  tags:
7
- - automatic-speech-recognition
8
- - mozilla-foundation/common_voice_16_1
9
- - mms
10
  - generated_from_trainer
11
  datasets:
12
- - common_voice_16_1
13
  metrics:
14
  - wer
15
  model-index:
@@ -19,11 +14,11 @@ model-index:
19
  name: Automatic Speech Recognition
20
  type: automatic-speech-recognition
21
  dataset:
22
- name: MOZILLA-FOUNDATION/COMMON_VOICE_16_1 - TW
23
- type: common_voice_16_1
24
  config: tw
25
  split: None
26
- args: 'Config: tw, Training split: train, Eval split: validation+test'
27
  metrics:
28
  - name: Wer
29
  type: wer
@@ -35,11 +30,11 @@ should probably proofread and complete it, then remove this comment. -->
35
 
36
  # wav2vec2-twi-adapter
37
 
38
- This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the MOZILLA-FOUNDATION/COMMON_VOICE_16_1 - TW dataset.
39
  It achieves the following results on the evaluation set:
40
- - Loss: 35.6947
41
  - Wer: 1.0
42
- - Cer: 5.1304
43
 
44
  ## Model description
45
 
@@ -70,11 +65,15 @@ The following hyperparameters were used during training:
70
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
71
  - lr_scheduler_type: linear
72
  - lr_scheduler_warmup_steps: 100
73
- - num_epochs: 4.0
74
  - mixed_precision_training: Native AMP
75
 
76
  ### Training results
77
 
 
 
 
 
78
 
79
 
80
  ### Framework versions
 
1
  ---
 
 
2
  license: apache-2.0
3
  base_model: facebook/wav2vec2-xls-r-300m
4
  tags:
 
 
 
5
  - generated_from_trainer
6
  datasets:
7
+ - common_voice_17_0
8
  metrics:
9
  - wer
10
  model-index:
 
14
  name: Automatic Speech Recognition
15
  type: automatic-speech-recognition
16
  dataset:
17
+ name: common_voice_17_0
18
+ type: common_voice_17_0
19
  config: tw
20
  split: None
21
+ args: tw
22
  metrics:
23
  - name: Wer
24
  type: wer
 
30
 
31
  # wav2vec2-twi-adapter
32
 
33
+ This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common_voice_17_0 dataset.
34
  It achieves the following results on the evaluation set:
35
+ - Loss: 2.4281
36
  - Wer: 1.0
37
+ - Cer: 1.0
38
 
39
  ## Model description
40
 
 
65
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
66
  - lr_scheduler_type: linear
67
  - lr_scheduler_warmup_steps: 100
68
+ - num_epochs: 30.0
69
  - mixed_precision_training: Native AMP
70
 
71
  ### Training results
72
 
73
+ | Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
74
+ |:-------------:|:-------:|:----:|:---------------:|:---:|:---:|
75
+ | No log | 11.1111 | 50 | 5.8930 | 1.0 | 1.0 |
76
+ | No log | 22.2222 | 100 | 2.4281 | 1.0 | 1.0 |
77
 
78
 
79
  ### Framework versions
config.json CHANGED
@@ -103,6 +103,6 @@
103
  "torch_dtype": "float32",
104
  "transformers_version": "4.41.1",
105
  "use_weighted_layer_sum": false,
106
- "vocab_size": 32,
107
  "xvector_output_dim": 512
108
  }
 
103
  "torch_dtype": "float32",
104
  "transformers_version": "4.41.1",
105
  "use_weighted_layer_sum": false,
106
+ "vocab_size": 38,
107
  "xvector_output_dim": 512
108
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a8642a139276c58649ec3962b6c38c21adb554c0ee7e666d71b9e975a3ae3f9d
3
- size 1265398792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45e47801a0c10552648c6368e667f1ed98845f5e54174df8f8970feb7816a7c0
3
+ size 1265423392
runs/May24_03-32-47_cc57b0143c57/events.out.tfevents.1716521591.cc57b0143c57.549.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a236f05672d60cc3b45a221cd78f74212a61781b26eb4790f5c273961d55261
3
+ size 4136
runs/May24_03-34-21_cc57b0143c57/events.out.tfevents.1716521688.cc57b0143c57.714.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84a152e8574f56a4bea0699260f83faec2a449b2e46fac82234a702755ac5a4b
3
+ size 7271
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b832995a5266615b0e3e4086f70522aa34d00f91dc165597fd1d810432779ba
3
  size 4667
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a02b74a29f3bd2fb75ea22c3890bcb31ee009c72ec0e942299216385158a9151
3
  size 4667
vocab.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "aka": {
3
- "[PAD]": 29,
4
- "[UNK]": 28,
5
  "a": 1,
6
  "b": 2,
7
  "c": 3,
@@ -23,12 +23,18 @@
23
  "u": 19,
24
  "w": 20,
25
  "y": 21,
 
26
  "|": 0,
27
- "á": 22,
28
- "ɔ": 23,
29
- "ɛ": 24,
30
- "": 25,
31
- "": 26,
32
- "": 27
 
 
 
 
 
33
  }
34
  }
 
1
  {
2
  "aka": {
3
+ "[PAD]": 35,
4
+ "[UNK]": 34,
5
  "a": 1,
6
  "b": 2,
7
  "c": 3,
 
23
  "u": 19,
24
  "w": 20,
25
  "y": 21,
26
+ "z": 22,
27
  "|": 0,
28
+ "·": 23,
29
+ "á": 24,
30
+ "ã": 25,
31
+ "ĩ": 26,
32
+ "ũ": 27,
33
+ "ɔ": 28,
34
+ "ɛ": 29,
35
+ "ʹ": 30,
36
+ "—": 31,
37
+ "’": 32,
38
+ "•": 33
39
  }
40
  }