EzraWilliam commited on
Commit
733cc9a
·
verified ·
1 Parent(s): 1515ea9

End of training

Browse files
README.md CHANGED
@@ -1,10 +1,10 @@
1
  ---
2
  license: apache-2.0
 
3
  tags:
4
  - generated_from_trainer
5
- metrics:
6
- - wer
7
- base_model: facebook/wav2vec2-xls-r-300m
8
  model-index:
9
  - name: wav2vec2-base-fleurs-CommonVoice-demo-google-colab-Ezra_William_Prod1
10
  results: []
@@ -15,10 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # wav2vec2-base-fleurs-CommonVoice-demo-google-colab-Ezra_William_Prod1
17
 
18
- This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the None dataset.
19
- It achieves the following results on the evaluation set:
20
- - Loss: 6.9691
21
- - Wer: 1.1680
22
 
23
  ## Model description
24
 
@@ -37,7 +34,7 @@ More information needed
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
- - learning_rate: 0.0003
41
  - train_batch_size: 8
42
  - eval_batch_size: 8
43
  - seed: 42
@@ -45,21 +42,17 @@ The following hyperparameters were used during training:
45
  - total_train_batch_size: 16
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
48
- - lr_scheduler_warmup_steps: 300
49
- - num_epochs: 600
50
  - mixed_precision_training: Native AMP
51
 
52
  ### Training results
53
 
54
- | Training Loss | Epoch | Step | Validation Loss | Wer |
55
- |:-------------:|:-----:|:----:|:---------------:|:------:|
56
- | 1.4925 | 300.0 | 300 | 6.2580 | 1.0725 |
57
- | 0.0626 | 600.0 | 600 | 6.9691 | 1.1680 |
58
 
59
 
60
  ### Framework versions
61
 
62
  - Transformers 4.37.1
63
- - Pytorch 2.1.0+cu121
64
  - Datasets 2.16.1
65
  - Tokenizers 0.15.1
 
1
  ---
2
  license: apache-2.0
3
+ base_model: facebook/wav2vec2-xls-r-300m
4
  tags:
5
  - generated_from_trainer
6
+ datasets:
7
+ - xtreme_s
 
8
  model-index:
9
  - name: wav2vec2-base-fleurs-CommonVoice-demo-google-colab-Ezra_William_Prod1
10
  results: []
 
15
 
16
  # wav2vec2-base-fleurs-CommonVoice-demo-google-colab-Ezra_William_Prod1
17
 
18
+ This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the xtreme_s dataset.
 
 
 
19
 
20
  ## Model description
21
 
 
34
  ### Training hyperparameters
35
 
36
  The following hyperparameters were used during training:
37
+ - learning_rate: 0.0005
38
  - train_batch_size: 8
39
  - eval_batch_size: 8
40
  - seed: 42
 
42
  - total_train_batch_size: 16
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
+ - lr_scheduler_warmup_steps: 200
46
+ - num_epochs: 6
47
  - mixed_precision_training: Native AMP
48
 
49
  ### Training results
50
 
 
 
 
 
51
 
52
 
53
  ### Framework versions
54
 
55
  - Transformers 4.37.1
56
+ - Pytorch 2.1.2+cu121
57
  - Datasets 2.16.1
58
  - Tokenizers 0.15.1
config.json CHANGED
@@ -78,7 +78,7 @@
78
  "num_hidden_layers": 24,
79
  "num_negatives": 100,
80
  "output_hidden_size": 1024,
81
- "pad_token_id": 30,
82
  "proj_codevector_dim": 768,
83
  "tdnn_dilation": [
84
  1,
@@ -104,6 +104,6 @@
104
  "torch_dtype": "float32",
105
  "transformers_version": "4.37.1",
106
  "use_weighted_layer_sum": false,
107
- "vocab_size": 33,
108
  "xvector_output_dim": 512
109
  }
 
78
  "num_hidden_layers": 24,
79
  "num_negatives": 100,
80
  "output_hidden_size": 1024,
81
+ "pad_token_id": 62,
82
  "proj_codevector_dim": 768,
83
  "tdnn_dilation": [
84
  1,
 
104
  "torch_dtype": "float32",
105
  "transformers_version": "4.37.1",
106
  "use_weighted_layer_sum": false,
107
+ "vocab_size": 65,
108
  "xvector_output_dim": 512
109
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9d57cd133f4f23063d568fc2e3d893d211e9f164527128a2b933e08461b7b51
3
- size 1261942780
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60fd368398d469bcee760020fc80995d9f95ffb04e121c2e25c4c17d9632ddf8
3
+ size 1262073980
preprocessor_config.json CHANGED
@@ -4,6 +4,6 @@
4
  "feature_size": 1,
5
  "padding_side": "right",
6
  "padding_value": 0.0,
7
- "return_attention_mask": false,
8
  "sampling_rate": 16000
9
  }
 
4
  "feature_size": 1,
5
  "padding_side": "right",
6
  "padding_value": 0.0,
7
+ "return_attention_mask": true,
8
  "sampling_rate": 16000
9
  }
runs/Jan25_09-29-39_78d99ad982a5/events.out.tfevents.1706174981.78d99ad982a5.339.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e65cd25a7ecd7d82af1c70c5c3669ad7db10d11bd1e762a537651cba1294a2bd
3
+ size 6303
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8ebc7b77aa0373ca415c790e73f42164d7c104d92ce907b99fc2d09bebe80958
3
  size 4792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3db2820d104018abf1e4975ba8fee41d6ee0e183e4e501b8ae2d139629675b76
3
  size 4792