w11wo commited on
Commit
eff41b2
1 Parent(s): 017e952

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +45 -38
README.md CHANGED
@@ -1,66 +1,73 @@
1
  ---
 
2
  license: apache-2.0
3
  tags:
4
- - generated_from_trainer
 
5
  metrics:
6
- - accuracy
7
- - f1
8
  model-index:
9
- - name: wav2vec2-xls-r-adult-child-cls
10
- results: []
11
  ---
12
 
13
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
- should probably proofread and complete it, then remove this comment. -->
15
 
16
- # wav2vec2-xls-r-adult-child-cls
17
 
18
- This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the None dataset.
19
- It achieves the following results on the evaluation set:
20
- - Loss: 0.1851
21
- - Accuracy: 0.9469
22
- - F1: 0.9508
23
 
24
- ## Model description
25
 
26
- More information needed
 
 
27
 
28
- ## Intended uses & limitations
29
 
30
- More information needed
31
 
32
- ## Training and evaluation data
33
-
34
- More information needed
35
 
36
  ## Training procedure
37
 
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
- - learning_rate: 3e-05
42
- - train_batch_size: 8
43
- - eval_batch_size: 8
44
- - seed: 42
45
- - gradient_accumulation_steps: 4
46
- - total_train_batch_size: 32
47
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
48
- - lr_scheduler_type: linear
49
- - lr_scheduler_warmup_ratio: 0.1
50
- - num_epochs: 5
 
51
 
52
  ### Training results
53
 
54
- | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
55
- |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
56
- | 0.2906 | 1.0 | 383 | 0.1856 | 0.9372 | 0.9421 |
57
- | 0.1749 | 2.0 | 766 | 0.1925 | 0.9418 | 0.9465 |
58
- | 0.1681 | 3.0 | 1149 | 0.1893 | 0.9414 | 0.9459 |
59
- | 0.1295 | 4.0 | 1532 | 0.1851 | 0.9469 | 0.9508 |
60
- | 0.2031 | 5.0 | 1915 | 0.1944 | 0.9423 | 0.9460 |
 
 
 
 
 
 
61
 
 
62
 
63
- ### Framework versions
64
 
65
  - Transformers 4.17.0.dev0
66
  - Pytorch 1.10.2+cu102
 
1
  ---
2
+ language: en
3
  license: apache-2.0
4
  tags:
5
+ - audio-classification
6
+ - generated_from_trainer
7
  metrics:
8
+ - accuracy
9
+ - f1
10
  model-index:
11
+ - name: wav2vec2-xls-r-adult-child-cls
12
+ results: []
13
  ---
14
 
15
+ # Wav2Vec2 XLS-R Adult/Child Speech Classifier
 
16
 
17
+ Wav2Vec2 XLS-R Adult/Child Speech Classifier is an audio classification model based on the [XLS-R](https://arxiv.org/abs/2111.09296) architecture. This model is a fine-tuned version of [wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on a private adult/child speech classification dataset.
18
 
19
+ This model was trained using HuggingFace's PyTorch framework. All training was done on a Tesla P100, provided by Kaggle. [Training metrics](https://huggingface.co/w11wo/wav2vec2-xls-r-adult-child-cls/tensorboard) were logged via Tensorboard.
 
 
 
 
20
 
21
+ ## Model
22
 
23
+ | Model | #params | Arch. | Training/Validation data (text) |
24
+ | -------------------------------- | ------- | ----- | ----------------------------------------- |
25
+ | `wav2vec2-xls-r-adult-child-cls` | 300M | XLS-R | Adult/Child Speech Classification Dataset |
26
 
27
+ ## Evaluation Results
28
 
29
+ The model achieves the following results on evaluation:
30
 
31
+ | Dataset | Loss | Accuracy | F1 |
32
+ | --------------------------------- | ------ | -------- | ------ |
33
+ | Adult/Child Speech Classification | 0.1851 | 94.69% | 0.9508 |
34
 
35
  ## Training procedure
36
 
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
+
41
+ - `learning_rate`: 3e-05
42
+ - `train_batch_size`: 8
43
+ - `eval_batch_size`: 8
44
+ - `seed`: 42
45
+ - `gradient_accumulation_steps`: 4
46
+ - `total_train_batch_size`: 32
47
+ - `optimizer`: Adam with `betas=(0.9,0.999)` and `epsilon=1e-08`
48
+ - `lr_scheduler_type`: linear
49
+ - `lr_scheduler_warmup_ratio`: 0.1
50
+ - `num_epochs`: 5
51
 
52
  ### Training results
53
 
54
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
55
+ | :-----------: | :---: | :--: | :-------------: | :------: | :----: |
56
+ | 0.2906 | 1.0 | 383 | 0.1856 | 0.9372 | 0.9421 |
57
+ | 0.1749 | 2.0 | 766 | 0.1925 | 0.9418 | 0.9465 |
58
+ | 0.1681 | 3.0 | 1149 | 0.1893 | 0.9414 | 0.9459 |
59
+ | 0.1295 | 4.0 | 1532 | 0.1851 | 0.9469 | 0.9508 |
60
+ | 0.2031 | 5.0 | 1915 | 0.1944 | 0.9423 | 0.9460 |
61
+
62
+ ## Disclaimer
63
+
64
+ Do consider the biases which came from pre-training datasets that may be carried over into the results of this model.
65
+
66
+ ## Authors
67
 
68
+ Wav2Vec2 XLS-R Adult/Child Speech Classifier was trained and evaluated by [Wilson Wongso](https://w11wo.github.io/). All computation and development are done on Kaggle.
69
 
70
+ ## Framework versions
71
 
72
  - Transformers 4.17.0.dev0
73
  - Pytorch 1.10.2+cu102