w11wo commited on
Commit
eb22a1c
1 Parent(s): 917e919

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +43 -36
README.md CHANGED
@@ -1,64 +1,71 @@
1
  ---
 
2
  license: apache-2.0
3
  tags:
4
- - generated_from_trainer
 
5
  metrics:
6
- - accuracy
7
- - f1
8
  model-index:
9
- - name: distil-wav2vec2-adult-child-cls-v3
10
- results: []
11
  ---
12
 
13
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
- should probably proofread and complete it, then remove this comment. -->
15
 
16
- # distil-wav2vec2-adult-child-cls-v3
17
 
18
- This model is a fine-tuned version of [w11wo/wav2vec2-adult-child-cls-v3](https://huggingface.co/w11wo/wav2vec2-adult-child-cls-v3) on the None dataset.
19
- It achieves the following results on the evaluation set:
20
- - Loss: 0.1301
21
- - Accuracy: 0.9603
22
- - F1: 0.9639
23
 
24
- ## Model description
25
 
26
- More information needed
 
 
27
 
28
- ## Intended uses & limitations
29
 
30
- More information needed
31
 
32
- ## Training and evaluation data
33
-
34
- More information needed
35
 
36
  ## Training procedure
37
 
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
- - learning_rate: 3e-05
42
- - train_batch_size: 32
43
- - eval_batch_size: 32
44
- - seed: 42
45
- - gradient_accumulation_steps: 4
46
- - total_train_batch_size: 128
47
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
48
- - lr_scheduler_type: linear
49
- - lr_scheduler_warmup_ratio: 0.1
50
- - num_epochs: 3
 
51
 
52
  ### Training results
53
 
54
- | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
55
- |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
56
- | 0.212 | 1.0 | 96 | 0.1561 | 0.9561 | 0.9596 |
57
- | 0.1523 | 2.0 | 192 | 0.1408 | 0.9575 | 0.9616 |
58
- | 0.0844 | 3.0 | 288 | 0.1301 | 0.9603 | 0.9639 |
 
 
 
 
 
 
59
 
 
60
 
61
- ### Framework versions
62
 
63
  - Transformers 4.16.2
64
  - Pytorch 1.10.2+cu102
 
1
  ---
2
+ language: en
3
  license: apache-2.0
4
  tags:
5
+ - audio-classification
6
+ - generated_from_trainer
7
  metrics:
8
+ - accuracy
9
+ - f1
10
  model-index:
11
+ - name: distil-wav2vec2-adult-child-cls-v3
12
+ results: []
13
  ---
14
 
15
+ # DistilWav2Vec2 Adult/Child Speech Classifier
 
16
 
17
+ DistilWav2Vec2 Adult/Child Speech Classifier is an audio classification model based on the [wav2vec 2.0](https://arxiv.org/abs/2006.11477) architecture. This model is a distilled version of [wav2vec2-adult-child-cls-v3](https://huggingface.co/w11wo/wav2vec2-adult-child-cls-v3) on a private adult/child speech classification dataset.
18
 
19
+ This model was trained using HuggingFace's PyTorch framework. All training was done on a Tesla P100, provided by Kaggle. [Training metrics](https://huggingface.co/w11wo/wav2vec2-adult-child-cls-v3/tensorboard) were logged via Tensorboard.
 
 
 
 
20
 
21
+ ## Model
22
 
23
+ | Model | #params | Arch. | Training/Validation data (text) |
24
+ | ------------------------------------ | ------- | ----------- | ----------------------------------------- |
25
+ | `distil-wav2vec2-adult-child-cls-v3` | 52M | wav2vec 2.0 | Adult/Child Speech Classification Dataset |
26
 
27
+ ## Evaluation Results
28
 
29
+ The model achieves the following results on evaluation:
30
 
31
+ | Dataset | Loss | Accuracy | F1 |
32
+ | --------------------------------- | ------ | -------- | ------ |
33
+ | Adult/Child Speech Classification | 0.1301 | 96.03% | 0.9639 |
34
 
35
  ## Training procedure
36
 
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
+
41
+ - `learning_rate`: 3e-05
42
+ - `train_batch_size`: 32
43
+ - `eval_batch_size`: 32
44
+ - `seed`: 42
45
+ - `gradient_accumulation_steps`: 4
46
+ - `total_train_batch_size`: 128
47
+ - `optimizer`: Adam with `betas=(0.9,0.999)` and `epsilon=1e-08`
48
+ - `lr_scheduler_type`: linear
49
+ - `lr_scheduler_warmup_ratio`: 0.1
50
+ - `num_epochs`: 3
51
 
52
  ### Training results
53
 
54
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
55
+ | :-----------: | :---: | :--: | :-------------: | :------: | :----: |
56
+ | 0.212 | 1.0 | 96 | 0.1561 | 0.9561 | 0.9596 |
57
+ | 0.1523 | 2.0 | 192 | 0.1408 | 0.9575 | 0.9616 |
58
+ | 0.0844 | 3.0 | 288 | 0.1301 | 0.9603 | 0.9639 |
59
+
60
+ ## Disclaimer
61
+
62
+ Do consider the biases which came from pre-training datasets that may be carried over into the results of this model.
63
+
64
+ ## Authors
65
 
66
+ DistilWav2Vec2 Adult/Child Speech Classifier was trained and evaluated by [Wilson Wongso](https://w11wo.github.io/). All computation and development are done on Kaggle.
67
 
68
+ ## Framework versions
69
 
70
  - Transformers 4.16.2
71
  - Pytorch 1.10.2+cu102