ahmmedasaad2772 commited on
Commit
48eaec6
1 Parent(s): ad47858

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +82 -0
README.md ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - generated_from_trainer
5
+ datasets:
6
+ - audiofolder
7
+ metrics:
8
+ - accuracy
9
+ model-index:
10
+ - name: wav2vec2-base-arabic_speech_emotion_recognition
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # wav2vec2-base-arabic_speech_emotion_recognition
18
+
19
+ This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the audiofolder dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.0430
22
+ - Accuracy: 0.9992
23
+ - Weighted f1: 0.9992
24
+ - Micro f1: 0.9992
25
+ - Macro f1: 0.9993
26
+ - Weighted recall: 0.9992
27
+ - Micro recall: 0.9992
28
+ - Macro recall: 0.9992
29
+ - Weighted precision: 0.9992
30
+ - Micro precision: 0.9992
31
+ - Macro precision: 0.9993
32
+
33
+ ## Model description
34
+
35
+ More information needed
36
+
37
+ ## Intended uses & limitations
38
+
39
+ More information needed
40
+
41
+ ## Training and evaluation data
42
+
43
+ More information needed
44
+
45
+ ## Training procedure
46
+
47
+ ### Training hyperparameters
48
+
49
+ The following hyperparameters were used during training:
50
+ - learning_rate: 3e-05
51
+ - train_batch_size: 32
52
+ - eval_batch_size: 32
53
+ - seed: 42
54
+ - gradient_accumulation_steps: 4
55
+ - total_train_batch_size: 128
56
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
57
+ - lr_scheduler_type: linear
58
+ - lr_scheduler_warmup_ratio: 0.1
59
+ - num_epochs: 10
60
+
61
+ ### Training results
62
+
63
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | Weighted f1 | Micro f1 | Macro f1 | Weighted recall | Micro recall | Macro recall | Weighted precision | Micro precision | Macro precision |
64
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:-----------:|:--------:|:--------:|:---------------:|:------------:|:------------:|:------------------:|:---------------:|:---------------:|
65
+ | 1.6078 | 0.98 | 29 | 1.4403 | 0.5227 | 0.4478 | 0.5227 | 0.4426 | 0.5227 | 0.5227 | 0.5187 | 0.4200 | 0.5227 | 0.4145 |
66
+ | 1.6078 | 1.98 | 58 | 0.7310 | 0.9808 | 0.9808 | 0.9808 | 0.9809 | 0.9808 | 0.9808 | 0.9799 | 0.9816 | 0.9808 | 0.9827 |
67
+ | 1.6078 | 2.98 | 87 | 0.2797 | 0.9992 | 0.9992 | 0.9992 | 0.9993 | 0.9992 | 0.9992 | 0.9992 | 0.9992 | 0.9992 | 0.9993 |
68
+ | 1.6078 | 3.98 | 116 | 0.1361 | 0.9992 | 0.9992 | 0.9992 | 0.9993 | 0.9992 | 0.9992 | 0.9992 | 0.9992 | 0.9992 | 0.9993 |
69
+ | 1.6078 | 4.98 | 145 | 0.0855 | 0.9992 | 0.9992 | 0.9992 | 0.9993 | 0.9992 | 0.9992 | 0.9992 | 0.9992 | 0.9992 | 0.9993 |
70
+ | 1.6078 | 5.98 | 174 | 0.0638 | 0.9992 | 0.9992 | 0.9992 | 0.9993 | 0.9992 | 0.9992 | 0.9992 | 0.9992 | 0.9992 | 0.9993 |
71
+ | 1.6078 | 6.98 | 203 | 0.0529 | 0.9992 | 0.9992 | 0.9992 | 0.9993 | 0.9992 | 0.9992 | 0.9992 | 0.9992 | 0.9992 | 0.9993 |
72
+ | 1.6078 | 7.98 | 232 | 0.0470 | 0.9992 | 0.9992 | 0.9992 | 0.9993 | 0.9992 | 0.9992 | 0.9992 | 0.9992 | 0.9992 | 0.9993 |
73
+ | 1.6078 | 8.98 | 261 | 0.0440 | 0.9992 | 0.9992 | 0.9992 | 0.9993 | 0.9992 | 0.9992 | 0.9992 | 0.9992 | 0.9992 | 0.9993 |
74
+ | 1.6078 | 9.98 | 290 | 0.0430 | 0.9992 | 0.9992 | 0.9992 | 0.9993 | 0.9992 | 0.9992 | 0.9992 | 0.9992 | 0.9992 | 0.9993 |
75
+
76
+
77
+ ### Framework versions
78
+
79
+ - Transformers 4.26.1
80
+ - Pytorch 2.1.0+cu121
81
+ - Datasets 2.18.0
82
+ - Tokenizers 0.13.3