Rolv-Arild commited on
Commit
43c0931
1 Parent(s): cef1531

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +130 -0
README.md ADDED
@@ -0,0 +1,130 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - generated_from_trainer
5
+ model-index:
6
+ - name: ''
7
+ results: []
8
+ ---
9
+
10
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
+ should probably proofread and complete it, then remove this comment. -->
12
+
13
+ #
14
+
15
+ This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the None dataset.
16
+ It achieves the following results on the evaluation set:
17
+ - Loss: 0.1957
18
+ - Wer: 0.1699
19
+
20
+ ## Model description
21
+
22
+ More information needed
23
+
24
+ ## Intended uses & limitations
25
+
26
+ More information needed
27
+
28
+ ## Training and evaluation data
29
+
30
+ More information needed
31
+
32
+ ## Training procedure
33
+
34
+ ### Training hyperparameters
35
+
36
+ The following hyperparameters were used during training:
37
+ - learning_rate: 7.5e-05
38
+ - train_batch_size: 16
39
+ - eval_batch_size: 16
40
+ - seed: 42
41
+ - gradient_accumulation_steps: 4
42
+ - total_train_batch_size: 64
43
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
+ - lr_scheduler_type: linear
45
+ - lr_scheduler_warmup_steps: 2000
46
+ - num_epochs: 20.0
47
+ - mixed_precision_training: Native AMP
48
+
49
+ ### Training results
50
+
51
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
52
+ |:-------------:|:-----:|:-----:|:---------------:|:------:|
53
+ | 4.4527 | 0.28 | 250 | 4.0144 | 1.0 |
54
+ | 3.1828 | 0.56 | 500 | 3.1369 | 1.0 |
55
+ | 2.9927 | 0.85 | 750 | 3.0183 | 1.0 |
56
+ | 2.9591 | 1.13 | 1000 | 2.9991 | 1.0 |
57
+ | 2.8989 | 1.41 | 1250 | 2.9000 | 1.0000 |
58
+ | 2.4286 | 1.69 | 1500 | 1.7688 | 0.9550 |
59
+ | 1.6765 | 1.98 | 1750 | 0.6842 | 0.4855 |
60
+ | 1.4521 | 2.26 | 2000 | 0.5096 | 0.3736 |
61
+ | 1.3589 | 2.54 | 2250 | 0.4479 | 0.3335 |
62
+ | 1.3136 | 2.82 | 2500 | 0.4056 | 0.3123 |
63
+ | 1.2856 | 3.11 | 2750 | 0.3870 | 0.2987 |
64
+ | 1.2283 | 3.39 | 3000 | 0.3646 | 0.2828 |
65
+ | 1.2053 | 3.67 | 3250 | 0.3499 | 0.2748 |
66
+ | 1.2087 | 3.95 | 3500 | 0.3345 | 0.2603 |
67
+ | 1.2002 | 4.24 | 3750 | 0.3320 | 0.2523 |
68
+ | 1.1383 | 4.52 | 4000 | 0.3117 | 0.2439 |
69
+ | 1.1364 | 4.8 | 4250 | 0.3198 | 0.2383 |
70
+ | 1.158 | 5.08 | 4500 | 0.3071 | 0.2342 |
71
+ | 1.108 | 5.37 | 4750 | 0.3011 | 0.2314 |
72
+ | 1.1025 | 5.65 | 5000 | 0.2875 | 0.2289 |
73
+ | 1.0697 | 5.93 | 5250 | 0.2926 | 0.2256 |
74
+ | 1.0904 | 6.21 | 5500 | 0.2695 | 0.2245 |
75
+ | 1.0802 | 6.5 | 5750 | 0.2602 | 0.2189 |
76
+ | 1.0882 | 6.78 | 6000 | 0.2603 | 0.2168 |
77
+ | 1.0881 | 7.06 | 6250 | 0.2540 | 0.2293 |
78
+ | 1.0378 | 7.34 | 6500 | 0.2614 | 0.2193 |
79
+ | 1.0397 | 7.63 | 6750 | 0.2707 | 0.2104 |
80
+ | 1.0296 | 7.91 | 7000 | 0.2483 | 0.2119 |
81
+ | 1.0249 | 8.19 | 7250 | 0.2483 | 0.2047 |
82
+ | 1.013 | 8.47 | 7500 | 0.2487 | 0.2042 |
83
+ | 1.0064 | 8.76 | 7750 | 0.2456 | 0.2016 |
84
+ | 1.0668 | 9.04 | 8000 | 0.2397 | 0.1995 |
85
+ | 1.0129 | 9.32 | 8250 | 0.2374 | 0.1994 |
86
+ | 1.0164 | 9.6 | 8500 | 0.2206 | 0.1992 |
87
+ | 0.975 | 9.89 | 8750 | 0.2247 | 0.1973 |
88
+ | 0.9849 | 10.17 | 9000 | 0.2325 | 0.1953 |
89
+ | 0.9826 | 10.45 | 9250 | 0.2301 | 0.1934 |
90
+ | 0.9835 | 10.73 | 9500 | 0.2192 | 0.1942 |
91
+ | 0.9676 | 11.02 | 9750 | 0.2266 | 0.1913 |
92
+ | 0.9627 | 11.3 | 10000 | 0.2193 | 0.1921 |
93
+ | 0.976 | 11.58 | 10250 | 0.2309 | 0.1882 |
94
+ | 0.969 | 11.86 | 10500 | 0.2268 | 0.1886 |
95
+ | 0.9611 | 12.15 | 10750 | 0.2322 | 0.1863 |
96
+ | 0.9397 | 12.43 | 11000 | 0.2197 | 0.1844 |
97
+ | 0.9601 | 12.71 | 11250 | 0.2211 | 0.1871 |
98
+ | 0.9718 | 12.99 | 11500 | 0.2079 | 0.1898 |
99
+ | 0.9347 | 13.28 | 11750 | 0.2054 | 0.1843 |
100
+ | 0.9377 | 13.56 | 12000 | 0.2031 | 0.1842 |
101
+ | 0.934 | 13.84 | 12250 | 0.2059 | 0.1806 |
102
+ | 0.9295 | 14.12 | 12500 | 0.2122 | 0.1861 |
103
+ | 0.935 | 14.41 | 12750 | 0.2072 | 0.1787 |
104
+ | 0.9021 | 14.69 | 13000 | 0.2105 | 0.1781 |
105
+ | 0.9193 | 14.97 | 13250 | 0.2035 | 0.1786 |
106
+ | 0.9214 | 15.25 | 13500 | 0.2035 | 0.1766 |
107
+ | 0.9048 | 15.54 | 13750 | 0.1964 | 0.1758 |
108
+ | 0.9006 | 15.82 | 14000 | 0.1984 | 0.1757 |
109
+ | 0.9027 | 16.1 | 14250 | 0.2022 | 0.1743 |
110
+ | 0.9083 | 16.38 | 14500 | 0.1969 | 0.1744 |
111
+ | 0.9761 | 16.67 | 14750 | 0.1963 | 0.1728 |
112
+ | 0.9311 | 16.95 | 15000 | 0.1960 | 0.1737 |
113
+ | 0.886 | 17.23 | 15250 | 0.1929 | 0.1726 |
114
+ | 0.8969 | 17.51 | 15500 | 0.1928 | 0.1734 |
115
+ | 0.9084 | 17.8 | 15750 | 0.1937 | 0.1713 |
116
+ | 0.8795 | 18.08 | 16000 | 0.1978 | 0.1709 |
117
+ | 0.8883 | 18.36 | 16250 | 0.1956 | 0.1703 |
118
+ | 0.8901 | 18.64 | 16500 | 0.1933 | 0.1705 |
119
+ | 0.8922 | 18.93 | 16750 | 0.1962 | 0.1711 |
120
+ | 0.8765 | 19.21 | 17000 | 0.1962 | 0.1711 |
121
+ | 0.8992 | 19.49 | 17250 | 0.1965 | 0.1703 |
122
+ | 0.8778 | 19.77 | 17500 | 0.1957 | 0.1699 |
123
+
124
+
125
+ ### Framework versions
126
+
127
+ - Transformers 4.17.0.dev0
128
+ - Pytorch 1.10.0+cu113
129
+ - Datasets 1.18.1
130
+ - Tokenizers 0.11.0