Transformers
PyTorch
wav2vec2
pretraining
mms
Inference Endpoints
patrickvonplaten commited on
Commit
59a1530
1 Parent(s): 56394f7

correct vocab

Browse files
Files changed (1) hide show
  1. README.md +218 -0
README.md ADDED
@@ -0,0 +1,218 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - mms
4
+ language:
5
+ - ab
6
+ - af
7
+ - ak
8
+ - am
9
+ - ar
10
+ - as
11
+ - av
12
+ - ay
13
+ - az
14
+ - ba
15
+ - bm
16
+ - be
17
+ - bn
18
+ - bi
19
+ - bo
20
+ - sh
21
+ - br
22
+ - bg
23
+ - ca
24
+ - cs
25
+ - ce
26
+ - cv
27
+ - ku
28
+ - cy
29
+ - da
30
+ - de
31
+ - dv
32
+ - dz
33
+ - el
34
+ - en
35
+ - eo
36
+ - et
37
+ - eu
38
+ - ee
39
+ - fo
40
+ - fa
41
+ - fj
42
+ - fi
43
+ - fr
44
+ - fy
45
+ - ff
46
+ - ga
47
+ - gl
48
+ - gn
49
+ - gu
50
+ - zh
51
+ - ht
52
+ - ha
53
+ - he
54
+ - hi
55
+ - sh
56
+ - hu
57
+ - hy
58
+ - ig
59
+ - ia
60
+ - ms
61
+ - is
62
+ - it
63
+ - jv
64
+ - ja
65
+ - kn
66
+ - ka
67
+ - kk
68
+ - kr
69
+ - km
70
+ - ki
71
+ - rw
72
+ - ky
73
+ - ko
74
+ - kv
75
+ - lo
76
+ - la
77
+ - lv
78
+ - ln
79
+ - lt
80
+ - lb
81
+ - lg
82
+ - mh
83
+ - ml
84
+ - mr
85
+ - ms
86
+ - mk
87
+ - mg
88
+ - mt
89
+ - mn
90
+ - mi
91
+ - my
92
+ - zh
93
+ - nl
94
+ - 'no'
95
+ - 'no'
96
+ - ne
97
+ - ny
98
+ - oc
99
+ - om
100
+ - or
101
+ - os
102
+ - pa
103
+ - pl
104
+ - pt
105
+ - ms
106
+ - ps
107
+ - qu
108
+ - qu
109
+ - qu
110
+ - qu
111
+ - qu
112
+ - qu
113
+ - qu
114
+ - qu
115
+ - qu
116
+ - qu
117
+ - qu
118
+ - qu
119
+ - qu
120
+ - qu
121
+ - qu
122
+ - qu
123
+ - qu
124
+ - qu
125
+ - qu
126
+ - qu
127
+ - qu
128
+ - qu
129
+ - ro
130
+ - rn
131
+ - ru
132
+ - sg
133
+ - sk
134
+ - sl
135
+ - sm
136
+ - sn
137
+ - sd
138
+ - so
139
+ - es
140
+ - sq
141
+ - su
142
+ - sv
143
+ - sw
144
+ - ta
145
+ - tt
146
+ - te
147
+ - tg
148
+ - tl
149
+ - th
150
+ - ti
151
+ - ts
152
+ - tr
153
+ - uk
154
+ - ms
155
+ - vi
156
+ - wo
157
+ - xh
158
+ - ms
159
+ - yo
160
+ - ms
161
+ - zu
162
+ - za
163
+ license: cc-by-sa-4.0
164
+ datasets:
165
+ - google/fleurs
166
+ metrics:
167
+ - wer
168
+ ---
169
+
170
+ # Massively Multilingual Speech (MMS) - 300m
171
+
172
+ Facebook's MMS counting *300m* parameters.
173
+
174
+ MMS is Facebook AI's massive multilingual pretrained model for speech ("MMS").
175
+ It is pretrained in with [Wav2Vec2's self-supervised training objective](https://ai.facebook.com/blog/wav2vec-20-learning-the-structure-of-speech-from-raw-audio/) on about 500,000 hours of speech data in over 1,400 languages.
176
+
177
+ When using the model make sure that your speech input is sampled at 16kHz.
178
+
179
+ **Note**: This model should be fine-tuned on a downstream task, like Automatic Speech Recognition, Translation, or Classification. Check out the [**How-to-fine section](#how-to-finetune) or [**this blog**](https://huggingface.co/blog/fine-tune-xlsr-wav2vec2) for more information about ASR.
180
+
181
+ ## Table Of Content
182
+
183
+ - [How to Finetune](#how-to-finetune)
184
+ - [Model details](#model-details)
185
+ - [Additional links](#additional-links)
186
+
187
+ ## How to finetune
188
+
189
+ Coming soon...
190
+
191
+ ## Model details
192
+
193
+ - **Developed by:** Vineel Pratap et al.
194
+ - **Model type:** Multi-Lingual Automatic Speech Recognition model
195
+ - **Language(s):** 1000+ languages, see [supported languages](#supported-languages)
196
+ - **License:** CC-BY-NC 4.0 license
197
+ - **Num parameters**: 300 million
198
+ - **Cite as:**
199
+
200
+ @article{pratap2023mms,
201
+ title={Scaling Speech Technology to 1,000+ Languages},
202
+ author={Vineel Pratap and Andros Tjandra and Bowen Shi and Paden Tomasello and Arun Babu and Sayani Kundu and Ali Elkahky and Zhaoheng Ni and Apoorv Vyas and Maryam Fazel-Zarandi and Alexei Baevski and Yossi Adi and Xiaohui Zhang and Wei-Ning Hsu and Alexis Conneau and Michael Auli},
203
+ journal={arXiv},
204
+ year={2023}
205
+ }
206
+
207
+ ## Additional Links
208
+
209
+ - [Blog post]( )
210
+ - [Transformers documentation](https://huggingface.co/docs/transformers/main/en/model_doc/mms).
211
+ - [Paper](https://arxiv.org/abs/2305.13516)
212
+ - [GitHub Repository](https://github.com/facebookresearch/fairseq/tree/main/examples/mms#asr)
213
+ - [Other **MMS** checkpoints](https://huggingface.co/models?other=mms)
214
+ - MMS ASR fine-tuned checkpoints:
215
+ - [facebook/mms-1b-all](https://huggingface.co/facebook/mms-1b-all)
216
+ - [facebook/mms-1b-l1107](https://huggingface.co/facebook/mms-1b-l1107)
217
+ - [facebook/mms-1b-fl102](https://huggingface.co/facebook/mms-1b-fl102)
218
+ - [Official Space](https://huggingface.co/spaces/facebook/MMS)