model documentation

#1
by nazneen - opened
Files changed (1) hide show
  1. README.md +162 -0
README.md ADDED
@@ -0,0 +1,162 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Model Card for tiny-wav2vec2-no-tokenizer
2
+
3
+
4
+ # Model Details
5
+
6
+ ## Model Description
7
+
8
+
9
+ - **Developed by:** More information needed
10
+ - **Shared by [Optional]:** Patrick von Platen
11
+ - **Model type:** Automatic Speech Recognition
12
+ - **Language(s) (NLP):** en
13
+ - **License:** More information needed
14
+ - **Related Models:**
15
+ - **Parent Model:** Wav2Vec2
16
+ - **Resources for more information:**
17
+ - [GitHub Repo](https://github.com/facebookresearch/fairseq/tree/main/examples/wav2vec#wav2vec-20)
18
+ - [Associated Paper](https://arxiv.org/abs/2006.11477)
19
+ - [Associated Model Doc](https://huggingface.co/docs/transformers/main/en/model_doc/wav2vec2)
20
+
21
+ # Uses
22
+
23
+
24
+ ## Direct Use
25
+
26
+ This model can be used for the task of Automatic Speech Recognition
27
+
28
+
29
+ ## Downstream Use [Optional]
30
+
31
+ More information needed
32
+
33
+ ## Out-of-Scope Use
34
+
35
+ The model should not be used to intentionally create hostile or alienating environments for people.
36
+
37
+ # Bias, Risks, and Limitations
38
+
39
+ Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
40
+
41
+
42
+ ## Recommendations
43
+
44
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
45
+
46
+
47
+ # Training Details
48
+
49
+ ## Training Data
50
+
51
+ More information needed
52
+
53
+ ## Training Procedure
54
+
55
+
56
+ ### Preprocessing
57
+
58
+ More information needed
59
+
60
+ ### Speeds, Sizes, Times
61
+
62
+ More information needed
63
+
64
+ # Evaluation
65
+
66
+
67
+ ## Testing Data, Factors & Metrics
68
+
69
+ ### Testing Data
70
+
71
+ More information needed
72
+
73
+ ### Factors
74
+
75
+
76
+ ### Metrics
77
+
78
+ More information needed
79
+ ## Results
80
+
81
+ More information needed
82
+
83
+ # Model Examination
84
+
85
+ More information needed
86
+
87
+ # Environmental Impact
88
+
89
+
90
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
91
+
92
+ - **Hardware Type:** More information needed
93
+ - **Hours used:** More information needed
94
+ - **Cloud Provider:** More information needed
95
+ - **Compute Region:** More information needed
96
+ - **Carbon Emitted:** More information needed
97
+
98
+ # Technical Specifications [optional]
99
+
100
+ ## Model Architecture and Objective
101
+
102
+ More information needed
103
+
104
+ ## Compute Infrastructure
105
+
106
+ More information needed
107
+
108
+ ### Hardware
109
+
110
+ More information needed
111
+
112
+ ### Software
113
+ More information needed
114
+
115
+ # Citation
116
+
117
+
118
+ **BibTeX:**
119
+ ```
120
+ @misc{https://doi.org/10.48550/arxiv.2006.11477,
121
+ doi = {10.48550/ARXIV.2006.11477},
122
+
123
+ url = {https://arxiv.org/abs/2006.11477},
124
+
125
+ author = {Baevski, Alexei and Zhou, Henry and Mohamed, Abdelrahman and Auli, Michael},
126
+
127
+ keywords = {Computation and Language (cs.CL), Machine Learning (cs.LG), Sound (cs.SD), Audio and Speech Processing (eess.AS), FOS: Computer and information sciences, FOS: Computer and information sciences, FOS: Electrical engineering, electronic engineering, information engineering, FOS: Electrical engineering, electronic engineering, information engineering},
128
+
129
+ title = {wav2vec 2.0: A Framework for Self-Supervised Learning of Speech Representations},
130
+
131
+ publisher = {arXiv},
132
+ ```
133
+
134
+
135
+ # Glossary [optional]
136
+ More information needed
137
+
138
+ # More Information [optional]
139
+
140
+ More information needed
141
+
142
+ # Model Card Authors [optional]
143
+
144
+
145
+ Patrick von Platen in collaboration with the Hugging Face team
146
+
147
+ # Model Card Contact
148
+
149
+ More information needed
150
+
151
+ # How to Get Started with the Model
152
+
153
+ Use the code below to get started with the model.
154
+
155
+ <details>
156
+ <summary> Click to expand </summary>
157
+ ```python
158
+ from transformers import AutoModel
159
+
160
+ model = AutoModel.from_pretrained("patrickvonplaten/tiny-wav2vec2-no-tokenizer")
161
+ ```
162
+ </details>