nazneen commited on
Commit
5976312
1 Parent(s): 43be426

model documentation

Browse files
Files changed (1) hide show
  1. README.md +164 -15
README.md CHANGED
@@ -1,21 +1,170 @@
1
- ---
2
- language:
3
- - ru
4
- tags:
5
- - PyTorch
6
- - Transformers
7
- - bert
8
- - exbert
9
- pipeline_tag: fill-mask
10
- thumbnail: "https://github.com/sberbank-ai/model-zoo"
11
- license: apache-2.0
12
- ---
13
 
14
- # ruBert-large
15
- Model was trained by [SberDevices](https://sberdevices.ru/) team.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  * Task: `mask filling`
17
  * Type: `encoder`
18
  * Tokenizer: `bpe`
19
  * Dict size: `120 138`
20
  * Num Parameters: `178 M`
21
- * Training Data Volume `30 GB`
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
 
2
+ # Model Card for ruBert-large
3
+
4
+
5
+ # Model Details
6
+
7
+ ## Model Description
8
+
9
+
10
+ - **Developed by:** Sberbank-ai
11
+ - **Shared by [Optional]:** Hugging Face
12
+ - **Model type:** Fill-Mask
13
+ - **Language(s) (NLP):** ru
14
+ - **License:** apache-2.0
15
+ - **Related Models:** exbert
16
+ - **Parent Model:** bert
17
+ - **Resources for more information:**
18
+ - [GitHub Repo](https://github.com/sberbank-ai/model-zoo)
19
+ - [Associated Paper](https://arxiv.org/abs/1810.04805)
20
+
21
+ # Uses
22
+
23
+ ## Direct Use
24
+
25
+ Fill-Mask
26
+
27
+ ## Downstream Use [Optional]
28
+
29
+
30
+ More information needed.
31
+
32
+
33
+ ## Out-of-Scope Use
34
+
35
+ The model should not be used to intentionally create hostile or alienating environments for people.
36
+
37
+
38
+ # Bias, Risks, and Limitations
39
+
40
+
41
+ Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
42
+
43
+
44
+ ## Recommendations
45
+
46
+
47
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recomendations.
48
+
49
+
50
+ # Training Details
51
+
52
+ ## Training Data
53
+
54
+ More information needed
55
+
56
+ ## Training Procedure
57
+
58
+
59
+
60
+ ### Preprocessing
61
+
62
+ More information needed
63
+
64
+ ### Speeds, Sizes, Times
65
+
66
  * Task: `mask filling`
67
  * Type: `encoder`
68
  * Tokenizer: `bpe`
69
  * Dict size: `120 138`
70
  * Num Parameters: `178 M`
71
+ * Training Data Volume `30 GB`
72
+
73
+
74
+ # Evaluation
75
+
76
+
77
+
78
+ ## Testing Data, Factors & Metrics
79
+
80
+ ### Testing Data
81
+
82
+
83
+
84
+ More information needed
85
+
86
+ ### Factors
87
+
88
+ More information needed
89
+
90
+ ### Metrics
91
+
92
+ More information needed
93
+
94
+ ## Results
95
+
96
+ | Model | Task | Type | Tokenizer | Dict size | Num Parameters | Training Data Volume |
97
+ |-----------------|----------------------|-----------------|-----------|-----------|-----------------|----------------------|
98
+ | ruBERT-large | mask filling | encoder | bpe | 120 138 | 427 M | 30 GB |
99
+
100
+ # Model Examination
101
+
102
+ More information needed
103
+ # Environmental Impact
104
+
105
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
106
+
107
+ - **Hardware Type:** More information needed
108
+ - **Hours used:** More information needed
109
+ - **Cloud Provider:** More information needed
110
+ - **Compute Region:** More information needed
111
+ - **Carbon Emitted:** More information needed
112
+
113
+ # Technical Specifications [optional]
114
+
115
+ ## Model Architecture and Objective
116
+
117
+ More information needed
118
+
119
+ ## Compute Infrastructure
120
+ More information needed
121
+
122
+ ### Hardware
123
+
124
+ More information needed
125
+
126
+ ### Software
127
+
128
+ More information needed
129
+
130
+
131
+ # Citation
132
+
133
+
134
+ **BibTeX:**
135
+
136
+ More information needed
137
+
138
+ **APA:**
139
+
140
+ More information needed
141
+
142
+ # Glossary [optional]
143
+
144
+ More information needed
145
+
146
+ # More Information [optional]
147
+
148
+ More information needed
149
+ # Model Card Authors [optional]
150
+
151
+ Sberbank-ai in collaberation with Ezi Ozoani and the Hugging Face team
152
+
153
+ # Model Card Contact
154
+ More information needed
155
+
156
+ # How to Get Started with the Model
157
+
158
+ Use the code below to get started with the model.
159
+
160
+ <details>
161
+ <summary> Click to expand </summary>
162
+ ```python
163
+ from transformers import AutoTokenizer, AutoModelForTokenClassification
164
+
165
+ tokenizer = AutoTokenizer.from_pretrained("asahi417/tner-xlm-roberta-base-ontonotes5")
166
+
167
+ model = AutoModelForTokenClassification.from_pretrained("asahi417/tner-xlm-roberta-base-ontonotes5")
168
+
169
+ ```
170
+ </details>