nazneen commited on
Commit
d615855
1 Parent(s): 54ee55d

model documentation

Browse files
Files changed (1) hide show
  1. README.md +170 -0
README.md CHANGED
@@ -1,3 +1,173 @@
 
1
  ---
2
  language: en
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
  ---
3
  language: en
4
  ---
5
+
6
+ # Model Card for ivila-row-layoutlm-finetuned-s2vl-v2
7
+
8
+
9
+
10
+ # Model Details
11
+
12
+ ## Model Description
13
+
14
+
15
+ - **Developed by:** Allen Institute for AI [allenai]
16
+ - **Shared by [Optional]:** More information needed
17
+ - **Model type:** Token Classification
18
+ - **Language(s) (NLP):** en
19
+ - **License:** More information needed
20
+ - **Related Models:**
21
+ - **Parent Model:** LayoutLM
22
+ - **Resources for more information:**
23
+ - [GitHub Repo](https://aka.ms/layoutlm)
24
+
25
+
26
+ # Uses
27
+
28
+
29
+ ## Direct Use
30
+
31
+ This model can be used for the task of Token Classification
32
+
33
+ ## Downstream Use [Optional]
34
+
35
+ More information needed
36
+
37
+ ## Out-of-Scope Use
38
+
39
+ The model should not be used to intentionally create hostile or alienating environments for people.
40
+
41
+ # Bias, Risks, and Limitations
42
+
43
+ Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
44
+
45
+
46
+ ## Recommendations
47
+
48
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
49
+
50
+
51
+ # Training Details
52
+
53
+ ## Training Data
54
+
55
+ See the [LayoutLM model card](https://huggingface.co/microsoft/layoutlm-base-uncased) for more information
56
+
57
+ > LayoutLM was pre-trained on IIT-CDIP Test Collection 1.0* dataset with two settings.
58
+ LayoutLM-Base, Uncased (11M documents, 2 epochs): 12-layer, 768-hidden, 12-heads, 113M parameters (This Model)
59
+ LayoutLM-Large, Uncased (11M documents, 2 epochs): 24-layer, 1024-hidden, 16-heads, 343M parameters
60
+
61
+
62
+ ## Training Procedure
63
+
64
+
65
+ ### Preprocessing
66
+
67
+ More information needed
68
+
69
+ ### Speeds, Sizes, Times
70
+
71
+ More information needed
72
+
73
+ # Evaluation
74
+
75
+
76
+ ## Testing Data, Factors & Metrics
77
+
78
+ ### Testing Data
79
+
80
+ More information needed
81
+
82
+ ### Factors
83
+
84
+
85
+ ### Metrics
86
+
87
+ More information needed
88
+ ## Results
89
+
90
+ More information needed
91
+
92
+ # Model Examination
93
+
94
+ More information needed
95
+
96
+ # Environmental Impact
97
+
98
+
99
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
100
+
101
+ - **Hardware Type:** More information needed
102
+ - **Hours used:** More information needed
103
+ - **Cloud Provider:** More information needed
104
+ - **Compute Region:** More information needed
105
+ - **Carbon Emitted:** More information needed
106
+
107
+ # Technical Specifications [optional]
108
+
109
+ ## Model Architecture and Objective
110
+
111
+ More information needed
112
+
113
+ ## Compute Infrastructure
114
+
115
+ More information needed
116
+
117
+ ### Hardware
118
+
119
+ More information needed
120
+
121
+ ### Software
122
+ * Transformers_version: 4.6.0
123
+
124
+ # Citation
125
+
126
+
127
+ **BibTeX:**
128
+ ```
129
+ @misc{xu2019layoutlm,
130
+ title={LayoutLM: Pre-training of Text and Layout for Document Image Understanding},
131
+ author={Yiheng Xu and Minghao Li and Lei Cui and Shaohan Huang and Furu Wei and Ming Zhou},
132
+ year={2019},
133
+ eprint={1912.13318},
134
+ archivePrefix={arXiv},
135
+ primaryClass={cs.CL}
136
+ }
137
+
138
+ ```
139
+
140
+
141
+ # Glossary [optional]
142
+ More information needed
143
+
144
+ # More Information [optional]
145
+
146
+ More information needed
147
+
148
+ # Model Card Authors [optional]
149
+
150
+
151
+ Allen Institute for AI [allenai] in collaboration with Ezi Ozoani and the Hugging Face team
152
+
153
+ # Model Card Contact
154
+
155
+ More information needed
156
+
157
+ # How to Get Started with the Model
158
+
159
+ Use the code below to get started with the model.
160
+
161
+ <details>
162
+ <summary> Click to expand </summary>
163
+
164
+ ```python
165
+ from transformers import AutoTokenizer, AutoModelForTokenClassification
166
+
167
+ tokenizer = AutoTokenizer.from_pretrained("allenai/ivila-row-layoutlm-finetuned-s2vl-v2")
168
+
169
+ model = AutoModelForTokenClassification.from_pretrained("allenai/ivila-row-layoutlm-finetuned-s2vl-v2")
170
+
171
+ ```
172
+ </details>
173
+