vmkhlv commited on
Commit
be5b020
1 Parent(s): a51dfa2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +42 -7
README.md CHANGED
@@ -34,6 +34,13 @@ configs:
34
  path: data/validation-*
35
  - split: test
36
  path: data/test-*
 
 
 
 
 
 
 
37
  ---
38
 
39
  # Dataset Card for Dataset Name
@@ -74,7 +81,34 @@ The dataset is intended to be used for NLP model development and benchmarking.
74
 
75
  <!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
76
 
77
- [More Information Needed]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
78
 
79
  ## Dataset Creation
80
 
@@ -135,15 +169,15 @@ The annotators were provided with a set of initial instructions, largely based o
135
  dataset (Rajpurkar et al., 2016) and the GermanQuAD data (Moller et al., 2021). These instructions were subsequently refined following regular
136
  meetings with the annotation team.
137
  The annotation guidelines provided to the annotators are available (here)[https://github.com/ltgoslo/NorQuAD/blob/main/guidelines.md].
138
- For annotation, we used the Haystack annotation tool which was designed for QA collection.
139
 
140
 
141
  #### Human validation
142
 
143
- In a separate stage, the annotators validated a subset of the NorQuAD dataset. In this phase each
144
  annotator replied to the questions created by the
145
  other annotator. We chose the question-answer
146
- pairs for validation at random. In total, 1378 questions from the set of question-answer pairs, were
147
  answered by validators.
148
 
149
 
@@ -154,7 +188,7 @@ answered by validators.
154
 
155
  Two students of the Master’s program in Natural Language Processing at the University of Oslo,
156
  both native Norwegian speakers, created question-answer pairs from the collected passages. Each
157
- student received separate set of passages for annotation. The students received financial remuneration for their efforts and are co-authors of the
158
  paper describing the resource.
159
 
160
 
@@ -164,7 +198,7 @@ paper describing the resource.
164
  <!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
165
 
166
  **BibTeX:**
167
-
168
  @inproceedings{
169
  ivanova2023norquad,
170
  title={NorQu{AD}: Norwegian Question Answering Dataset},
@@ -173,6 +207,7 @@ booktitle={The 24th Nordic Conference on Computational Linguistics},
173
  year={2023},
174
  url={https://aclanthology.org/2023.nodalida-1.17.pdf}
175
  }
 
176
  **APA:**
177
 
178
  [More Information Needed]
@@ -184,4 +219,4 @@ Vladislav Mikhailov and Lilja Øvrelid
184
 
185
  ## Dataset Card Contact
186
 
187
- vladism@ifi.uio.no and liljao@ifi.uio.no
 
34
  path: data/validation-*
35
  - split: test
36
  path: data/test-*
37
+ license: cc0-1.0
38
+ task_categories:
39
+ - question-answering
40
+ language:
41
+ - nb
42
+ size_categories:
43
+ - 1K<n<10K
44
  ---
45
 
46
  # Dataset Card for Dataset Name
 
81
 
82
  <!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
83
 
84
+ **Data Instances**
85
+
86
+ ```
87
+ {
88
+ "id": "1",
89
+ "context": "This is a test context",
90
+ "question": "This is a question",
91
+ "answers": {
92
+ "answer_start": [1],
93
+ "text": ["This is an answer"]
94
+ },
95
+ }
96
+ ```
97
+
98
+ **Data Fields**
99
+
100
+ ```
101
+ id: a string feature.
102
+ context: a string feature.
103
+ question: a string feature.
104
+ answers: a dictionary feature containing:
105
+ text: a string feature.
106
+ answer_start: a int32 feature.
107
+ ```
108
+
109
+ **Dataset Splits**
110
+
111
+ NorQuAD consists of training (3808 examples), validation (472), and public test (472) sets.
112
 
113
  ## Dataset Creation
114
 
 
169
  dataset (Rajpurkar et al., 2016) and the GermanQuAD data (Moller et al., 2021). These instructions were subsequently refined following regular
170
  meetings with the annotation team.
171
  The annotation guidelines provided to the annotators are available (here)[https://github.com/ltgoslo/NorQuAD/blob/main/guidelines.md].
172
+ For annotation, we used the Haystack annotation tool, which was designed for QA collection.
173
 
174
 
175
  #### Human validation
176
 
177
+ In a separate stage, the annotators validated a subset of the NorQuAD dataset. In this phase, each
178
  annotator replied to the questions created by the
179
  other annotator. We chose the question-answer
180
+ pairs for validation at random. In total, 1378 questions from the set of question-answer pairs were
181
  answered by validators.
182
 
183
 
 
188
 
189
  Two students of the Master’s program in Natural Language Processing at the University of Oslo,
190
  both native Norwegian speakers, created question-answer pairs from the collected passages. Each
191
+ student received a separate set of passages for annotation. The students received financial remuneration for their efforts and are co-authors of the
192
  paper describing the resource.
193
 
194
 
 
198
  <!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
199
 
200
  **BibTeX:**
201
+ ```
202
  @inproceedings{
203
  ivanova2023norquad,
204
  title={NorQu{AD}: Norwegian Question Answering Dataset},
 
207
  year={2023},
208
  url={https://aclanthology.org/2023.nodalida-1.17.pdf}
209
  }
210
+ ```
211
  **APA:**
212
 
213
  [More Information Needed]
 
219
 
220
  ## Dataset Card Contact
221
 
222
+ vladism@ifi.uio.no and liljao@ifi.uio.no