File size: 5,226 Bytes
2f216d8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
---
annotations_creators:
- no-annotation
language_creators:
- other
languages:
- ca
- en
- de
- es
- fr
- ru
- ja
- it
- zh
- pt
- nl
- tr
- pl
- vi
- ar
- id
- uk
- ro
- 'no'
- th
- sv
- el
- fi
- he
- da
- cs
- ko
- fa
- hi
- hu
- sk
- lt
- et
- hr
- is
- lv
- ms
- bg
- sr
- ca 
licenses:
- cc0-1.0
multilinguality:
- multilingual
pretty_name: MQA - a Multilingual FAQ and CQA Dataset
size_categories:
- unknown
source_datasets:
- original
task_categories:
- question-answering
task_ids:
- multiple-choice-qa
---
# MQA
MQA is a multilingual corpus of questions and answers parsed from the [Common Crawl](https://commoncrawl.org/). Questions are divided between *Frequently Asked Questions* pages and *Community Question Answering* pages.
```
from datasets import load_dataset
load_dataset("clips/mqa", language="en")
{
  "name": "the title of the question (if any)",
  "text": "the body of the question (if any)",
  "answers": [{
    "text": "the text of the answer",
    "is_accepted": "true|false"
  }]
}
```

## Languages
We collected around 234M pairs of questions and answers in 39 different languages. To download a language specific subset you need to specify the language key as configuration. See below for an example.
```
load_dataset("clips/mqa", language="en") # replace "en" by any language listed below
```

| Language   | Key | Pairs     | Pages     |
|------------|-----|-----------|-----------|
| All        | all | 6,346,693 | 1,035,649 |
| English    | en  | 3,719,484 | 608,796   |
| German     | de  | 829,098   | 111,618   |
| Spanish    | es  | 482,818   | 75,489    |
| French     | fr  | 351,458   | 56,317    |
| Italian    | it  | 155,296   | 24,562    |
| Dutch      | nl  | 150,819   | 32,574    |
| Portuguese | pt  | 138,778   | 26,169    |
| Turkish    | tr  | 102,373   | 19,002    |
| Russian    | ru  | 91,771    | 22,643    |
| Polish     | pl  | 65,182    | 10,695    |
| Indonesian | id  | 45,839    | 7,910     |
| Norwegian  | no  | 37,711    | 5,143     |
| Swedish    | sv  | 37,003    | 5,270     |
| Danish     | da  | 32,655    | 5,279     |
| Vietnamese | vi  | 27,157    | 5,261     |
| Finnish    | fi  | 20,485    | 2,795     |
| Romanian   | ro  | 17,066    | 3,554     |
| Czech      | cs  | 16,675    | 2,568     |
| Hebrew     | he  | 11,212    | 1,921     |
| Hungarian  | hu  | 8,598     | 1,264     |
| Croatian   | hr  | 5,215     | 819       |

## Data Fields
#### Nested (per page - default)
The data is organized by page. Each page contains a list of questions and answers.
- **id** 
- **language**
- **num_pairs**: the number of FAQs on the page
- **domain**: source web domain of the FAQs
- **qa_pairs**: a list of questions and answers
  - **question**
  - **answer**
  - **language**
  
#### Flattened
The data is organized by pair (i.e. pages are flattened). You can access the flat version of any language by appending `_flat` to the configuration (e.g. `en_flat`). The data will be returned pair-by-pair instead of page-by-page. 
- **domain_id** 
- **pair_id**
- **language**
- **domain**: source web domain of the FAQs
- **question**
- **answer**

## Source Data

This section was adapted from the source data description of [OSCAR](https://huggingface.co/datasets/oscar#source-data)

Common Crawl is a non-profit foundation which produces and maintains an open repository of web crawled data that is both accessible and analysable. Common Crawl's complete web archive consists of petabytes of data collected over 8 years of web crawling. The repository contains raw web page HTML data (WARC files), metdata extracts (WAT files) and plain text extracts (WET files). The organisation's crawlers has always respected nofollow and robots.txt policies.

To construct MFAQ, the WARC files of Common Crawl were used. We looked for `FAQPage` markup in the HTML and subsequently parsed the `FAQItem` from the page. 

## People
This model was developed by [Maxime De Bruyn](https://www.linkedin.com/in/maximedebruyn/), Ehsan Lotfi, Jeska Buhmann and Walter Daelemans.

## Licensing Information
```
These data are released under this licensing scheme.
We do not own any of the text from which these data has been extracted.
We license the actual packaging of these data under the Creative Commons CC0 license ("no rights reserved") http://creativecommons.org/publicdomain/zero/1.0/

Should you consider that our data contains material that is owned by you and should therefore not be reproduced here, please:
* Clearly identify yourself, with detailed contact data such as an address, telephone number or email address at which you can be contacted.
* Clearly identify the copyrighted work claimed to be infringed.
* Clearly identify the material that is claimed to be infringing and information reasonably sufficient to allow us to locate the material.

We will comply to legitimate requests by removing the affected sources from the next release of the corpus.
```

## Citation information
```
@misc{debruyn2021mfaq,
      title={MFAQ: a Multilingual FAQ Dataset}, 
      author={Maxime {De Bruyn} and Ehsan Lotfi and Jeska Buhmann and Walter Daelemans},
      year={2021},
      eprint={2109.12870},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}
```