pquad / README.md
Gholamreza's picture
Update README.md
2d5d844
|
raw
history blame
No virus
5.15 kB
---
pretty_name: PQuAD
annotations_creators:
- crowdsourced
language_creators:
- crowdsourced
language:
- fa
license:
- cc-by-sa-4.0
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- question-answering
task_ids:
- open-domain-qa
- extractive-qa
paperswithcode_id: squad
train-eval-index:
- config: pquad
task: question-answering
task_id: extractive_question_answering
splits:
train_split: train
eval_split: validation
col_mapping:
question: question
context: context
answers:
text: text
answer_start: answer_start
metrics:
- type: pquad
name: PQuAD
dataset_info:
features:
- name: id
dtype: int32
- name: title
dtype: string
- name: context
dtype: string
- name: question
dtype: string
- name: answers
sequence:
- name: text
dtype: string
- name: answer_start
dtype: int32
config_name: pquad
splits:
- name: train
num_bytes: ...
num_examples: 63994
- name: validation
num_bytes: ...
num_examples: 7976
- name: test
num_bytes: ...
num_examples: 8002
download_size: ...
dataset_size: ...
---
# Dataset Card for "pquad"
## PQuAD Description
**THIS IS A NON-OFFICIAL VERSION OF THE DATASET UPLOADED TO HUGGINGFACE BY [Gholamreza Dar](https://huggingface.co/Gholamreza)**
*The original repository for the dataset is https://github.com/AUT-NLP/PQuAD*
PQuAD is a crowd- sourced reading comprehension dataset on Persian Language. It includes 80,000
questions along with their answers, with 25% of the questions being unanswerable. As a reading
comprehension dataset, it requires a system to read a passage and then answer the given questions
from the passage. PQuAD's questions are based on Persian Wikipedia articles and cover a wide
variety of subjects. Articles used for question generation are quality checked and include few
number of non-Persian words.
## Dataset Splits
The dataset is divided into three categories including train, validation, and test sets and the
statistics of these sets are as follows:
```
+----------------------------+-------+------------+------+-------+
| | Train | Validation | Test | Total |
+----------------------------+-------+------------+------+-------+
| Total Questions | 63994 | 7976 | 8002 | 79972 |
| Unanswerable Questions | 15721 | 1981 | 1914 | 19616 |
| Mean # of paragraph tokens | 125 | 121 | 124 | 125 |
| Mean # of question tokens | 10 | 11 | 11 | 10 |
| Mean # of answer tokens | 5 | 6 | 5 | 5 |
+----------------------------+-------+------------+------+-------+
```
Workers were encouraged to use paraphrased sentences in their questions and avoid choosing the
answers comprising non-Persian words. Another group of crowdworkers validated the questions and
answers in the test and validation set to ensure their quality. They also provided additional
answers to the questions in test and validation sets if possible. This helps to consider all
possible types of answers and have a better evaluation of models.
PQuAD is stored in the JSON format and consists of passages where each passage is linked to a
set of questions. Answer(s) of the questions is specified with answer's span (start and end
point of answer in paragraph). Also, the unanswerable questions are marked as unanswerable.
## Results
The estimated human performance on the test set is 88.3% for F1 and 80.3% for EM. We have
evaluated PQuAD using two pre-trained transformer-based language models, namely ParsBERT
(Farahani et al., 2021) and XLM-RoBERTa (Conneau et al., 2020), as well as BiDAF (Levy et
al., 2017) which is an attention-based model proposed for MRC.
```
+-------------+------+------+-----------+-----------+-------------+
| Model | EM | F1 | HasAns_EM | HasAns_F1 | NoAns_EM/F1 |
+-------------+------+------+-----------+-----------+-------------+
| BNA | 54.4 | 71.4 | 43.9 | 66.4 | 87.6 |
| ParsBERT | 68.1 | 82.0 | 61.5 | 79.8 | 89.0 |
| XLM-RoBERTa | 74.8 | 87.6 | 69.1 | 86.0 | 92.7 |
| Human | 80.3 | 88.3 | 74.9 | 85.6 | 96.8 |
+-------------+------+------+-----------+-----------+-------------+
```
## LICENSE
PQuAD is developed by Mabna Intelligent Computing at Amirkabir Science and Technology Park with
collaboration of the NLP lab of the Amirkabir University of Technology and is supported by the
Vice Presidency for Scientific and Technology. By releasing this dataset, we aim to ease research
on Persian reading comprehension and the development of Persian question answering systems.
This work is licensed under a
[Creative Commons Attribution-ShareAlike 4.0 International License][cc-by-sa].
[![CC BY-SA 4.0][cc-by-sa-image]][cc-by-sa]
[cc-by-sa]: http://creativecommons.org/licenses/by-sa/4.0/
[cc-by-sa-image]: https://licensebuttons.net/l/by-sa/4.0/88x31.png
[cc-by-sa-shield]: https://img.shields.io/badge/License-CC%20BY--SA%204.0-lightgrey.svg
# Dataset Card for "pquad"