Datasets:
Tasks:
Text Generation
Modalities:
Text
Sub-tasks:
language-modeling
Languages:
English
Size:
10K - 100K
ArXiv:
Tags:
question-generation
License:
init
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- README.md +122 -0
- data/processed/dev00.jsonl +0 -0
- data/processed/dev01.jsonl +0 -0
- data/processed/dev02.jsonl +0 -0
- data/processed/dev03.jsonl +0 -0
- data/processed/test00.jsonl +0 -0
- data/processed/test01.jsonl +0 -0
- data/processed/test02.jsonl +0 -0
- data/processed/test03.jsonl +0 -0
- data/processed/train00.jsonl +0 -0
- data/processed/train01.jsonl +0 -0
- data/processed/train02.jsonl +0 -0
- data/processed/train03.jsonl +0 -0
- data/processed/train04.jsonl +0 -0
- data/processed/train05.jsonl +0 -0
- data/processed/train06.jsonl +0 -0
- data/processed/train07.jsonl +0 -0
- data/processed/train08.jsonl +0 -0
- data/processed/train09.jsonl +0 -0
- data/processed/train10.jsonl +0 -0
- data/processed/train11.jsonl +0 -0
- data/processed/train12.jsonl +0 -0
- data/processed/train13.jsonl +0 -0
- data/processed/train14.jsonl +0 -0
- data/processed/train15.jsonl +0 -0
- data/processed/train16.jsonl +0 -0
- data/processed/train17.jsonl +0 -0
- data/processed/train18.jsonl +0 -0
- data/processed/train19.jsonl +0 -0
- data/processed/train20.jsonl +0 -0
- data/processed/train21.jsonl +0 -0
- data/processed/train22.jsonl +0 -0
- data/raw/dev.jsonl +0 -0
- data/raw/test.jsonl +0 -0
- data/raw/train00.jsonl +0 -0
- data/raw/train01.jsonl +0 -0
- data/raw/train02.jsonl +0 -0
- data/raw/train03.jsonl +0 -0
- data/raw/train04.jsonl +0 -0
- data/raw/train05.jsonl +0 -0
- data/raw/train06.jsonl +0 -0
- data/raw/train07.jsonl +0 -0
- data/raw/train08.jsonl +0 -0
- data/raw/train09.jsonl +0 -0
- data/raw/train10.jsonl +0 -0
- process.py +102 -0
- qg_squad.py +64 -0
- reference_files/ans-dev-normalized.txt +0 -0
- reference_files/ans-dev.txt +0 -0
- reference_files/ans-test-normalized.txt +0 -0
README.md
CHANGED
@@ -1,3 +1,125 @@
|
|
1 |
---
|
2 |
license: mit
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: mit
|
3 |
+
pretty_name: SQuAD QG
|
4 |
+
languages: en
|
5 |
+
multilinguality: monolingual
|
6 |
+
size_categories: 10K<n<100K
|
7 |
+
source_datasets: extended|wikipedia
|
8 |
+
task_categories: question-generation
|
9 |
+
task_ids: question-generation
|
10 |
---
|
11 |
+
|
12 |
+
# Dataset Card for "qg_squad"
|
13 |
+
|
14 |
+
## Table of Contents
|
15 |
+
- [Dataset Description](#dataset-description)
|
16 |
+
- [Dataset Summary](#dataset-summary)
|
17 |
+
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
|
18 |
+
- [Languages](#languages)
|
19 |
+
- [Dataset Structure](#dataset-structure)
|
20 |
+
- [Data Instances](#data-instances)
|
21 |
+
- [Data Fields](#data-fields)
|
22 |
+
- [Data Splits](#data-splits)
|
23 |
+
- [Dataset Creation](#dataset-creation)
|
24 |
+
- [Curation Rationale](#curation-rationale)
|
25 |
+
- [Source Data](#source-data)
|
26 |
+
- [Annotations](#annotations)
|
27 |
+
- [Personal and Sensitive Information](#personal-and-sensitive-information)
|
28 |
+
- [Considerations for Using the Data](#considerations-for-using-the-data)
|
29 |
+
- [Social Impact of Dataset](#social-impact-of-dataset)
|
30 |
+
- [Discussion of Biases](#discussion-of-biases)
|
31 |
+
- [Other Known Limitations](#other-known-limitations)
|
32 |
+
- [Additional Information](#additional-information)
|
33 |
+
- [Dataset Curators](#dataset-curators)
|
34 |
+
- [Licensing Information](#licensing-information)
|
35 |
+
- [Citation Information](#citation-information)
|
36 |
+
- [Contributions](#contributions)
|
37 |
+
|
38 |
+
## Dataset Description
|
39 |
+
- **Repository:** [https://github.com/asahi417/lm-question-generation](https://github.com/asahi417/lm-question-generation)
|
40 |
+
- **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
|
41 |
+
- **Point of Contact:** [Asahi Ushio](http://asahiushio.com/)
|
42 |
+
- **Size of downloaded dataset files:** 33.51 MB
|
43 |
+
- **Size of the generated dataset:** 85.75 MB
|
44 |
+
- **Total amount of disk used:** 119.27 MB
|
45 |
+
|
46 |
+
### Dataset Summary
|
47 |
+
[SQuAD](https://rajpurkar.github.io/SQuAD-explorer/) dataset for question generation (QG) task. The split
|
48 |
+
of train/development/test set follows the ["Neural Question Generation"](https://arxiv.org/abs/1705.00106) work and is
|
49 |
+
compatible with the [leader board](https://paperswithcode.com/sota/question-generation-on-squad11).
|
50 |
+
|
51 |
+
|
52 |
+
### Supported Tasks and Leaderboards
|
53 |
+
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
|
54 |
+
|
55 |
+
### Languages
|
56 |
+
English (en)
|
57 |
+
|
58 |
+
## Dataset Structure
|
59 |
+
### Data Instances
|
60 |
+
#### plain_text
|
61 |
+
- **Size of downloaded dataset files:** 33.51 MB
|
62 |
+
- **Size of the generated dataset:** 85.75 MB
|
63 |
+
- **Total amount of disk used:** 119.27 MB
|
64 |
+
An example of 'train' looks as follows.
|
65 |
+
```
|
66 |
+
{
|
67 |
+
"question": "What is heresy mainly at odds with?",
|
68 |
+
"passage": "Heresy is any provocative belief or theory that is strongly at variance with established beliefs or customs."
|
69 |
+
"A heretic is a proponent of such claims or beliefs. Heresy is distinct from both apostasy, which is the explicit"
|
70 |
+
"renunciation of one's religion, principles or cause, and blasphemy, which is an impious utterance or action concerning God or sacred things.",
|
71 |
+
"answer": "established beliefs or customs",
|
72 |
+
"sentence": "Heresy is any provocative belief or theory that is strongly at variance with established beliefs or customs .",
|
73 |
+
"passage_sentence": "<hl> Heresy is any provocative belief or theory that is strongly at variance with established beliefs or customs . <hl> A heretic is a proponent of such claims or beliefs. Heresy is distinct from both apostasy, which is the explicit renunciation of one's religion, principles or cause, and blasphemy, which is an impious utterance or action concerning God or sacred things.",
|
74 |
+
"passage_answer": "Heresy is any provocative belief or theory that is strongly at variance with <hl> established beliefs or customs <hl>. A heretic is a proponent of such claims or beliefs. Heresy is distinct from both apostasy, which is the explicit renunciation of one's religion, principles or cause, and blasphemy, which is an impious utterance or action concerning God or sacred things.",
|
75 |
+
"sentence_answer": "Heresy is any provocative belief or theory that is strongly at variance with <hl> established beliefs or customs <hl> ."
|
76 |
+
}
|
77 |
+
```
|
78 |
+
### Data Fields
|
79 |
+
The data fields are the same among all splits.
|
80 |
+
#### plain_text
|
81 |
+
- `question`: a `string` feature.
|
82 |
+
- `passage`: a `string` feature.
|
83 |
+
- `answer`: a `string` feature.
|
84 |
+
- `sentence`: a `string` feature.
|
85 |
+
- `passage_sentence`: a `string` feature.
|
86 |
+
- `passage_answer`: a `string` feature.
|
87 |
+
- `sentence_answer`: a `string` feature.
|
88 |
+
|
89 |
+
### Data Splits
|
90 |
+
|
91 |
+
| name |train|validation|test |
|
92 |
+
|----------|----:|---------:|----:|
|
93 |
+
|plain_text|75722| 10570|11877|
|
94 |
+
|
95 |
+
## Dataset Creation
|
96 |
+
### Curation Rationale
|
97 |
+
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
|
98 |
+
### Source Data
|
99 |
+
#### Initial Data Collection and Normalization
|
100 |
+
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
|
101 |
+
#### Who are the source language producers?
|
102 |
+
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
|
103 |
+
### Annotations
|
104 |
+
#### Annotation process
|
105 |
+
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
|
106 |
+
#### Who are the annotators?
|
107 |
+
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
|
108 |
+
### Personal and Sensitive Information
|
109 |
+
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
|
110 |
+
## Considerations for Using the Data
|
111 |
+
### Social Impact of Dataset
|
112 |
+
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
|
113 |
+
### Discussion of Biases
|
114 |
+
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
|
115 |
+
### Other Known Limitations
|
116 |
+
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
|
117 |
+
## Additional Information
|
118 |
+
### Dataset Curators
|
119 |
+
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
|
120 |
+
### Licensing Information
|
121 |
+
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
|
122 |
+
### Citation Information
|
123 |
+
```
|
124 |
+
TBA
|
125 |
+
```
|
data/processed/dev00.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/dev01.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/dev02.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/dev03.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/test00.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/test01.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/test02.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/test03.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train00.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train01.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train02.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train03.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train04.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train05.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train06.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train07.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train08.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train09.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train10.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train11.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train12.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train13.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train14.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train15.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train16.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train17.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train18.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train19.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train20.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train21.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/processed/train22.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/raw/dev.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/raw/test.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/raw/train00.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/raw/train01.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/raw/train02.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/raw/train03.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/raw/train04.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/raw/train05.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/raw/train06.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/raw/train07.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/raw/train08.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/raw/train09.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
data/raw/train10.jsonl
ADDED
The diff for this file is too large to render.
See raw diff
|
|
process.py
ADDED
@@ -0,0 +1,102 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
""" Script to process raw SQuAD file for Question Generation format
|
2 |
+
You need to run `python -m spacy download en_core_web_sm`.
|
3 |
+
Split when uploading to dataset hub by
|
4 |
+
```
|
5 |
+
gsplit -l 3300 -d --additional-suffix=.jsonl train.jsonl train
|
6 |
+
gsplit -l 3300 -d --additional-suffix=.jsonl test.jsonl test
|
7 |
+
gsplit -l 3300 -d --additional-suffix=.jsonl dev.jsonl dev
|
8 |
+
```
|
9 |
+
"""
|
10 |
+
import json
|
11 |
+
import os
|
12 |
+
import re
|
13 |
+
from glob import glob
|
14 |
+
from tqdm import tqdm
|
15 |
+
from typing import List, Dict
|
16 |
+
|
17 |
+
import spacy
|
18 |
+
|
19 |
+
SPLITTER = spacy.load('en_core_web_sm')
|
20 |
+
HIGHLIGHT_TOKEN = '<hl>'
|
21 |
+
|
22 |
+
|
23 |
+
def get_sentence(document: str):
|
24 |
+
return [str(s) for s in SPLITTER(document).sents]
|
25 |
+
|
26 |
+
|
27 |
+
def jsonline_reader(filename: str):
|
28 |
+
with open(filename, 'r') as f_reader:
|
29 |
+
examples = [json.loads(i) for i in f_reader.read().split('\n') if len(i) > 0]
|
30 |
+
return examples
|
31 |
+
|
32 |
+
|
33 |
+
def process_single_data(data: Dict):
|
34 |
+
""" Convert single raw json data into QG format """
|
35 |
+
example = {'question': data["question"], 'passage': data["context"], 'answer': data["answer"]}
|
36 |
+
|
37 |
+
# get sentence
|
38 |
+
position = example['passage'].find(example['answer'])
|
39 |
+
assert position != -1
|
40 |
+
before_tmp = get_sentence(example['passage'][:position])
|
41 |
+
if len(before_tmp) == 0:
|
42 |
+
before = ''
|
43 |
+
before_sentence = ''
|
44 |
+
else:
|
45 |
+
if before_tmp[-1].endswith('.'):
|
46 |
+
before = ' '.join(before_tmp)
|
47 |
+
before_sentence = ''
|
48 |
+
else:
|
49 |
+
before = ' '.join(before_tmp[:-1])
|
50 |
+
before_sentence = before_tmp[-1]
|
51 |
+
before_sentence = before_sentence if before_sentence.endswith(' ') else '{} '.format(before_sentence)
|
52 |
+
after_tmp = get_sentence(example['passage'][position + len(example['answer']):])
|
53 |
+
if len(after_tmp) == 0:
|
54 |
+
after = ''
|
55 |
+
after_sentence = ''
|
56 |
+
else:
|
57 |
+
after = ' '.join(after_tmp[1:])
|
58 |
+
after_sentence = after_tmp[0]
|
59 |
+
after_sentence = after_sentence if after_sentence.startswith(' ') else ' {}'.format(after_sentence)
|
60 |
+
example['sentence'] = '{}{}{}'.format(before_sentence, example['answer'], after_sentence)
|
61 |
+
|
62 |
+
# get passage_sentence
|
63 |
+
before = '' if before == '' else '{} '.format(before)
|
64 |
+
after = '' if after == '' else ' {}'.format(after)
|
65 |
+
source_text = '{0}{1} {2} {1}{3}'.format(before, HIGHLIGHT_TOKEN, example['sentence'], after)
|
66 |
+
example['passage_sentence'] = re.sub(r'\s+', ' ', source_text)
|
67 |
+
|
68 |
+
# get passage_answer
|
69 |
+
source_text = '{0}{1} {2} {1}{3}'.format(
|
70 |
+
example['passage'][:position], HIGHLIGHT_TOKEN, example['answer'],
|
71 |
+
example['passage'][position + len(example['answer']):])
|
72 |
+
example['passage_answer'] = re.sub(r'\s+', ' ', source_text)
|
73 |
+
|
74 |
+
# get sentence_answer
|
75 |
+
before = get_sentence(example['passage'][:position])
|
76 |
+
if len(before) == 0 or before[-1].endswith('.'):
|
77 |
+
before = ''
|
78 |
+
else:
|
79 |
+
before = before[-1] if before[-1].endswith(' ') else '{} '.format(before[-1])
|
80 |
+
after = get_sentence(example['passage'][position + len(example['answer']):])
|
81 |
+
if len(after) == 0:
|
82 |
+
after = ''
|
83 |
+
else:
|
84 |
+
after = after[0] if after[0].startswith(' ') else ' {}'.format(after[0])
|
85 |
+
source_text = '{0}{1} {2} {1}{3}'.format(before, HIGHLIGHT_TOKEN, example['answer'], after)
|
86 |
+
example['sentence_answer'] = re.sub(r'\s+', ' ', source_text)
|
87 |
+
|
88 |
+
return example
|
89 |
+
|
90 |
+
|
91 |
+
if __name__ == '__main__':
|
92 |
+
output = './data/processed'
|
93 |
+
os.makedirs(output, exist_ok=True)
|
94 |
+
path = {'train': 'data/raw/train*.jsonl', 'dev': 'data/raw/dev.jsonl', 'test': 'data/raw/test.jsonl'}
|
95 |
+
for k, v in path.items():
|
96 |
+
json_data = []
|
97 |
+
for _file in sorted(glob(v)):
|
98 |
+
json_data += jsonline_reader(_file)
|
99 |
+
with open('{}/{}.jsonl'.format(output, k), 'w') as f:
|
100 |
+
for single_data in tqdm(json_data):
|
101 |
+
single_data = process_single_data(single_data)
|
102 |
+
f.write(json.dumps(single_data) + '\n')
|
qg_squad.py
ADDED
@@ -0,0 +1,64 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
from datasets import load_dataset
|
2 |
+
from datasets.tasks import Summarization
|
3 |
+
|
4 |
+
_DESCRIPTION = """
|
5 |
+
[SQuAD](https://rajpurkar.github.io/SQuAD-explorer/) evaluation set for the question generation (QG) models. The split
|
6 |
+
of test and development set follows the ["Neural Question Generation"](https://arxiv.org/abs/1705.00106) work and is
|
7 |
+
compatible with the [leader board](https://paperswithcode.com/sota/question-generation-on-squad11).
|
8 |
+
"""
|
9 |
+
_URL = 'https://huggingface.co/datasets/asahi417/squad_qg/data/processed'
|
10 |
+
_URLS = {
|
11 |
+
'train': ['{}/train{:02d}.jsonl'.format(_URL, i) for i in range(23)],
|
12 |
+
'test': ['{}/test{:02d}.jsonl'.format(_URL, i) for i in range(4)],
|
13 |
+
'validation': ['{}/dev{:02d}.jsonl'.format(_URL, i) for i in range(4)]
|
14 |
+
}
|
15 |
+
|
16 |
+
|
17 |
+
class SquadQGConfig(datasets.BuilderConfig):
|
18 |
+
"""BuilderConfig for SquadQG"""
|
19 |
+
|
20 |
+
def __init__(self, **kwargs):
|
21 |
+
"""BuilderConfig for SquadQG.
|
22 |
+
Args:
|
23 |
+
**kwargs: keyword arguments forwarded to super.
|
24 |
+
"""
|
25 |
+
super(SquadQGConfig, self).__init__(**kwargs)
|
26 |
+
|
27 |
+
|
28 |
+
class SquadQG(datasets.GeneratorBasedBuilder):
|
29 |
+
|
30 |
+
def _info(self):
|
31 |
+
return datasets.DatasetInfo(
|
32 |
+
description=_DESCRIPTION,
|
33 |
+
features=datasets.Features(
|
34 |
+
{
|
35 |
+
"answer": datasets.Value("string"),
|
36 |
+
"question": datasets.Value("string"),
|
37 |
+
"sentence": datasets.Value("string"),
|
38 |
+
"passage": datasets.Value("string"),
|
39 |
+
"sentence_answer": datasets.Value("string"),
|
40 |
+
"passage_answer": datasets.Value("string"),
|
41 |
+
"passage_sentence": datasets.Value("string")
|
42 |
+
}
|
43 |
+
),
|
44 |
+
supervised_keys=None,
|
45 |
+
task_templates=[
|
46 |
+
Summarization(task='question generation', text_column="passage_answer", summary_column='question')
|
47 |
+
],
|
48 |
+
homepage="https://github.com/asahi417/lm-question-generation"
|
49 |
+
)
|
50 |
+
|
51 |
+
def _split_generators(self, dl_manager):
|
52 |
+
downloaded_file = dl_manager.download_and_extract(_URLS)
|
53 |
+
return [
|
54 |
+
datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": downloaded_file["train"]}),
|
55 |
+
datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={"filepath": downloaded_file["validation"]}),
|
56 |
+
datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"filepath": downloaded_file["test"]}),
|
57 |
+
]
|
58 |
+
|
59 |
+
def _generate_examples(self, filepath):
|
60 |
+
logger.info("generating examples from = %s", filepath)
|
61 |
+
with open(filepath, encoding="utf-8") as f:
|
62 |
+
for _id, i in enumerate(f.read().split('\n')):
|
63 |
+
data = json.loads(i)
|
64 |
+
yield _id, data
|
reference_files/ans-dev-normalized.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
reference_files/ans-dev.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
reference_files/ans-test-normalized.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|