Datasets:
lmqg
/

Languages:
Russian
ArXiv:
License:
This view is limited to 50 files because it contains too many changes.  See the raw diff here.
Files changed (50) hide show
  1. .gitattributes +0 -74
  2. README.md +0 -85
  3. data/processed/test00.jsonl +0 -0
  4. data/processed/test01.jsonl +0 -0
  5. data/processed/test02.jsonl +0 -0
  6. data/processed/test03.jsonl +0 -0
  7. data/processed/test04.jsonl +0 -0
  8. data/processed/test05.jsonl +0 -0
  9. data/processed/test06.jsonl +0 -0
  10. data/processed/test07.jsonl +0 -0
  11. data/processed/test08.jsonl +0 -0
  12. data/processed/test09.jsonl +0 -0
  13. data/processed/test10.jsonl +0 -0
  14. data/processed/test11.jsonl +0 -0
  15. data/processed/test12.jsonl +0 -0
  16. data/processed/test13.jsonl +0 -0
  17. data/processed/test14.jsonl +0 -0
  18. data/processed/test15.jsonl +0 -0
  19. data/processed/test16.jsonl +0 -0
  20. data/processed/test17.jsonl +0 -0
  21. data/processed/test18.jsonl +0 -0
  22. data/processed/test19.jsonl +0 -0
  23. data/processed/test20.jsonl +0 -0
  24. data/processed/test21.jsonl +0 -0
  25. data/processed/test22.jsonl +0 -0
  26. data/processed/test23.jsonl +0 -0
  27. data/processed/test24.jsonl +0 -0
  28. data/processed/test25.jsonl +0 -0
  29. data/processed/test26.jsonl +0 -0
  30. data/processed/test27.jsonl +0 -0
  31. data/processed/test28.jsonl +0 -0
  32. data/processed/test29.jsonl +0 -0
  33. data/processed/test30.jsonl +0 -0
  34. data/processed/test31.jsonl +0 -0
  35. data/processed/test32.jsonl +0 -0
  36. data/processed/test33.jsonl +0 -0
  37. data/processed/test34.jsonl +0 -0
  38. data/processed/test35.jsonl +0 -0
  39. data/processed/test36.jsonl +0 -0
  40. data/processed/test37.jsonl +0 -0
  41. data/processed/test38.jsonl +0 -0
  42. data/processed/test39.jsonl +0 -0
  43. data/processed/train00.jsonl +0 -0
  44. data/processed/train01.jsonl +0 -0
  45. data/processed/train02.jsonl +0 -0
  46. data/processed/train03.jsonl +0 -0
  47. data/processed/train04.jsonl +0 -0
  48. data/processed/train05.jsonl +0 -0
  49. data/processed/train06.jsonl +0 -0
  50. data/processed/train07.jsonl +0 -0
.gitattributes CHANGED
@@ -38,77 +38,3 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
38
  data/processed/test.jsonl filter=lfs diff=lfs merge=lfs -text
39
  data/processed/train.jsonl filter=lfs diff=lfs merge=lfs -text
40
  data/processed/validation.jsonl filter=lfs diff=lfs merge=lfs -text
41
- data/processed/train03.jsonl filter=lfs diff=lfs merge=lfs -text
42
- data/processed/train48.jsonl filter=lfs diff=lfs merge=lfs -text
43
- data/processed/train57.jsonl filter=lfs diff=lfs merge=lfs -text
44
- data/processed/train14.jsonl filter=lfs diff=lfs merge=lfs -text
45
- data/processed/train20.jsonl filter=lfs diff=lfs merge=lfs -text
46
- data/processed/train30.jsonl filter=lfs diff=lfs merge=lfs -text
47
- data/processed/test00.jsonl filter=lfs diff=lfs merge=lfs -text
48
- data/processed/train01.jsonl filter=lfs diff=lfs merge=lfs -text
49
- data/processed/train02.jsonl filter=lfs diff=lfs merge=lfs -text
50
- data/processed/train19.jsonl filter=lfs diff=lfs merge=lfs -text
51
- data/processed/train35.jsonl filter=lfs diff=lfs merge=lfs -text
52
- data/processed/train46.jsonl filter=lfs diff=lfs merge=lfs -text
53
- data/processed/validation01.jsonl filter=lfs diff=lfs merge=lfs -text
54
- data/processed/validation03.jsonl filter=lfs diff=lfs merge=lfs -text
55
- data/processed/test06.jsonl filter=lfs diff=lfs merge=lfs -text
56
- data/processed/train18.jsonl filter=lfs diff=lfs merge=lfs -text
57
- data/processed/train39.jsonl filter=lfs diff=lfs merge=lfs -text
58
- data/processed/train40.jsonl filter=lfs diff=lfs merge=lfs -text
59
- data/processed/train06.jsonl filter=lfs diff=lfs merge=lfs -text
60
- data/processed/train28.jsonl filter=lfs diff=lfs merge=lfs -text
61
- data/processed/train41.jsonl filter=lfs diff=lfs merge=lfs -text
62
- data/processed/train52.jsonl filter=lfs diff=lfs merge=lfs -text
63
- data/processed/test02.jsonl filter=lfs diff=lfs merge=lfs -text
64
- data/processed/train38.jsonl filter=lfs diff=lfs merge=lfs -text
65
- data/processed/train42.jsonl filter=lfs diff=lfs merge=lfs -text
66
- data/processed/train47.jsonl filter=lfs diff=lfs merge=lfs -text
67
- data/processed/validation00.jsonl filter=lfs diff=lfs merge=lfs -text
68
- data/processed/validation05.jsonl filter=lfs diff=lfs merge=lfs -text
69
- data/processed/test01.jsonl filter=lfs diff=lfs merge=lfs -text
70
- data/processed/test03.jsonl filter=lfs diff=lfs merge=lfs -text
71
- data/processed/test04.jsonl filter=lfs diff=lfs merge=lfs -text
72
- data/processed/train04.jsonl filter=lfs diff=lfs merge=lfs -text
73
- data/processed/train13.jsonl filter=lfs diff=lfs merge=lfs -text
74
- data/processed/train34.jsonl filter=lfs diff=lfs merge=lfs -text
75
- data/processed/train07.jsonl filter=lfs diff=lfs merge=lfs -text
76
- data/processed/train21.jsonl filter=lfs diff=lfs merge=lfs -text
77
- data/processed/train55.jsonl filter=lfs diff=lfs merge=lfs -text
78
- data/processed/test05.jsonl filter=lfs diff=lfs merge=lfs -text
79
- data/processed/train27.jsonl filter=lfs diff=lfs merge=lfs -text
80
- data/processed/train36.jsonl filter=lfs diff=lfs merge=lfs -text
81
- data/processed/train50.jsonl filter=lfs diff=lfs merge=lfs -text
82
- data/processed/validation04.jsonl filter=lfs diff=lfs merge=lfs -text
83
- data/processed/validation07.jsonl filter=lfs diff=lfs merge=lfs -text
84
- data/processed/test07.jsonl filter=lfs diff=lfs merge=lfs -text
85
- data/processed/train09.jsonl filter=lfs diff=lfs merge=lfs -text
86
- data/processed/train15.jsonl filter=lfs diff=lfs merge=lfs -text
87
- data/processed/train23.jsonl filter=lfs diff=lfs merge=lfs -text
88
- data/processed/train44.jsonl filter=lfs diff=lfs merge=lfs -text
89
- data/processed/train45.jsonl filter=lfs diff=lfs merge=lfs -text
90
- data/processed/train51.jsonl filter=lfs diff=lfs merge=lfs -text
91
- data/processed/train16.jsonl filter=lfs diff=lfs merge=lfs -text
92
- data/processed/train29.jsonl filter=lfs diff=lfs merge=lfs -text
93
- data/processed/train49.jsonl filter=lfs diff=lfs merge=lfs -text
94
- data/processed/train53.jsonl filter=lfs diff=lfs merge=lfs -text
95
- data/processed/train56.jsonl filter=lfs diff=lfs merge=lfs -text
96
- data/processed/train08.jsonl filter=lfs diff=lfs merge=lfs -text
97
- data/processed/train11.jsonl filter=lfs diff=lfs merge=lfs -text
98
- data/processed/train25.jsonl filter=lfs diff=lfs merge=lfs -text
99
- data/processed/train05.jsonl filter=lfs diff=lfs merge=lfs -text
100
- data/processed/train22.jsonl filter=lfs diff=lfs merge=lfs -text
101
- data/processed/train26.jsonl filter=lfs diff=lfs merge=lfs -text
102
- data/processed/validation02.jsonl filter=lfs diff=lfs merge=lfs -text
103
- data/processed/train00.jsonl filter=lfs diff=lfs merge=lfs -text
104
- data/processed/train10.jsonl filter=lfs diff=lfs merge=lfs -text
105
- data/processed/train12.jsonl filter=lfs diff=lfs merge=lfs -text
106
- data/processed/train17.jsonl filter=lfs diff=lfs merge=lfs -text
107
- data/processed/train24.jsonl filter=lfs diff=lfs merge=lfs -text
108
- data/processed/train32.jsonl filter=lfs diff=lfs merge=lfs -text
109
- data/processed/train43.jsonl filter=lfs diff=lfs merge=lfs -text
110
- data/processed/train54.jsonl filter=lfs diff=lfs merge=lfs -text
111
- data/processed/validation06.jsonl filter=lfs diff=lfs merge=lfs -text
112
- data/processed/train31.jsonl filter=lfs diff=lfs merge=lfs -text
113
- data/processed/train33.jsonl filter=lfs diff=lfs merge=lfs -text
114
- data/processed/train37.jsonl filter=lfs diff=lfs merge=lfs -text
 
38
  data/processed/test.jsonl filter=lfs diff=lfs merge=lfs -text
39
  data/processed/train.jsonl filter=lfs diff=lfs merge=lfs -text
40
  data/processed/validation.jsonl filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
README.md CHANGED
@@ -1,85 +0,0 @@
1
- ---
2
- license: cc-by-4.0
3
- pretty_name: SberQuAD for question generation
4
- language: ru
5
- multilinguality: monolingual
6
- size_categories: 10K<n<100K
7
- source_datasets: deepset/germanquad
8
- task_categories:
9
- - text-generation
10
- task_ids:
11
- - language-modeling
12
- tags:
13
- - question-generation
14
- ---
15
-
16
- # Dataset Card for "lmqg/qg_ruquad"
17
-
18
-
19
- ## Dataset Description
20
- - **Repository:** [https://github.com/asahi417/lm-question-generation](https://github.com/asahi417/lm-question-generation)
21
- - **Paper:** [https://arxiv.org/abs/2210.03992](https://arxiv.org/abs/2210.03992)
22
- - **Point of Contact:** [Asahi Ushio](http://asahiushio.com/)
23
-
24
- ### Dataset Summary
25
- This is a subset of [QG-Bench](https://github.com/asahi417/lm-question-generation/blob/master/QG_BENCH.md#datasets), a unified question generation benchmark proposed in
26
- ["Generative Language Models for Paragraph-Level Question Generation: A Unified Benchmark and Evaluation, EMNLP 2022 main conference"](https://arxiv.org/abs/2210.03992).
27
- This is a modified version of [SberQuaD](https://huggingface.co/datasets/sberquad) for question generation (QG) task.
28
- Since the original dataset only contains training/validation set, we manually sample test set from training set, which
29
- has no overlap in terms of the paragraph with the training set.
30
-
31
- ### Supported Tasks and Leaderboards
32
- * `question-generation`: The dataset is assumed to be used to train a model for question generation.
33
- Success on this task is typically measured by achieving a high BLEU4/METEOR/ROUGE-L/BERTScore/MoverScore (see our paper for more in detail).
34
-
35
- ### Languages
36
- Russian (ru)
37
-
38
- ## Dataset Structure
39
- An example of 'train' looks as follows.
40
- ```
41
- {
42
- 'answer': 'известковыми выделениями сине-зелёных водорослей',
43
- 'question': 'чем представлены органические остатки?',
44
- 'sentence': 'Они представлены известковыми выделениями сине-зелёных водорослей , ходами червей, остатками кишечнополостных.'
45
- 'paragraph': "В протерозойских отложениях органические остатки встречаются намного чаще, чем в архейских. Они представлены..."
46
- 'sentence_answer': "Они представлены <hl> известковыми выделениями сине-зелёных водорослей <hl> , ход...",
47
- 'paragraph_answer': "В протерозойских отложениях органические остатки встречаются намного чаще, чем в архейских. Они представлены <hl> известковыми выделениям...",
48
- 'paragraph_sentence': "В протерозойских отложениях органические остатки встречаются намного чаще, чем в архейских. <hl> Они представлены известковыми выделениями сине-зелёных водорослей , ходами червей, остатками кишечнополостных. <hl> Кроме..."
49
- }
50
- ```
51
- The data fields are the same among all splits.
52
- - `question`: a `string` feature.
53
- - `paragraph`: a `string` feature.
54
- - `answer`: a `string` feature.
55
- - `sentence`: a `string` feature.
56
- - `paragraph_answer`: a `string` feature, which is same as the paragraph but the answer is highlighted by a special token `<hl>`.
57
- - `paragraph_sentence`: a `string` feature, which is same as the paragraph but a sentence containing the answer is highlighted by a special token `<hl>`.
58
- - `sentence_answer`: a `string` feature, which is same as the sentence but the answer is highlighted by a special token `<hl>`.
59
-
60
- Each of `paragraph_answer`, `paragraph_sentence`, and `sentence_answer` feature is assumed to be used to train a question generation model,
61
- but with different information. The `paragraph_answer` and `sentence_answer` features are for answer-aware question generation and
62
- `paragraph_sentence` feature is for sentence-aware question generation.
63
-
64
- ## Data Splits
65
-
66
- |train|validation|test |
67
- |----:|---------:|----:|
68
- | 45327 | 5036 |23936 |
69
-
70
-
71
- ## Citation Information
72
-
73
- ```
74
- @inproceedings{ushio-etal-2022-generative,
75
- title = "{G}enerative {L}anguage {M}odels for {P}aragraph-{L}evel {Q}uestion {G}eneration",
76
- author = "Ushio, Asahi and
77
- Alva-Manchego, Fernando and
78
- Camacho-Collados, Jose",
79
- booktitle = "Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing",
80
- month = dec,
81
- year = "2022",
82
- address = "Abu Dhabi, U.A.E.",
83
- publisher = "Association for Computational Linguistics",
84
- }
85
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
data/processed/test00.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
data/processed/test01.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
data/processed/test02.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
data/processed/test03.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
data/processed/test04.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
data/processed/test05.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
data/processed/test06.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
data/processed/test07.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
data/processed/test08.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test09.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test10.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test11.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test12.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test13.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test14.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test15.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test16.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test17.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test18.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test19.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test20.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test21.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test22.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test23.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test24.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test25.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test26.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test27.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test28.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test29.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test30.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test31.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test32.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test33.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test34.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test35.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test36.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test37.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test38.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/test39.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
data/processed/train00.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
data/processed/train01.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
data/processed/train02.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
data/processed/train03.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
data/processed/train04.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
data/processed/train05.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
data/processed/train06.jsonl CHANGED
The diff for this file is too large to render. See raw diff
 
data/processed/train07.jsonl CHANGED
The diff for this file is too large to render. See raw diff