File size: 1,205 Bytes
29e8a78
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f7379d8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29e8a78
f7379d8
29e8a78
f7379d8
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
---
dataset_info:
  features:
  - name: url
    dtype: string
  - name: title
    dtype: string
  - name: summary
    dtype: string
  - name: article
    dtype: string
  - name: step_headers
    dtype: string
  splits:
  - name: train
    num_bytes: 315275236
    num_examples: 35775
  - name: test
    num_bytes: 17584216
    num_examples: 2000
  - name: validation
    num_bytes: 17880851
    num_examples: 2000
  download_size: 194202865
  dataset_size: 350740303
license:
- unknown
task_categories:
- summarization
language:
- en
multilinguality:
- monolingual
tags:
- abstractive-summarization
- wiki
- abstractive
pretty_name: 'WikiSum: Coherent Summarization Dataset for Efficient Human-Evaluation'
size_categories:
- 10K<n<100K
source_datasets:
- original
paperswithcode_id: wikisum
---
# wikisum

## Dataset Description

- **Homepage:** https://registry.opendata.aws/wikisum/
- **Repository:** https://github.com/tensorflow/tensor2tensor/tree/master/tensor2tensor/data_generators/wikisum
- **Paper:** [Generating Wikipedia by Summarizing Long Sequences](https://arxiv.org/abs/1801.10198)
- **Leaderboard:** [More Information Needed]
- **Point of Contact:** [nachshon](mailto:nachshon@amazon.com)