gigaword_tiny / README.md
SpeedOfMagic's picture
Update README.md
f2876f5 verified
---
dataset_info:
features:
- name: document
dtype: string
- name: summary
dtype: string
splits:
- name: train
num_bytes: 23380
num_examples: 100
- name: validation
num_bytes: 23634
num_examples: 100
- name: test
num_bytes: 24038
num_examples: 100
download_size: 55197
dataset_size: 71052
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
- split: test
path: data/test-*
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
This is a tiny version of [https://huggingface.co/datasets/Harvard/gigaword](Harvard/gigaword), used for testing purposes.
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
This is a tiny version of [https://huggingface.co/datasets/Harvard/gigaword](Harvard/gigaword).
It was created by selecting only first 100 samples from each split.
- **Language(s) (NLP):** English
### Dataset Sources [optional]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
It is supposed be used only for testing purposes.
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
It is supposed be used only when you want to test that your code works.
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
Dataset contains two string columns: `document` and `summary`. `document` shows source document, and `summary` is summarization of document.
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
To test code without waiting a long time to download gigawords.