Datasets:
Tasks:
Text Generation
Modalities:
Text
Formats:
parquet
Languages:
English
Size:
10K - 100K
ArXiv:
License:
license: mit | |
dataset_info: | |
features: | |
- name: text | |
dtype: string | |
splits: | |
- name: train | |
num_bytes: 129624 | |
num_examples: 10000 | |
- name: validation_top1 | |
num_bytes: 10754 | |
num_examples: 1000 | |
- name: test_top1 | |
num_bytes: 10948 | |
num_examples: 1000 | |
- name: validation_1_10 | |
num_bytes: 11618 | |
num_examples: 1000 | |
- name: test_1_10 | |
num_bytes: 11692 | |
num_examples: 1000 | |
- name: validation_10_20 | |
num_bytes: 13401 | |
num_examples: 1000 | |
- name: test_10_20 | |
num_bytes: 13450 | |
num_examples: 1000 | |
- name: validation_20_30 | |
num_bytes: 15112 | |
num_examples: 1000 | |
- name: test_20_30 | |
num_bytes: 15069 | |
num_examples: 1000 | |
- name: validation_bottom50 | |
num_bytes: 15204 | |
num_examples: 1000 | |
- name: test_bottom50 | |
num_bytes: 15076 | |
num_examples: 1000 | |
download_size: 241234 | |
dataset_size: 261948 | |
language: | |
- en | |
viewer: true | |
task_categories: | |
- text-generation | |
size_categories: | |
- 1K<n<10K | |
# WikiSpell | |
## Description | |
This dataset is a **custom implementation** of the WikiSpell dataset introduced in [Character-Aware Models Improve Visual Text Rendering](https://arxiv.org/pdf/2212.10562.pdf) by Liu et al. (2022). | |
Similarly to the original WikiSpell dataset, the training set is composed of 5000 words taken uniformly from the 50% least common Wiktionary words (taken from [this Wiktionary extraction](https://kaikki.org/dictionary/rawdata.html)), and 5000 words sampled according to their frequencies taken from the 50% most common Wiktionary words. | |
The validation and test are splitted in 5 sets, sampled depending on their frequency in the corpus: | |
- 1% most common words | |
- 1 - 10% most common words | |
- 10 - 20% most common words | |
- 20 - 30% most common words | |
- 50% least common words | |
Contrary to the original WikiSpell dataset, we compute the frequency of the words using the first 100k sentences from OpenWebText ([Skylion007/openwebtext](https://huggingface.co/datasets/Skylion007/openwebtext)) instead of mC4. | |
## Usage | |
This dataset is used for testing spelling in Large Language Models. To do so, the labels should be computed like in the following snippet: | |
```python | |
sample = ds["train"][0] | |
label = " ".join(sample["text"]) | |
``` | |
**The labels are not included in the dataset files directly.** | |
## Citation | |
Please cite the original paper introducing WikiSpell if you're using this dataset: | |
``` | |
@inproceedings{liu-etal-2023-character, | |
title = "Character-Aware Models Improve Visual Text Rendering", | |
author = "Liu, Rosanne and | |
Garrette, Dan and | |
Saharia, Chitwan and | |
Chan, William and | |
Roberts, Adam and | |
Narang, Sharan and | |
Blok, Irina and | |
Mical, Rj and | |
Norouzi, Mohammad and | |
Constant, Noah", | |
booktitle = "Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", | |
month = jul, | |
year = "2023", | |
address = "Toronto, Canada", | |
publisher = "Association for Computational Linguistics", | |
url = "https://aclanthology.org/2023.acl-long.900", | |
pages = "16270--16297", | |
} | |
``` |