lince / README.md
albertvillanova's picture
Convert dataset sizes from base 2 to base 10 in the dataset card (#2)
2ad124d
---
paperswithcode_id: lince
pretty_name: Linguistic Code-switching Evaluation Dataset
dataset_info:
- config_name: lid_spaeng
features:
- name: idx
dtype: int32
- name: words
sequence: string
- name: lid
sequence: string
splits:
- name: train
num_bytes: 4745003
num_examples: 21030
- name: validation
num_bytes: 739950
num_examples: 3332
- name: test
num_bytes: 1337727
num_examples: 8289
download_size: 1188861
dataset_size: 6822680
- config_name: lid_hineng
features:
- name: idx
dtype: int32
- name: words
sequence: string
- name: lid
sequence: string
splits:
- name: train
num_bytes: 1662284
num_examples: 4823
- name: validation
num_bytes: 268930
num_examples: 744
- name: test
num_bytes: 456850
num_examples: 1854
download_size: 432854
dataset_size: 2388064
- config_name: lid_msaea
features:
- name: idx
dtype: int32
- name: words
sequence: string
- name: lid
sequence: string
splits:
- name: train
num_bytes: 3804156
num_examples: 8464
- name: validation
num_bytes: 490566
num_examples: 1116
- name: test
num_bytes: 590488
num_examples: 1663
download_size: 803806
dataset_size: 4885210
- config_name: lid_nepeng
features:
- name: idx
dtype: int32
- name: words
sequence: string
- name: lid
sequence: string
splits:
- name: train
num_bytes: 2239014
num_examples: 8451
- name: validation
num_bytes: 351649
num_examples: 1332
- name: test
num_bytes: 620512
num_examples: 3228
download_size: 545342
dataset_size: 3211175
- config_name: pos_spaeng
features:
- name: idx
dtype: int32
- name: words
sequence: string
- name: lid
sequence: string
- name: pos
sequence: string
splits:
- name: train
num_bytes: 5467832
num_examples: 27893
- name: validation
num_bytes: 840593
num_examples: 4298
- name: test
num_bytes: 1758626
num_examples: 10720
download_size: 819657
dataset_size: 8067051
- config_name: pos_hineng
features:
- name: idx
dtype: int32
- name: words
sequence: string
- name: lid
sequence: string
- name: pos
sequence: string
splits:
- name: train
num_bytes: 537541
num_examples: 1030
- name: validation
num_bytes: 80886
num_examples: 160
- name: test
num_bytes: 131192
num_examples: 299
download_size: 113872
dataset_size: 749619
- config_name: ner_spaeng
features:
- name: idx
dtype: int32
- name: words
sequence: string
- name: lid
sequence: string
- name: ner
sequence: string
splits:
- name: train
num_bytes: 9836312
num_examples: 33611
- name: validation
num_bytes: 2980990
num_examples: 10085
- name: test
num_bytes: 6530956
num_examples: 23527
download_size: 3075520
dataset_size: 19348258
- config_name: ner_msaea
features:
- name: idx
dtype: int32
- name: words
sequence: string
- name: ner
sequence: string
splits:
- name: train
num_bytes: 3887684
num_examples: 10103
- name: validation
num_bytes: 431414
num_examples: 1122
- name: test
num_bytes: 367310
num_examples: 1110
download_size: 938671
dataset_size: 4686408
- config_name: ner_hineng
features:
- name: idx
dtype: int32
- name: words
sequence: string
- name: lid
sequence: string
- name: ner
sequence: string
splits:
- name: train
num_bytes: 474639
num_examples: 1243
- name: validation
num_bytes: 121403
num_examples: 314
- name: test
num_bytes: 185220
num_examples: 522
download_size: 141285
dataset_size: 781262
- config_name: sa_spaeng
features:
- name: idx
dtype: int32
- name: words
sequence: string
- name: lid
sequence: string
- name: sa
dtype: string
splits:
- name: train
num_bytes: 3587783
num_examples: 12194
- name: validation
num_bytes: 546692
num_examples: 1859
- name: test
num_bytes: 1349407
num_examples: 4736
download_size: 1031412
dataset_size: 5483882
---
# Dataset Card for "lince"
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [http://ritual.uh.edu/lince](http://ritual.uh.edu/lince)
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 9.09 MB
- **Size of the generated dataset:** 56.42 MB
- **Total amount of disk used:** 65.52 MB
### Dataset Summary
LinCE is a centralized Linguistic Code-switching Evaluation benchmark
(https://ritual.uh.edu/lince/) that contains data for training and evaluating
NLP systems on code-switching tasks.
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Dataset Structure
### Data Instances
#### lid_hineng
- **Size of downloaded dataset files:** 0.43 MB
- **Size of the generated dataset:** 2.39 MB
- **Total amount of disk used:** 2.82 MB
An example of 'validation' looks as follows.
```
{
"idx": 0,
"lid": ["other", "other", "lang1", "lang1", "lang1", "other", "lang1", "lang1", "lang1", "lang1", "lang1", "lang1", "lang1", "mixed", "lang1", "lang1", "other"],
"words": ["@ZahirJ", "@BinyavangaW", "Loved", "the", "ending", "!", "I", "could", "have", "offered", "you", "some", "ironic", "chai-tea", "for", "it", ";)"]
}
```
#### lid_msaea
- **Size of downloaded dataset files:** 0.81 MB
- **Size of the generated dataset:** 4.89 MB
- **Total amount of disk used:** 5.69 MB
An example of 'train' looks as follows.
```
This example was too long and was cropped:
{
"idx": 0,
"lid": ["ne", "lang2", "other", "lang2", "lang2", "other", "other", "lang2", "lang2", "lang2", "lang2", "lang2", "lang2", "lang2", "lang2", "lang2", "lang2", "lang2", "lang2", "lang2", "lang2", "other", "lang2", "lang2", "lang2", "ne", "lang2", "lang2"],
"words": "[\"علاء\", \"بخير\", \"،\", \"معنوياته\", \"كويسة\", \".\", \"..\", \"اسخف\", \"حاجة\", \"بس\", \"ان\", \"كل\", \"واحد\", \"منهم\", \"بييقى\", \"مقفول\", \"عليه\"..."
}
```
#### lid_nepeng
- **Size of downloaded dataset files:** 0.55 MB
- **Size of the generated dataset:** 3.21 MB
- **Total amount of disk used:** 3.75 MB
An example of 'validation' looks as follows.
```
{
"idx": 1,
"lid": ["other", "lang2", "lang2", "lang2", "lang2", "lang1", "lang1", "lang1", "lang1", "lang1", "lang2", "lang2", "other", "mixed", "lang2", "lang2", "other", "other", "other", "other"],
"words": ["@nirvikdada", "la", "hamlai", "bhetna", "paayeko", "will", "be", "your", "greatest", "gift", "ni", "dada", ";P", "#TreatChaiyo", "j", "hos", ";)", "@zappylily", "@AsthaGhm", "@ayacs_asis"]
}
```
#### lid_spaeng
- **Size of downloaded dataset files:** 1.18 MB
- **Size of the generated dataset:** 6.83 MB
- **Total amount of disk used:** 8.01 MB
An example of 'train' looks as follows.
```
{
"idx": 0,
"lid": ["other", "other", "lang1", "lang1", "lang1", "other", "lang1", "lang1"],
"words": ["11:11", ".....", "make", "a", "wish", ".......", "night", "night"]
}
```
#### ner_hineng
- **Size of downloaded dataset files:** 0.14 MB
- **Size of the generated dataset:** 0.79 MB
- **Total amount of disk used:** 0.92 MB
An example of 'train' looks as follows.
```
{
"idx": 1,
"lid": ["en", "en", "en", "en", "en", "en", "hi", "hi", "hi", "hi", "hi", "hi", "hi", "en", "en", "en", "en", "rest"],
"ner": ["O", "O", "O", "O", "O", "O", "O", "O", "O", "O", "O", "B-PERSON", "I-PERSON", "O", "O", "O", "B-PERSON", "I-PERSON"],
"words": ["I", "liked", "a", "@YouTube", "video", "https://t.co/DmVqhZbdaI", "Kabhi", "Palkon", "Pe", "Aasoon", "Hai-", "Kishore", "Kumar", "-Vocal", "Cover", "By", "Stephen", "Qadir"]
}
```
### Data Fields
The data fields are the same among all splits.
#### lid_hineng
- `idx`: a `int32` feature.
- `words`: a `list` of `string` features.
- `lid`: a `list` of `string` features.
#### lid_msaea
- `idx`: a `int32` feature.
- `words`: a `list` of `string` features.
- `lid`: a `list` of `string` features.
#### lid_nepeng
- `idx`: a `int32` feature.
- `words`: a `list` of `string` features.
- `lid`: a `list` of `string` features.
#### lid_spaeng
- `idx`: a `int32` feature.
- `words`: a `list` of `string` features.
- `lid`: a `list` of `string` features.
#### ner_hineng
- `idx`: a `int32` feature.
- `words`: a `list` of `string` features.
- `lid`: a `list` of `string` features.
- `ner`: a `list` of `string` features.
### Data Splits
| name |train|validation|test|
|----------|----:|---------:|---:|
|lid_hineng| 4823| 744|1854|
|lid_msaea | 8464| 1116|1663|
|lid_nepeng| 8451| 1332|3228|
|lid_spaeng|21030| 3332|8289|
|ner_hineng| 1243| 314| 522|
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Citation Information
```
@inproceedings{aguilar-etal-2020-lince,
title = "{L}in{CE}: A Centralized Benchmark for Linguistic Code-switching Evaluation",
author = "Aguilar, Gustavo and
Kar, Sudipta and
Solorio, Thamar",
booktitle = "Proceedings of The 12th Language Resources and Evaluation Conference",
month = may,
year = "2020",
address = "Marseille, France",
publisher = "European Language Resources Association",
url = "https://www.aclweb.org/anthology/2020.lrec-1.223",
pages = "1803--1813",
language = "English",
ISBN = "979-10-95546-34-4",
}
```
Note that each LinCE dataset has its own citation too. Please see [here](https://ritual.uh.edu/lince/datasets)
for the correct citation on each dataset.
### Contributions
Thanks to [@lhoestq](https://github.com/lhoestq), [@thomwolf](https://github.com/thomwolf), [@gaguilar](https://github.com/gaguilar) for adding this dataset.