sha
stringlengths
40
40
text
stringlengths
1
13.4M
id
stringlengths
2
117
tags
sequencelengths
1
7.91k
created_at
stringlengths
25
25
metadata
stringlengths
2
875k
last_modified
stringlengths
25
25
arxiv
sequencelengths
0
25
languages
sequencelengths
0
7.91k
tags_str
stringlengths
17
159k
text_str
stringlengths
1
447k
text_lists
sequencelengths
0
352
processed_texts
sequencelengths
1
353
tokens_length
sequencelengths
1
353
input_texts
sequencelengths
1
40
e9ee727feeeb20d3a005f9c677abf5311323f1b6
# Dataset Card for TabFact ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [TabFact](https://tabfact.github.io/index.html) - **Repository:** [GitHub](https://github.com/wenhuchen/Table-Fact-Checking) - **Paper:** [TabFact: A Large-scale Dataset for Table-based Fact Verification](https://arxiv.org/abs/1909.02164) - **Leaderboard:** [Leaderboard](https://competitions.codalab.org/competitions/21611) - **Point of Contact:** [Wenhu Chen](wenhuchen@cs.ucsb.edu) ### Dataset Summary The problem of verifying whether a textual hypothesis holds the truth based on the given evidence, also known as fact verification, plays an important role in the study of natural language understanding and semantic representation. However, existing studies are restricted to dealing with unstructured textual evidence (e.g., sentences and passages, a pool of passages), while verification using structured forms of evidence, such as tables, graphs, and databases, remains unexplored. TABFACT is large scale dataset with 16k Wikipedia tables as evidence for 118k human annotated statements designed for fact verification with semi-structured evidence. The statements are labeled as either ENTAILED or REFUTED. TABFACT is challenging since it involves both soft linguistic reasoning and hard symbolic reasoning. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ``` @inproceedings{2019TabFactA, title={TabFact : A Large-scale Dataset for Table-based Fact Verification}, author={Wenhu Chen, Hongmin Wang, Jianshu Chen, Yunkai Zhang, Hong Wang, Shiyang Li, Xiyou Zhou and William Yang Wang}, booktitle = {International Conference on Learning Representations (ICLR)}, address = {Addis Ababa, Ethiopia}, month = {April}, year = {2020} } ``` ### Contributions Thanks to [@patil-suraj](https://github.com/patil-suraj) for adding this dataset.
tab_fact
[ "task_categories:text-classification", "task_ids:fact-checking", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:en", "license:cc-by-4.0", "arxiv:1909.02164", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["fact-checking"], "paperswithcode_id": "tabfact", "pretty_name": "TabFact", "dataset_info": [{"config_name": "tab_fact", "features": [{"name": "id", "dtype": "int32"}, {"name": "table_id", "dtype": "string"}, {"name": "table_text", "dtype": "string"}, {"name": "table_caption", "dtype": "string"}, {"name": "statement", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "refuted", "1": "entailed"}}}}], "splits": [{"name": "train", "num_bytes": 99852664, "num_examples": 92283}, {"name": "validation", "num_bytes": 13846872, "num_examples": 12792}, {"name": "test", "num_bytes": 13493391, "num_examples": 12779}], "download_size": 196508436, "dataset_size": 127192927}, {"config_name": "blind_test", "features": [{"name": "id", "dtype": "int32"}, {"name": "table_id", "dtype": "string"}, {"name": "table_text", "dtype": "string"}, {"name": "table_caption", "dtype": "string"}, {"name": "statement", "dtype": "string"}, {"name": "test_id", "dtype": "string"}], "splits": [{"name": "test", "num_bytes": 10954442, "num_examples": 9750}], "download_size": 196508436, "dataset_size": 10954442}]}
2024-01-18T11:16:41+00:00
[ "1909.02164" ]
[ "en" ]
TAGS #task_categories-text-classification #task_ids-fact-checking #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-4.0 #arxiv-1909.02164 #region-us
# Dataset Card for TabFact ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: TabFact - Repository: GitHub - Paper: TabFact: A Large-scale Dataset for Table-based Fact Verification - Leaderboard: Leaderboard - Point of Contact: Wenhu Chen ### Dataset Summary The problem of verifying whether a textual hypothesis holds the truth based on the given evidence, also known as fact verification, plays an important role in the study of natural language understanding and semantic representation. However, existing studies are restricted to dealing with unstructured textual evidence (e.g., sentences and passages, a pool of passages), while verification using structured forms of evidence, such as tables, graphs, and databases, remains unexplored. TABFACT is large scale dataset with 16k Wikipedia tables as evidence for 118k human annotated statements designed for fact verification with semi-structured evidence. The statements are labeled as either ENTAILED or REFUTED. TABFACT is challenging since it involves both soft linguistic reasoning and hard symbolic reasoning. ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @patil-suraj for adding this dataset.
[ "# Dataset Card for TabFact", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: TabFact\n- Repository: GitHub\n- Paper: TabFact: A Large-scale Dataset for Table-based Fact Verification\n- Leaderboard: Leaderboard\n- Point of Contact: Wenhu Chen", "### Dataset Summary\n\nThe problem of verifying whether a textual hypothesis holds the truth based on the given evidence, also known as fact verification, plays an important role in the study of natural language understanding and semantic representation. However, existing studies are restricted to dealing with unstructured textual evidence (e.g., sentences and passages, a pool of passages), while verification using structured forms of evidence, such as tables, graphs, and databases, remains unexplored. TABFACT is large scale dataset with 16k Wikipedia tables as evidence for 118k human annotated statements designed for fact verification with semi-structured evidence. The statements are labeled as either ENTAILED or REFUTED. TABFACT is challenging since it involves both soft linguistic reasoning and hard symbolic reasoning.", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @patil-suraj for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #task_ids-fact-checking #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-4.0 #arxiv-1909.02164 #region-us \n", "# Dataset Card for TabFact", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: TabFact\n- Repository: GitHub\n- Paper: TabFact: A Large-scale Dataset for Table-based Fact Verification\n- Leaderboard: Leaderboard\n- Point of Contact: Wenhu Chen", "### Dataset Summary\n\nThe problem of verifying whether a textual hypothesis holds the truth based on the given evidence, also known as fact verification, plays an important role in the study of natural language understanding and semantic representation. However, existing studies are restricted to dealing with unstructured textual evidence (e.g., sentences and passages, a pool of passages), while verification using structured forms of evidence, such as tables, graphs, and databases, remains unexplored. TABFACT is large scale dataset with 16k Wikipedia tables as evidence for 118k human annotated statements designed for fact verification with semi-structured evidence. The statements are labeled as either ENTAILED or REFUTED. TABFACT is challenging since it involves both soft linguistic reasoning and hard symbolic reasoning.", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @patil-suraj for adding this dataset." ]
[ 100, 8, 120, 53, 191, 10, 4, 6, 6, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 19 ]
[ "passage: TAGS\n#task_categories-text-classification #task_ids-fact-checking #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-4.0 #arxiv-1909.02164 #region-us \n# Dataset Card for TabFact## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: TabFact\n- Repository: GitHub\n- Paper: TabFact: A Large-scale Dataset for Table-based Fact Verification\n- Leaderboard: Leaderboard\n- Point of Contact: Wenhu Chen### Dataset Summary\n\nThe problem of verifying whether a textual hypothesis holds the truth based on the given evidence, also known as fact verification, plays an important role in the study of natural language understanding and semantic representation. However, existing studies are restricted to dealing with unstructured textual evidence (e.g., sentences and passages, a pool of passages), while verification using structured forms of evidence, such as tables, graphs, and databases, remains unexplored. TABFACT is large scale dataset with 16k Wikipedia tables as evidence for 118k human annotated statements designed for fact verification with semi-structured evidence. The statements are labeled as either ENTAILED or REFUTED. TABFACT is challenging since it involves both soft linguistic reasoning and hard symbolic reasoning.### Supported Tasks and Leaderboards### Languages## Dataset Structure### Data Instances### Data Fields" ]
2f417fcf209e7c5fbe223ebf2a49be5822680391
# Dataset Card for Tamilmixsentiment ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Tamilmixsentiment Homepage](https://dravidian-codemix.github.io/2020/index.html) - **Repository:** [Tamilmixsentiment repository](https://dravidian-codemix.github.io/2020/datasets.html) - **Paper:** [Corpus Creation for Sentiment Analysis in Code-Mixed Tamil-English Text](https://www.aclweb.org/anthology/2020.sltu-1.28/) - **Leaderboard:** [Rank list](https://drive.google.com/file/d/1Mf8-No-63koGRwdF13RrO01NAFBlNmI0/view?usp=sharing) - **Point of Contact:** [Bharathi Raja Chakravarthi](mailto:bharathiraja.akr@gmail.com) ### Dataset Summary The first gold standard Tamil-English code-switched, sentiment-annotated corpus containing 15,744 comment posts from YouTube. This makes the largest general domain sentiment dataset for this relatively low-resource language with code-mixing phenomenon. The comment/post may contain more than one sentence but the average sentence length of the corpora is 1. Each comment/post is annotated with sentiment polarity at the comment/post level. This dataset also has class imbalance problems depicting real-world scenarios. ### Supported Tasks and Leaderboards To identify sentiment polarity of the code-mixed dataset of comments/posts in Tamil-English collected from social media. ### Languages Tamil-English code-switched. The dataset contains all the three types of code-mixed sentences - Inter-Sentential switch, Intra-Sentential switch and Tag switching. Most comments were written in Roman script with either Tamil grammar with English lexicon or English grammar with Tamil lexicon. Some comments were written in Tamil script with English expressions in between. ## Dataset Structure ### Data Instances An example from the Tamilmixsentiment train set looks as follows: ``` text label Trailer late ah parthavanga like podunga Positive ``` ### Data Fields - `text`: Tamil-English code-mixed comment. - `label`: list of the possible sentiments "Positive", "Negative", "Mixed_feelings", "unknown_state", "not-Tamil" ### Data Splits The entire dataset of 15,744 sentences was randomly shuffled and split into three parts as follows: | | train | validation | test | |------------------------------|------:|-----------:|-----:| | Tamilmixsentiment | 11335 | 1260 | 3149 | ## Dataset Creation ### Curation Rationale Sentiment analysis has become important in social media research (Yang and Eisenstein, 2017). Until recently these applications were created for high-resourced languages which analysed monolingual utterances. But social media in multilingual communities contains more code-mixed text. Code-mixing is common among speakers in a bilingual speech community. As English is seen as the language of prestige and education, the influence of lexicon, connectives and phrases from English language is common in spoken Tamil. Tamil has little annotated data for code-mixed scenarios. An annotated corpus developed for monolingual data cannot deal with code-mixed usage and therefore it fails to yield good results due to mixture of languages at different levels of linguistic analysis. Therefore this dataset of code-mixed Tamil-English sentiment annotated corpus is created. ### Source Data #### Initial Data Collection and Normalization The data was scraped from Youtube. In total 184,573 sentences for Tamil from YouTube comments from the trailers of a movies released in 2019. Many of the them contained sentences that were either entirely written in English or code-mixed Tamil-English or fully written in Tamil. So we filtered out a non-code-mixed corpus based on language identification at comment level using the langdetect library. The comment is written fully in Tamil or English, we discarded that comment since monolingual resources are available for these languages. We also identified if the sentences were written in other languages such as Hindi, Malayalam, Urdu, Telugu, and Kannada. We preprocessed the comments by removing the emoticons and applying a sentence length filter. We want to create a code-mixed corpus of reasonable size with sentences that have fairly defined sentiments which will be useful for future research. Thus our filter removed sentences with less than five words and more than 15 words after cleaning the data. In the end we got 15,744 Tanglish sentences. #### Who are the source language producers? Youtube users ### Annotations #### Annotation process Three steps complete the annotation setup. First, each sentence was annotated by two people. In the second step, the data were collected if both of them agreed. In the case of conflict, a third person annotated the sentence. In the third step, if all the three of them did not agree, then two more annotators annotated the sentences. #### Who are the annotators? Eleven volunteers were involved in the process. All of them were native speakers of Tamil with diversity in gender, educational level and medium of instruction in their school education. ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ``` @inproceedings{chakravarthi-etal-2020-corpus, title = "Corpus Creation for Sentiment Analysis in Code-Mixed {T}amil-{E}nglish Text", author = "Chakravarthi, Bharathi Raja and Muralidaran, Vigneshwaran and Priyadharshini, Ruba and McCrae, John Philip", booktitle = "Proceedings of the 1st Joint Workshop on Spoken Language Technologies for Under-resourced languages (SLTU) and Collaboration and Computing for Under-Resourced Languages (CCURL)", month = may, year = "2020", address = "Marseille, France", publisher = "European Language Resources association", url = "https://www.aclweb.org/anthology/2020.sltu-1.28", pages = "202--210", abstract = "Understanding the sentiment of a comment from a video or an image is an essential task in many applications. Sentiment analysis of a text can be useful for various decision-making processes. One such application is to analyse the popular sentiments of videos on social media based on viewer comments. However, comments from social media do not follow strict rules of grammar, and they contain mixing of more than one language, often written in non-native scripts. Non-availability of annotated code-mixed data for a low-resourced language like Tamil also adds difficulty to this problem. To overcome this, we created a gold standard Tamil-English code-switched, sentiment-annotated corpus containing 15,744 comment posts from YouTube. In this paper, we describe the process of creating the corpus and assigning polarities. We present inter-annotator agreement and show the results of sentiment analysis trained on this corpus as a benchmark.", language = "English", ISBN = "979-10-95546-35-1", } ``` ### Contributions Thanks to [@jamespaultg](https://github.com/jamespaultg) for adding this dataset.
tamilmixsentiment
[ "task_categories:text-classification", "task_ids:sentiment-classification", "annotations_creators:expert-generated", "language_creators:crowdsourced", "multilinguality:multilingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "language:ta", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["crowdsourced"], "language": ["en", "ta"], "license": ["unknown"], "multilinguality": ["multilingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["sentiment-classification"], "pretty_name": "Tamilmixsentiment", "dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "Positive", "1": "Negative", "2": "Mixed_feelings", "3": "unknown_state", "4": "not-Tamil"}}}}], "splits": [{"name": "train", "num_bytes": 790132, "num_examples": 11335}, {"name": "validation", "num_bytes": 89618, "num_examples": 1260}, {"name": "test", "num_bytes": 218764, "num_examples": 3149}], "download_size": 1150792, "dataset_size": 1098514}}
2023-06-16T12:07:45+00:00
[]
[ "en", "ta" ]
TAGS #task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #language_creators-crowdsourced #multilinguality-multilingual #size_categories-10K<n<100K #source_datasets-original #language-English #language-Tamil #license-unknown #region-us
Dataset Card for Tamilmixsentiment ================================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: Tamilmixsentiment Homepage * Repository: Tamilmixsentiment repository * Paper: Corpus Creation for Sentiment Analysis in Code-Mixed Tamil-English Text * Leaderboard: Rank list * Point of Contact: Bharathi Raja Chakravarthi ### Dataset Summary The first gold standard Tamil-English code-switched, sentiment-annotated corpus containing 15,744 comment posts from YouTube. This makes the largest general domain sentiment dataset for this relatively low-resource language with code-mixing phenomenon. The comment/post may contain more than one sentence but the average sentence length of the corpora is 1. Each comment/post is annotated with sentiment polarity at the comment/post level. This dataset also has class imbalance problems depicting real-world scenarios. ### Supported Tasks and Leaderboards To identify sentiment polarity of the code-mixed dataset of comments/posts in Tamil-English collected from social media. ### Languages Tamil-English code-switched. The dataset contains all the three types of code-mixed sentences - Inter-Sentential switch, Intra-Sentential switch and Tag switching. Most comments were written in Roman script with either Tamil grammar with English lexicon or English grammar with Tamil lexicon. Some comments were written in Tamil script with English expressions in between. Dataset Structure ----------------- ### Data Instances An example from the Tamilmixsentiment train set looks as follows: ### Data Fields * 'text': Tamil-English code-mixed comment. * 'label': list of the possible sentiments "Positive", "Negative", "Mixed\_feelings", "unknown\_state", "not-Tamil" ### Data Splits The entire dataset of 15,744 sentences was randomly shuffled and split into three parts as follows: Dataset Creation ---------------- ### Curation Rationale Sentiment analysis has become important in social media research (Yang and Eisenstein, 2017). Until recently these applications were created for high-resourced languages which analysed monolingual utterances. But social media in multilingual communities contains more code-mixed text. Code-mixing is common among speakers in a bilingual speech community. As English is seen as the language of prestige and education, the influence of lexicon, connectives and phrases from English language is common in spoken Tamil. Tamil has little annotated data for code-mixed scenarios. An annotated corpus developed for monolingual data cannot deal with code-mixed usage and therefore it fails to yield good results due to mixture of languages at different levels of linguistic analysis. Therefore this dataset of code-mixed Tamil-English sentiment annotated corpus is created. ### Source Data #### Initial Data Collection and Normalization The data was scraped from Youtube. In total 184,573 sentences for Tamil from YouTube comments from the trailers of a movies released in 2019. Many of the them contained sentences that were either entirely written in English or code-mixed Tamil-English or fully written in Tamil. So we filtered out a non-code-mixed corpus based on language identification at comment level using the langdetect library. The comment is written fully in Tamil or English, we discarded that comment since monolingual resources are available for these languages. We also identified if the sentences were written in other languages such as Hindi, Malayalam, Urdu, Telugu, and Kannada. We preprocessed the comments by removing the emoticons and applying a sentence length filter. We want to create a code-mixed corpus of reasonable size with sentences that have fairly defined sentiments which will be useful for future research. Thus our filter removed sentences with less than five words and more than 15 words after cleaning the data. In the end we got 15,744 Tanglish sentences. #### Who are the source language producers? Youtube users ### Annotations #### Annotation process Three steps complete the annotation setup. First, each sentence was annotated by two people. In the second step, the data were collected if both of them agreed. In the case of conflict, a third person annotated the sentence. In the third step, if all the three of them did not agree, then two more annotators annotated the sentences. #### Who are the annotators? Eleven volunteers were involved in the process. All of them were native speakers of Tamil with diversity in gender, educational level and medium of instruction in their school education. ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @jamespaultg for adding this dataset.
[ "### Dataset Summary\n\n\nThe first gold standard Tamil-English code-switched, sentiment-annotated corpus containing 15,744 comment posts from YouTube. This makes the largest general domain sentiment dataset for this relatively low-resource language with code-mixing phenomenon. The comment/post may contain more than one sentence but the average sentence length of the corpora is 1. Each comment/post is annotated with sentiment polarity at the comment/post level. This dataset also has class imbalance problems depicting real-world scenarios.", "### Supported Tasks and Leaderboards\n\n\nTo identify sentiment polarity of the code-mixed dataset of comments/posts in Tamil-English collected from social media.", "### Languages\n\n\nTamil-English code-switched. The dataset contains all the three types of code-mixed sentences - Inter-Sentential switch, Intra-Sentential switch and Tag switching. Most comments were written in Roman script with either Tamil grammar with English lexicon or English grammar with Tamil lexicon. Some comments were written in Tamil script with English expressions in between.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nAn example from the Tamilmixsentiment train set looks as follows:", "### Data Fields\n\n\n* 'text': Tamil-English code-mixed comment.\n* 'label': list of the possible sentiments \"Positive\", \"Negative\", \"Mixed\\_feelings\", \"unknown\\_state\", \"not-Tamil\"", "### Data Splits\n\n\nThe entire dataset of 15,744 sentences was randomly shuffled and split into three parts as follows:\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nSentiment analysis has become important in social media research (Yang and Eisenstein, 2017). Until recently these applications were created for high-resourced languages which analysed monolingual utterances. But social media in multilingual communities contains more code-mixed text. Code-mixing is common among speakers in a bilingual speech community. As English is seen as the language of prestige and education, the influence of lexicon, connectives and phrases from English language is common in spoken Tamil. Tamil has little annotated data for code-mixed scenarios. An annotated corpus developed for monolingual data cannot deal with code-mixed usage and therefore it fails to yield good results due to mixture of languages at different levels of linguistic analysis. Therefore this dataset of code-mixed Tamil-English sentiment annotated corpus is created.", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nThe data was scraped from Youtube. In total 184,573 sentences for Tamil from YouTube comments from the trailers of a movies released in 2019. Many of the them contained sentences\nthat were either entirely written in English or code-mixed Tamil-English or fully written in Tamil. So we filtered out a non-code-mixed corpus based on language identification\nat comment level using the langdetect library. The comment is written fully in Tamil or English, we discarded that comment since monolingual resources are available for these languages. We also identified if the sentences were written in other languages such as Hindi, Malayalam, Urdu, Telugu, and Kannada. We preprocessed the comments by removing the emoticons and applying a sentence\nlength filter. We want to create a code-mixed corpus of reasonable size with sentences that have fairly defined sentiments which will be useful for future research. Thus our filter removed sentences with less than five words and more than 15 words after cleaning the data. In the end we got 15,744 Tanglish sentences.", "#### Who are the source language producers?\n\n\nYoutube users", "### Annotations", "#### Annotation process\n\n\nThree steps complete the annotation setup. First, each sentence was annotated by two people. In the second step, the data were collected if both of them agreed. In the case of conflict, a third person annotated the sentence. In the third step, if all the three of them did not agree, then two more annotators annotated the sentences.", "#### Who are the annotators?\n\n\nEleven volunteers were involved in the process. All of them were native speakers of Tamil with diversity in gender, educational level and medium of instruction in their school education.", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @jamespaultg for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #language_creators-crowdsourced #multilinguality-multilingual #size_categories-10K<n<100K #source_datasets-original #language-English #language-Tamil #license-unknown #region-us \n", "### Dataset Summary\n\n\nThe first gold standard Tamil-English code-switched, sentiment-annotated corpus containing 15,744 comment posts from YouTube. This makes the largest general domain sentiment dataset for this relatively low-resource language with code-mixing phenomenon. The comment/post may contain more than one sentence but the average sentence length of the corpora is 1. Each comment/post is annotated with sentiment polarity at the comment/post level. This dataset also has class imbalance problems depicting real-world scenarios.", "### Supported Tasks and Leaderboards\n\n\nTo identify sentiment polarity of the code-mixed dataset of comments/posts in Tamil-English collected from social media.", "### Languages\n\n\nTamil-English code-switched. The dataset contains all the three types of code-mixed sentences - Inter-Sentential switch, Intra-Sentential switch and Tag switching. Most comments were written in Roman script with either Tamil grammar with English lexicon or English grammar with Tamil lexicon. Some comments were written in Tamil script with English expressions in between.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nAn example from the Tamilmixsentiment train set looks as follows:", "### Data Fields\n\n\n* 'text': Tamil-English code-mixed comment.\n* 'label': list of the possible sentiments \"Positive\", \"Negative\", \"Mixed\\_feelings\", \"unknown\\_state\", \"not-Tamil\"", "### Data Splits\n\n\nThe entire dataset of 15,744 sentences was randomly shuffled and split into three parts as follows:\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nSentiment analysis has become important in social media research (Yang and Eisenstein, 2017). Until recently these applications were created for high-resourced languages which analysed monolingual utterances. But social media in multilingual communities contains more code-mixed text. Code-mixing is common among speakers in a bilingual speech community. As English is seen as the language of prestige and education, the influence of lexicon, connectives and phrases from English language is common in spoken Tamil. Tamil has little annotated data for code-mixed scenarios. An annotated corpus developed for monolingual data cannot deal with code-mixed usage and therefore it fails to yield good results due to mixture of languages at different levels of linguistic analysis. Therefore this dataset of code-mixed Tamil-English sentiment annotated corpus is created.", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nThe data was scraped from Youtube. In total 184,573 sentences for Tamil from YouTube comments from the trailers of a movies released in 2019. Many of the them contained sentences\nthat were either entirely written in English or code-mixed Tamil-English or fully written in Tamil. So we filtered out a non-code-mixed corpus based on language identification\nat comment level using the langdetect library. The comment is written fully in Tamil or English, we discarded that comment since monolingual resources are available for these languages. We also identified if the sentences were written in other languages such as Hindi, Malayalam, Urdu, Telugu, and Kannada. We preprocessed the comments by removing the emoticons and applying a sentence\nlength filter. We want to create a code-mixed corpus of reasonable size with sentences that have fairly defined sentiments which will be useful for future research. Thus our filter removed sentences with less than five words and more than 15 words after cleaning the data. In the end we got 15,744 Tanglish sentences.", "#### Who are the source language producers?\n\n\nYoutube users", "### Annotations", "#### Annotation process\n\n\nThree steps complete the annotation setup. First, each sentence was annotated by two people. In the second step, the data were collected if both of them agreed. In the case of conflict, a third person annotated the sentence. In the third step, if all the three of them did not agree, then two more annotators annotated the sentences.", "#### Who are the annotators?\n\n\nEleven volunteers were involved in the process. All of them were native speakers of Tamil with diversity in gender, educational level and medium of instruction in their school education.", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @jamespaultg for adding this dataset." ]
[ 95, 121, 39, 96, 21, 63, 35, 198, 4, 239, 12, 5, 84, 47, 18, 7, 8, 14, 6, 6, 18 ]
[ "passage: TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #language_creators-crowdsourced #multilinguality-multilingual #size_categories-10K<n<100K #source_datasets-original #language-English #language-Tamil #license-unknown #region-us \n### Dataset Summary\n\n\nThe first gold standard Tamil-English code-switched, sentiment-annotated corpus containing 15,744 comment posts from YouTube. This makes the largest general domain sentiment dataset for this relatively low-resource language with code-mixing phenomenon. The comment/post may contain more than one sentence but the average sentence length of the corpora is 1. Each comment/post is annotated with sentiment polarity at the comment/post level. This dataset also has class imbalance problems depicting real-world scenarios.### Supported Tasks and Leaderboards\n\n\nTo identify sentiment polarity of the code-mixed dataset of comments/posts in Tamil-English collected from social media.### Languages\n\n\nTamil-English code-switched. The dataset contains all the three types of code-mixed sentences - Inter-Sentential switch, Intra-Sentential switch and Tag switching. Most comments were written in Roman script with either Tamil grammar with English lexicon or English grammar with Tamil lexicon. Some comments were written in Tamil script with English expressions in between.\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nAn example from the Tamilmixsentiment train set looks as follows:### Data Fields\n\n\n* 'text': Tamil-English code-mixed comment.\n* 'label': list of the possible sentiments \"Positive\", \"Negative\", \"Mixed\\_feelings\", \"unknown\\_state\", \"not-Tamil\"### Data Splits\n\n\nThe entire dataset of 15,744 sentences was randomly shuffled and split into three parts as follows:\n\n\n\nDataset Creation\n----------------" ]
9f1a9abb59ca202c77241cecd41055cd4b04d9c4
# Dataset Card for tanzil ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** http://opus.nlpl.eu/Tanzil.php - **Repository:** None - **Paper:** http://www.lrec-conf.org/proceedings/lrec2012/pdf/463_Paper.pdf - **Leaderboard:** [More Information Needed] - **Point of Contact:** [More Information Needed] ### Dataset Summary To load a language pair which isn't part of the config, all you need to do is specify the language code as pairs. You can find the valid pairs in Homepage section of Dataset Description: http://opus.nlpl.eu/Tanzil.php E.g. `dataset = load_dataset("tanzil", lang1="en", lang2="ru")` ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances Here are some examples of questions and facts: ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@abhishekkrthakur](https://github.com/abhishekkrthakur) for adding this dataset.
tanzil
[ "task_categories:translation", "annotations_creators:found", "language_creators:found", "multilinguality:multilingual", "size_categories:100K<n<1M", "source_datasets:original", "language:am", "language:ar", "language:az", "language:bg", "language:bn", "language:bs", "language:cs", "language:de", "language:dv", "language:en", "language:es", "language:fa", "language:fr", "language:ha", "language:hi", "language:id", "language:it", "language:ja", "language:ko", "language:ku", "language:ml", "language:ms", "language:nl", "language:no", "language:pl", "language:pt", "language:ro", "language:ru", "language:sd", "language:so", "language:sq", "language:sv", "language:sw", "language:ta", "language:tg", "language:th", "language:tr", "language:tt", "language:ug", "language:ur", "language:uz", "language:zh", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["am", "ar", "az", "bg", "bn", "bs", "cs", "de", "dv", "en", "es", "fa", "fr", "ha", "hi", "id", "it", "ja", "ko", "ku", "ml", "ms", "nl", "no", "pl", "pt", "ro", "ru", "sd", "so", "sq", "sv", "sw", "ta", "tg", "th", "tr", "tt", "ug", "ur", "uz", "zh"], "license": ["unknown"], "multilinguality": ["multilingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["translation"], "task_ids": [], "pretty_name": "tanzil", "dataset_info": [{"config_name": "bg-en", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["bg", "en"]}}}], "splits": [{"name": "train", "num_bytes": 34473016, "num_examples": 135477}], "download_size": 9305292, "dataset_size": 34473016}, {"config_name": "bn-hi", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["bn", "hi"]}}}], "splits": [{"name": "train", "num_bytes": 18869103, "num_examples": 24942}], "download_size": 3542740, "dataset_size": 18869103}, {"config_name": "fa-sv", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["fa", "sv"]}}}], "splits": [{"name": "train", "num_bytes": 29281634, "num_examples": 68601}], "download_size": 8550826, "dataset_size": 29281634}, {"config_name": "ru-zh", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["ru", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 59736143, "num_examples": 99779}], "download_size": 16214659, "dataset_size": 59736143}, {"config_name": "en-tr", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "tr"]}}}], "splits": [{"name": "train", "num_bytes": 255891913, "num_examples": 1189967}], "download_size": 82954694, "dataset_size": 255891913}]}
2024-01-18T11:16:42+00:00
[]
[ "am", "ar", "az", "bg", "bn", "bs", "cs", "de", "dv", "en", "es", "fa", "fr", "ha", "hi", "id", "it", "ja", "ko", "ku", "ml", "ms", "nl", "no", "pl", "pt", "ro", "ru", "sd", "so", "sq", "sv", "sw", "ta", "tg", "th", "tr", "tt", "ug", "ur", "uz", "zh" ]
TAGS #task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-100K<n<1M #source_datasets-original #language-Amharic #language-Arabic #language-Azerbaijani #language-Bulgarian #language-Bengali #language-Bosnian #language-Czech #language-German #language-Dhivehi #language-English #language-Spanish #language-Persian #language-French #language-Hausa #language-Hindi #language-Indonesian #language-Italian #language-Japanese #language-Korean #language-Kurdish #language-Malayalam #language-Malay (macrolanguage) #language-Dutch #language-Norwegian #language-Polish #language-Portuguese #language-Romanian #language-Russian #language-Sindhi #language-Somali #language-Albanian #language-Swedish #language-Swahili (macrolanguage) #language-Tamil #language-Tajik #language-Thai #language-Turkish #language-Tatar #language-Uighur #language-Urdu #language-Uzbek #language-Chinese #license-unknown #region-us
# Dataset Card for tanzil ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: None - Paper: URL - Leaderboard: - Point of Contact: ### Dataset Summary To load a language pair which isn't part of the config, all you need to do is specify the language code as pairs. You can find the valid pairs in Homepage section of Dataset Description: URL E.g. 'dataset = load_dataset("tanzil", lang1="en", lang2="ru")' ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances Here are some examples of questions and facts: ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @abhishekkrthakur for adding this dataset.
[ "# Dataset Card for tanzil", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: URL\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\n\nTo load a language pair which isn't part of the config, all you need to do is specify the language code as pairs.\nYou can find the valid pairs in Homepage section of Dataset Description: URL\nE.g.\n\n'dataset = load_dataset(\"tanzil\", lang1=\"en\", lang2=\"ru\")'", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances\n\nHere are some examples of questions and facts:", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ "TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-100K<n<1M #source_datasets-original #language-Amharic #language-Arabic #language-Azerbaijani #language-Bulgarian #language-Bengali #language-Bosnian #language-Czech #language-German #language-Dhivehi #language-English #language-Spanish #language-Persian #language-French #language-Hausa #language-Hindi #language-Indonesian #language-Italian #language-Japanese #language-Korean #language-Kurdish #language-Malayalam #language-Malay (macrolanguage) #language-Dutch #language-Norwegian #language-Polish #language-Portuguese #language-Romanian #language-Russian #language-Sindhi #language-Somali #language-Albanian #language-Swedish #language-Swahili (macrolanguage) #language-Tamil #language-Tajik #language-Thai #language-Turkish #language-Tatar #language-Uighur #language-Urdu #language-Uzbek #language-Chinese #license-unknown #region-us \n", "# Dataset Card for tanzil", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: URL\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\n\nTo load a language pair which isn't part of the config, all you need to do is specify the language code as pairs.\nYou can find the valid pairs in Homepage section of Dataset Description: URL\nE.g.\n\n'dataset = load_dataset(\"tanzil\", lang1=\"en\", lang2=\"ru\")'", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances\n\nHere are some examples of questions and facts:", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ 303, 7, 120, 28, 80, 10, 4, 6, 17, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 20 ]
[ "passage: TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-100K<n<1M #source_datasets-original #language-Amharic #language-Arabic #language-Azerbaijani #language-Bulgarian #language-Bengali #language-Bosnian #language-Czech #language-German #language-Dhivehi #language-English #language-Spanish #language-Persian #language-French #language-Hausa #language-Hindi #language-Indonesian #language-Italian #language-Japanese #language-Korean #language-Kurdish #language-Malayalam #language-Malay (macrolanguage) #language-Dutch #language-Norwegian #language-Polish #language-Portuguese #language-Romanian #language-Russian #language-Sindhi #language-Somali #language-Albanian #language-Swedish #language-Swahili (macrolanguage) #language-Tamil #language-Tajik #language-Thai #language-Turkish #language-Tatar #language-Uighur #language-Urdu #language-Uzbek #language-Chinese #license-unknown #region-us \n# Dataset Card for tanzil## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: URL\n- Leaderboard: \n- Point of Contact:" ]
58b71ecd5016350847357ad7de52bf19b5d4aa6f
# Dataset Card for TaPaCo Corpus ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [TaPaCo: A Corpus of Sentential Paraphrases for 73 Languages](https://zenodo.org/record/3707949#.X9Dh0cYza3I) - **Paper:** [TaPaCo: A Corpus of Sentential Paraphrases for 73 Languages](https://www.aclweb.org/anthology/2020.lrec-1.848.pdf) - **Data:** https://doi.org/10.5281/zenodo.3707949 - **Point of Contact:** [Yves Scherrer](https://blogs.helsinki.fi/yvesscherrer/) ### Dataset Summary A freely available paraphrase corpus for 73 languages extracted from the Tatoeba database. Tatoeba is a crowdsourcing project mainly geared towards language learners. Its aim is to provide example sentences and translations for particular linguistic constructions and words. The paraphrase corpus is created by populating a graph with Tatoeba sentences and equivalence links between sentences “meaning the same thing”. This graph is then traversed to extract sets of paraphrases. Several language-independent filters and pruning steps are applied to remove uninteresting sentences. A manual evaluation performed on three languages shows that between half and three quarters of inferred paraphrases are correct and that most remaining ones are either correct but trivial, or near-paraphrases that neutralize a morphological distinction. The corpus contains a total of 1.9 million sentences, with 200 – 250 000 sentences per language. It covers a range of languages for which, to our knowledge, no other paraphrase dataset exists. ### Supported Tasks and Leaderboards Paraphrase detection and generation have become popular tasks in NLP and are increasingly integrated into a wide variety of common downstream tasks such as machine translation , information retrieval, question answering, and semantic parsing. Most of the existing datasets cover only a single language – in most cases English – or a small number of languages. Furthermore, some paraphrase datasets focus on lexical and phrasal rather than sentential paraphrases, while others are created (semi -)automatically using machine translation. The number of sentences per language ranges from 200 to 250 000, which makes the dataset more suitable for fine-tuning and evaluation purposes than for training. It is well-suited for multi-reference evaluation of paraphrase generation models, as there is generally not a single correct way of paraphrasing a given input sentence. ### Languages The dataset contains paraphrases in Afrikaans, Arabic, Azerbaijani, Belarusian, Berber languages, Bulgarian, Bengali , Breton, Catalan; Valencian, Chavacano, Mandarin, Czech, Danish, German, Greek, Modern (1453-), English, Esperanto , Spanish; Castilian, Estonian, Basque, Finnish, French, Galician, Gronings, Hebrew, Hindi, Croatian, Hungarian , Armenian, Interlingua (International Auxiliary Language Association), Indonesian, Interlingue; Occidental, Ido , Icelandic, Italian, Japanese, Lojban, Kabyle, Korean, Cornish, Latin, Lingua Franca Nova\t, Lithuanian, Macedonian , Marathi, Bokmål, Norwegian; Norwegian Bokmål, Low German; Low Saxon; German, Low; Saxon, Low, Dutch; Flemish, ]Old Russian, Turkish, Ottoman (1500-1928), Iranian Persian, Polish, Portuguese, Rundi, Romanian; Moldavian; Moldovan, Russian, Slovenian, Serbian, Swedish, Turkmen, Tagalog, Klingon; tlhIngan-Hol, Toki Pona, Turkish, Tatar, Uighur; Uyghur, Ukrainian, Urdu, Vietnamese, Volapük, Waray, Wu Chinese and Yue Chinese ## Dataset Structure ### Data Instances Each data instance corresponds to a paraphrase, e.g.: ``` { 'paraphrase_set_id': '1483', 'sentence_id': '5778896', 'paraphrase': 'Ɣremt adlis-a.', 'lists': ['7546'], 'tags': [''], 'language': 'ber' } ``` ### Data Fields Each dialogue instance has the following fields: - `paraphrase_set_id`: a running number that groups together all sentences that are considered paraphrases of each other - `sentence_id`: OPUS sentence id - `paraphrase`: Sentential paraphrase in a given language for a given paraphrase_set_id - `lists`: Contributors can add sentences to list in order to specify the original source of the data - `tags`: Indicates morphological or phonological properties of the sentence when available - `language`: Language identifier, one of the 73 languages that belong to this dataset. ### Data Splits The dataset is having a single `train` split, contains a total of 1.9 million sentences, with 200 – 250 000 sentences per language ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Creative Commons Attribution 2.0 Generic ### Citation Information ``` @dataset{scherrer_yves_2020_3707949, author = {Scherrer, Yves}, title = {{TaPaCo: A Corpus of Sentential Paraphrases for 73 Languages}}, month = mar, year = 2020, publisher = {Zenodo}, version = {1.0}, doi = {10.5281/zenodo.3707949}, url = {https://doi.org/10.5281/zenodo.3707949} } ``` ### Contributions Thanks to [@pacman100](https://github.com/pacman100) for adding this dataset.
tapaco
[ "task_categories:text2text-generation", "task_categories:translation", "task_categories:text-classification", "task_ids:semantic-similarity-classification", "annotations_creators:machine-generated", "language_creators:crowdsourced", "multilinguality:multilingual", "size_categories:100K<n<1M", "size_categories:10K<n<100K", "size_categories:1K<n<10K", "size_categories:1M<n<10M", "size_categories:n<1K", "source_datasets:extended|other-tatoeba", "language:af", "language:ar", "language:az", "language:be", "language:ber", "language:bg", "language:bn", "language:br", "language:ca", "language:cbk", "language:cmn", "language:cs", "language:da", "language:de", "language:el", "language:en", "language:eo", "language:es", "language:et", "language:eu", "language:fi", "language:fr", "language:gl", "language:gos", "language:he", "language:hi", "language:hr", "language:hu", "language:hy", "language:ia", "language:id", "language:ie", "language:io", "language:is", "language:it", "language:ja", "language:jbo", "language:kab", "language:ko", "language:kw", "language:la", "language:lfn", "language:lt", "language:mk", "language:mr", "language:nb", "language:nds", "language:nl", "language:orv", "language:ota", "language:pes", "language:pl", "language:pt", "language:rn", "language:ro", "language:ru", "language:sl", "language:sr", "language:sv", "language:tk", "language:tl", "language:tlh", "language:tok", "language:tr", "language:tt", "language:ug", "language:uk", "language:ur", "language:vi", "language:vo", "language:war", "language:wuu", "language:yue", "license:cc-by-2.0", "paraphrase-generation", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["machine-generated"], "language_creators": ["crowdsourced"], "language": ["af", "ar", "az", "be", "ber", "bg", "bn", "br", "ca", "cbk", "cmn", "cs", "da", "de", "el", "en", "eo", "es", "et", "eu", "fi", "fr", "gl", "gos", "he", "hi", "hr", "hu", "hy", "ia", "id", "ie", "io", "is", "it", "ja", "jbo", "kab", "ko", "kw", "la", "lfn", "lt", "mk", "mr", "nb", "nds", "nl", "orv", "ota", "pes", "pl", "pt", "rn", "ro", "ru", "sl", "sr", "sv", "tk", "tl", "tlh", "tok", "tr", "tt", "ug", "uk", "ur", "vi", "vo", "war", "wuu", "yue"], "license": ["cc-by-2.0"], "multilinguality": ["multilingual"], "size_categories": ["100K<n<1M", "10K<n<100K", "1K<n<10K", "1M<n<10M", "n<1K"], "source_datasets": ["extended|other-tatoeba"], "task_categories": ["text2text-generation", "translation", "text-classification"], "task_ids": ["semantic-similarity-classification"], "paperswithcode_id": "tapaco", "pretty_name": "TaPaCo Corpus", "config_names": ["af", "all_languages", "ar", "az", "be", "ber", "bg", "bn", "br", "ca", "cbk", "cmn", "cs", "da", "de", "el", "en", "eo", "es", "et", "eu", "fi", "fr", "gl", "gos", "he", "hi", "hr", "hu", "hy", "ia", "id", "ie", "io", "is", "it", "ja", "jbo", "kab", "ko", "kw", "la", "lfn", "lt", "mk", "mr", "nb", "nds", "nl", "orv", "ota", "pes", "pl", "pt", "rn", "ro", "ru", "sl", "sr", "sv", "tk", "tl", "tlh", "tok", "tr", "tt", "ug", "uk", "ur", "vi", "vo", "war", "wuu", "yue"], "tags": ["paraphrase-generation"], "dataset_info": [{"config_name": "all_languages", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 162802556, "num_examples": 1926192}], "download_size": 32213126, "dataset_size": 162802556}, {"config_name": "af", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 21219, "num_examples": 307}], "download_size": 32213126, "dataset_size": 21219}, {"config_name": "ar", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 546200, "num_examples": 6446}], "download_size": 32213126, "dataset_size": 546200}, {"config_name": "az", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 44461, "num_examples": 624}], "download_size": 32213126, "dataset_size": 44461}, {"config_name": "be", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 140376, "num_examples": 1512}], "download_size": 32213126, "dataset_size": 140376}, {"config_name": "ber", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 5118620, "num_examples": 67484}], "download_size": 32213126, "dataset_size": 5118620}, {"config_name": "bg", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 590535, "num_examples": 6324}], "download_size": 32213126, "dataset_size": 590535}, {"config_name": "bn", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 146654, "num_examples": 1440}], "download_size": 32213126, "dataset_size": 146654}, {"config_name": "br", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 177919, "num_examples": 2536}], "download_size": 32213126, "dataset_size": 177919}, {"config_name": "ca", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 39404, "num_examples": 518}], "download_size": 32213126, "dataset_size": 39404}, {"config_name": "cbk", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 19404, "num_examples": 262}], "download_size": 32213126, "dataset_size": 19404}, {"config_name": "cmn", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 964514, "num_examples": 12549}], "download_size": 32213126, "dataset_size": 964514}, {"config_name": "cs", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 482292, "num_examples": 6659}], "download_size": 32213126, "dataset_size": 482292}, {"config_name": "da", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 848886, "num_examples": 11220}], "download_size": 32213126, "dataset_size": 848886}, {"config_name": "de", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 10593377, "num_examples": 125091}], "download_size": 32213126, "dataset_size": 10593377}, {"config_name": "el", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 926054, "num_examples": 10072}], "download_size": 32213126, "dataset_size": 926054}, {"config_name": "en", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 15070349, "num_examples": 158053}], "download_size": 32213126, "dataset_size": 15070349}, {"config_name": "eo", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 16810965, "num_examples": 207105}], "download_size": 32213126, "dataset_size": 16810965}, {"config_name": "es", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 6851135, "num_examples": 85064}], "download_size": 32213126, "dataset_size": 6851135}, {"config_name": "et", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 17127, "num_examples": 241}], "download_size": 32213126, "dataset_size": 17127}, {"config_name": "eu", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 42702, "num_examples": 573}], "download_size": 32213126, "dataset_size": 42702}, {"config_name": "fi", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 2520167, "num_examples": 31753}], "download_size": 32213126, "dataset_size": 2520167}, {"config_name": "fr", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 9481426, "num_examples": 116733}], "download_size": 32213126, "dataset_size": 9481426}, {"config_name": "gl", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 26551, "num_examples": 351}], "download_size": 32213126, "dataset_size": 26551}, {"config_name": "gos", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 18442, "num_examples": 279}], "download_size": 32213126, "dataset_size": 18442}, {"config_name": "he", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 6024345, "num_examples": 68350}], "download_size": 32213126, "dataset_size": 6024345}, {"config_name": "hi", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 209382, "num_examples": 1913}], "download_size": 32213126, "dataset_size": 209382}, {"config_name": "hr", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 36638, "num_examples": 505}], "download_size": 32213126, "dataset_size": 36638}, {"config_name": "hu", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 5289610, "num_examples": 67964}], "download_size": 32213126, "dataset_size": 5289610}, {"config_name": "hy", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 49230, "num_examples": 603}], "download_size": 32213126, "dataset_size": 49230}, {"config_name": "ia", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 194035, "num_examples": 2548}], "download_size": 32213126, "dataset_size": 194035}, {"config_name": "id", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 124568, "num_examples": 1602}], "download_size": 32213126, "dataset_size": 124568}, {"config_name": "ie", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 31956, "num_examples": 488}], "download_size": 32213126, "dataset_size": 31956}, {"config_name": "io", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 33892, "num_examples": 480}], "download_size": 32213126, "dataset_size": 33892}, {"config_name": "is", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 132062, "num_examples": 1641}], "download_size": 32213126, "dataset_size": 132062}, {"config_name": "it", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 15073750, "num_examples": 198919}], "download_size": 32213126, "dataset_size": 15073750}, {"config_name": "ja", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 4314423, "num_examples": 44267}], "download_size": 32213126, "dataset_size": 4314423}, {"config_name": "jbo", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 201564, "num_examples": 2704}], "download_size": 32213126, "dataset_size": 201564}, {"config_name": "kab", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1211051, "num_examples": 15944}], "download_size": 32213126, "dataset_size": 1211051}, {"config_name": "ko", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 40458, "num_examples": 503}], "download_size": 32213126, "dataset_size": 40458}, {"config_name": "kw", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 88577, "num_examples": 1328}], "download_size": 32213126, "dataset_size": 88577}, {"config_name": "la", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 485749, "num_examples": 6889}], "download_size": 32213126, "dataset_size": 485749}, {"config_name": "lfn", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 203383, "num_examples": 2313}], "download_size": 32213126, "dataset_size": 203383}, {"config_name": "lt", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 599166, "num_examples": 8042}], "download_size": 32213126, "dataset_size": 599166}, {"config_name": "mk", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1240185, "num_examples": 14678}], "download_size": 32213126, "dataset_size": 1240185}, {"config_name": "mr", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1838921, "num_examples": 16413}], "download_size": 32213126, "dataset_size": 1838921}, {"config_name": "nb", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 85371, "num_examples": 1094}], "download_size": 32213126, "dataset_size": 85371}, {"config_name": "nds", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 195021, "num_examples": 2633}], "download_size": 32213126, "dataset_size": 195021}, {"config_name": "nl", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1790975, "num_examples": 23561}], "download_size": 32213126, "dataset_size": 1790975}, {"config_name": "orv", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 40484, "num_examples": 471}], "download_size": 32213126, "dataset_size": 40484}, {"config_name": "ota", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 44996, "num_examples": 486}], "download_size": 32213126, "dataset_size": 44996}, {"config_name": "pes", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 433406, "num_examples": 4285}], "download_size": 32213126, "dataset_size": 433406}, {"config_name": "pl", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1722188, "num_examples": 22391}], "download_size": 32213126, "dataset_size": 1722188}, {"config_name": "pt", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 6141178, "num_examples": 78430}], "download_size": 32213126, "dataset_size": 6141178}, {"config_name": "rn", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 47387, "num_examples": 648}], "download_size": 32213126, "dataset_size": 47387}, {"config_name": "ro", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 162955, "num_examples": 2092}], "download_size": 32213126, "dataset_size": 162955}, {"config_name": "ru", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 24540667, "num_examples": 251263}], "download_size": 32213126, "dataset_size": 24540667}, {"config_name": "sl", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 49610, "num_examples": 706}], "download_size": 32213126, "dataset_size": 49610}, {"config_name": "sr", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 667308, "num_examples": 8175}], "download_size": 32213126, "dataset_size": 667308}, {"config_name": "sv", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 509884, "num_examples": 7005}], "download_size": 32213126, "dataset_size": 509884}, {"config_name": "tk", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 95047, "num_examples": 1165}], "download_size": 32213126, "dataset_size": 95047}, {"config_name": "tl", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 76059, "num_examples": 1017}], "download_size": 32213126, "dataset_size": 76059}, {"config_name": "tlh", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 185309, "num_examples": 2804}], "download_size": 32213126, "dataset_size": 185309}, {"config_name": "toki", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 310864, "num_examples": 3738}], "download_size": 32213126, "dataset_size": 310864}, {"config_name": "tr", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 11271158, "num_examples": 142088}], "download_size": 32213126, "dataset_size": 11271158}, {"config_name": "tt", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 277269, "num_examples": 2398}], "download_size": 32213126, "dataset_size": 277269}, {"config_name": "ug", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 118474, "num_examples": 1183}], "download_size": 32213126, "dataset_size": 118474}, {"config_name": "uk", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 4885677, "num_examples": 54431}], "download_size": 32213126, "dataset_size": 4885677}, {"config_name": "ur", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 24075, "num_examples": 252}], "download_size": 32213126, "dataset_size": 24075}, {"config_name": "vi", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 84773, "num_examples": 962}], "download_size": 32213126, "dataset_size": 84773}, {"config_name": "vo", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 22164, "num_examples": 328}], "download_size": 32213126, "dataset_size": 22164}, {"config_name": "war", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 25759, "num_examples": 327}], "download_size": 32213126, "dataset_size": 25759}, {"config_name": "wuu", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 31640, "num_examples": 408}], "download_size": 32213126, "dataset_size": 31640}, {"config_name": "yue", "features": [{"name": "paraphrase_set_id", "dtype": "string"}, {"name": "sentence_id", "dtype": "string"}, {"name": "paraphrase", "dtype": "string"}, {"name": "lists", "sequence": "string"}, {"name": "tags", "sequence": "string"}, {"name": "language", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 42766, "num_examples": 561}], "download_size": 32213126, "dataset_size": 42766}]}
2023-06-08T12:14:46+00:00
[]
[ "af", "ar", "az", "be", "ber", "bg", "bn", "br", "ca", "cbk", "cmn", "cs", "da", "de", "el", "en", "eo", "es", "et", "eu", "fi", "fr", "gl", "gos", "he", "hi", "hr", "hu", "hy", "ia", "id", "ie", "io", "is", "it", "ja", "jbo", "kab", "ko", "kw", "la", "lfn", "lt", "mk", "mr", "nb", "nds", "nl", "orv", "ota", "pes", "pl", "pt", "rn", "ro", "ru", "sl", "sr", "sv", "tk", "tl", "tlh", "tok", "tr", "tt", "ug", "uk", "ur", "vi", "vo", "war", "wuu", "yue" ]
TAGS #task_categories-text2text-generation #task_categories-translation #task_categories-text-classification #task_ids-semantic-similarity-classification #annotations_creators-machine-generated #language_creators-crowdsourced #multilinguality-multilingual #size_categories-100K<n<1M #size_categories-10K<n<100K #size_categories-1K<n<10K #size_categories-1M<n<10M #size_categories-n<1K #source_datasets-extended|other-tatoeba #language-Afrikaans #language-Arabic #language-Azerbaijani #language-Belarusian #language-ber #language-Bulgarian #language-Bengali #language-Breton #language-Catalan #language-Chavacano #language-Mandarin Chinese #language-Czech #language-Danish #language-German #language-Modern Greek (1453-) #language-English #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Finnish #language-French #language-Galician #language-Gronings #language-Hebrew #language-Hindi #language-Croatian #language-Hungarian #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Indonesian #language-Interlingue #language-Ido #language-Icelandic #language-Italian #language-Japanese #language-Lojban #language-Kabyle #language-Korean #language-Cornish #language-Latin #language-Lingua Franca Nova #language-Lithuanian #language-Macedonian #language-Marathi #language-Norwegian Bokmål #language-Low German #language-Dutch #language-Old Russian #language-Ottoman Turkish (1500-1928) #language-Iranian Persian #language-Polish #language-Portuguese #language-Rundi #language-Romanian #language-Russian #language-Slovenian #language-Serbian #language-Swedish #language-Turkmen #language-Tagalog #language-Klingon #language-Toki Pona #language-Turkish #language-Tatar #language-Uighur #language-Ukrainian #language-Urdu #language-Vietnamese #language-Volapük #language-Waray (Philippines) #language-Wu Chinese #language-Yue Chinese #license-cc-by-2.0 #paraphrase-generation #region-us
# Dataset Card for TaPaCo Corpus ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: TaPaCo: A Corpus of Sentential Paraphrases for 73 Languages - Paper: TaPaCo: A Corpus of Sentential Paraphrases for 73 Languages - Data: URL - Point of Contact: Yves Scherrer ### Dataset Summary A freely available paraphrase corpus for 73 languages extracted from the Tatoeba database. Tatoeba is a crowdsourcing project mainly geared towards language learners. Its aim is to provide example sentences and translations for particular linguistic constructions and words. The paraphrase corpus is created by populating a graph with Tatoeba sentences and equivalence links between sentences “meaning the same thing”. This graph is then traversed to extract sets of paraphrases. Several language-independent filters and pruning steps are applied to remove uninteresting sentences. A manual evaluation performed on three languages shows that between half and three quarters of inferred paraphrases are correct and that most remaining ones are either correct but trivial, or near-paraphrases that neutralize a morphological distinction. The corpus contains a total of 1.9 million sentences, with 200 – 250 000 sentences per language. It covers a range of languages for which, to our knowledge, no other paraphrase dataset exists. ### Supported Tasks and Leaderboards Paraphrase detection and generation have become popular tasks in NLP and are increasingly integrated into a wide variety of common downstream tasks such as machine translation , information retrieval, question answering, and semantic parsing. Most of the existing datasets cover only a single language – in most cases English – or a small number of languages. Furthermore, some paraphrase datasets focus on lexical and phrasal rather than sentential paraphrases, while others are created (semi -)automatically using machine translation. The number of sentences per language ranges from 200 to 250 000, which makes the dataset more suitable for fine-tuning and evaluation purposes than for training. It is well-suited for multi-reference evaluation of paraphrase generation models, as there is generally not a single correct way of paraphrasing a given input sentence. ### Languages The dataset contains paraphrases in Afrikaans, Arabic, Azerbaijani, Belarusian, Berber languages, Bulgarian, Bengali , Breton, Catalan; Valencian, Chavacano, Mandarin, Czech, Danish, German, Greek, Modern (1453-), English, Esperanto , Spanish; Castilian, Estonian, Basque, Finnish, French, Galician, Gronings, Hebrew, Hindi, Croatian, Hungarian , Armenian, Interlingua (International Auxiliary Language Association), Indonesian, Interlingue; Occidental, Ido , Icelandic, Italian, Japanese, Lojban, Kabyle, Korean, Cornish, Latin, Lingua Franca Nova\t, Lithuanian, Macedonian , Marathi, Bokmål, Norwegian; Norwegian Bokmål, Low German; Low Saxon; German, Low; Saxon, Low, Dutch; Flemish, ]Old Russian, Turkish, Ottoman (1500-1928), Iranian Persian, Polish, Portuguese, Rundi, Romanian; Moldavian; Moldovan, Russian, Slovenian, Serbian, Swedish, Turkmen, Tagalog, Klingon; tlhIngan-Hol, Toki Pona, Turkish, Tatar, Uighur; Uyghur, Ukrainian, Urdu, Vietnamese, Volapük, Waray, Wu Chinese and Yue Chinese ## Dataset Structure ### Data Instances Each data instance corresponds to a paraphrase, e.g.: ### Data Fields Each dialogue instance has the following fields: - 'paraphrase_set_id': a running number that groups together all sentences that are considered paraphrases of each other - 'sentence_id': OPUS sentence id - 'paraphrase': Sentential paraphrase in a given language for a given paraphrase_set_id - 'lists': Contributors can add sentences to list in order to specify the original source of the data - 'tags': Indicates morphological or phonological properties of the sentence when available - 'language': Language identifier, one of the 73 languages that belong to this dataset. ### Data Splits The dataset is having a single 'train' split, contains a total of 1.9 million sentences, with 200 – 250 000 sentences per language ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Creative Commons Attribution 2.0 Generic ### Contributions Thanks to @pacman100 for adding this dataset.
[ "# Dataset Card for TaPaCo Corpus", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: TaPaCo: A Corpus of Sentential Paraphrases for 73 Languages\n- Paper: TaPaCo: A Corpus of Sentential Paraphrases for 73 Languages\n- Data: URL\n- Point of Contact: Yves Scherrer", "### Dataset Summary\nA freely available paraphrase corpus for 73 languages extracted from the Tatoeba database. \nTatoeba is a crowdsourcing project mainly geared towards language learners. Its aim is to provide example sentences \nand translations for particular linguistic constructions and words. The paraphrase corpus is created by populating a \ngraph with Tatoeba sentences and equivalence links between sentences “meaning the same thing”. This graph is then \ntraversed to extract sets of paraphrases. Several language-independent filters and pruning steps are applied to \nremove uninteresting sentences. A manual evaluation performed on three languages shows that between half and three \nquarters of inferred paraphrases are correct and that most remaining ones are either correct but trivial, \nor near-paraphrases that neutralize a morphological distinction. The corpus contains a total of 1.9 million \nsentences, with 200 – 250 000 sentences per language. It covers a range of languages for which, to our knowledge,\nno other paraphrase dataset exists.", "### Supported Tasks and Leaderboards\nParaphrase detection and generation have become popular tasks in NLP\nand are increasingly integrated into a wide variety of common downstream tasks such as machine translation\n, information retrieval, question answering, and semantic parsing. Most of the existing datasets\n cover only a single language – in most cases English – or a small number of languages. Furthermore, some paraphrase\n datasets focus on lexical and phrasal rather than sentential paraphrases, while others are created (semi\n -)automatically using machine translation.\n\nThe number of sentences per language ranges from 200 to 250 000, which makes the dataset\nmore suitable for fine-tuning and evaluation purposes than\nfor training. It is well-suited for multi-reference evaluation\nof paraphrase generation models, as there is generally not a\nsingle correct way of paraphrasing a given input sentence.", "### Languages\n\nThe dataset contains paraphrases in Afrikaans, Arabic, Azerbaijani, Belarusian, Berber languages, Bulgarian, Bengali\n, Breton, Catalan; Valencian, Chavacano, Mandarin, Czech, Danish, German, Greek, Modern (1453-), English, Esperanto\n, Spanish; Castilian, Estonian, Basque, Finnish, French, Galician, Gronings, Hebrew, Hindi, Croatian, Hungarian\n, Armenian, Interlingua (International Auxiliary Language Association), Indonesian, Interlingue; Occidental, Ido\n, Icelandic, Italian, Japanese, Lojban, Kabyle, Korean, Cornish, Latin, Lingua Franca Nova\\t, Lithuanian, Macedonian\n, Marathi, Bokmål, Norwegian; Norwegian Bokmål, Low German; Low Saxon; German, Low; Saxon, Low, Dutch; Flemish, ]Old\n Russian, Turkish, Ottoman (1500-1928), Iranian Persian, Polish, Portuguese, Rundi, Romanian; Moldavian; Moldovan, \n Russian, Slovenian, Serbian, Swedish, Turkmen, Tagalog, Klingon; tlhIngan-Hol, Toki Pona, Turkish, Tatar, \n Uighur; Uyghur, Ukrainian, Urdu, Vietnamese, Volapük, Waray, Wu Chinese and Yue Chinese", "## Dataset Structure", "### Data Instances\nEach data instance corresponds to a paraphrase, e.g.:", "### Data Fields\nEach dialogue instance has the following fields:\n- 'paraphrase_set_id': a running number that groups together all sentences that are considered paraphrases of each\n other\n- 'sentence_id': OPUS sentence id\n- 'paraphrase': Sentential paraphrase in a given language for a given paraphrase_set_id\n- 'lists': Contributors can add sentences to list in order to specify the original source of the data\n- 'tags': Indicates morphological or phonological properties of the sentence when available\n- 'language': Language identifier, one of the 73 languages that belong to this dataset.", "### Data Splits\n\nThe dataset is having a single 'train' split, contains a total of 1.9 million sentences, with 200 – 250 000\n sentences per language", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCreative Commons Attribution 2.0 Generic", "### Contributions\n\nThanks to @pacman100 for adding this dataset." ]
[ "TAGS\n#task_categories-text2text-generation #task_categories-translation #task_categories-text-classification #task_ids-semantic-similarity-classification #annotations_creators-machine-generated #language_creators-crowdsourced #multilinguality-multilingual #size_categories-100K<n<1M #size_categories-10K<n<100K #size_categories-1K<n<10K #size_categories-1M<n<10M #size_categories-n<1K #source_datasets-extended|other-tatoeba #language-Afrikaans #language-Arabic #language-Azerbaijani #language-Belarusian #language-ber #language-Bulgarian #language-Bengali #language-Breton #language-Catalan #language-Chavacano #language-Mandarin Chinese #language-Czech #language-Danish #language-German #language-Modern Greek (1453-) #language-English #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Finnish #language-French #language-Galician #language-Gronings #language-Hebrew #language-Hindi #language-Croatian #language-Hungarian #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Indonesian #language-Interlingue #language-Ido #language-Icelandic #language-Italian #language-Japanese #language-Lojban #language-Kabyle #language-Korean #language-Cornish #language-Latin #language-Lingua Franca Nova #language-Lithuanian #language-Macedonian #language-Marathi #language-Norwegian Bokmål #language-Low German #language-Dutch #language-Old Russian #language-Ottoman Turkish (1500-1928) #language-Iranian Persian #language-Polish #language-Portuguese #language-Rundi #language-Romanian #language-Russian #language-Slovenian #language-Serbian #language-Swedish #language-Turkmen #language-Tagalog #language-Klingon #language-Toki Pona #language-Turkish #language-Tatar #language-Uighur #language-Ukrainian #language-Urdu #language-Vietnamese #language-Volapük #language-Waray (Philippines) #language-Wu Chinese #language-Yue Chinese #license-cc-by-2.0 #paraphrase-generation #region-us \n", "# Dataset Card for TaPaCo Corpus", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: TaPaCo: A Corpus of Sentential Paraphrases for 73 Languages\n- Paper: TaPaCo: A Corpus of Sentential Paraphrases for 73 Languages\n- Data: URL\n- Point of Contact: Yves Scherrer", "### Dataset Summary\nA freely available paraphrase corpus for 73 languages extracted from the Tatoeba database. \nTatoeba is a crowdsourcing project mainly geared towards language learners. Its aim is to provide example sentences \nand translations for particular linguistic constructions and words. The paraphrase corpus is created by populating a \ngraph with Tatoeba sentences and equivalence links between sentences “meaning the same thing”. This graph is then \ntraversed to extract sets of paraphrases. Several language-independent filters and pruning steps are applied to \nremove uninteresting sentences. A manual evaluation performed on three languages shows that between half and three \nquarters of inferred paraphrases are correct and that most remaining ones are either correct but trivial, \nor near-paraphrases that neutralize a morphological distinction. The corpus contains a total of 1.9 million \nsentences, with 200 – 250 000 sentences per language. It covers a range of languages for which, to our knowledge,\nno other paraphrase dataset exists.", "### Supported Tasks and Leaderboards\nParaphrase detection and generation have become popular tasks in NLP\nand are increasingly integrated into a wide variety of common downstream tasks such as machine translation\n, information retrieval, question answering, and semantic parsing. Most of the existing datasets\n cover only a single language – in most cases English – or a small number of languages. Furthermore, some paraphrase\n datasets focus on lexical and phrasal rather than sentential paraphrases, while others are created (semi\n -)automatically using machine translation.\n\nThe number of sentences per language ranges from 200 to 250 000, which makes the dataset\nmore suitable for fine-tuning and evaluation purposes than\nfor training. It is well-suited for multi-reference evaluation\nof paraphrase generation models, as there is generally not a\nsingle correct way of paraphrasing a given input sentence.", "### Languages\n\nThe dataset contains paraphrases in Afrikaans, Arabic, Azerbaijani, Belarusian, Berber languages, Bulgarian, Bengali\n, Breton, Catalan; Valencian, Chavacano, Mandarin, Czech, Danish, German, Greek, Modern (1453-), English, Esperanto\n, Spanish; Castilian, Estonian, Basque, Finnish, French, Galician, Gronings, Hebrew, Hindi, Croatian, Hungarian\n, Armenian, Interlingua (International Auxiliary Language Association), Indonesian, Interlingue; Occidental, Ido\n, Icelandic, Italian, Japanese, Lojban, Kabyle, Korean, Cornish, Latin, Lingua Franca Nova\\t, Lithuanian, Macedonian\n, Marathi, Bokmål, Norwegian; Norwegian Bokmål, Low German; Low Saxon; German, Low; Saxon, Low, Dutch; Flemish, ]Old\n Russian, Turkish, Ottoman (1500-1928), Iranian Persian, Polish, Portuguese, Rundi, Romanian; Moldavian; Moldovan, \n Russian, Slovenian, Serbian, Swedish, Turkmen, Tagalog, Klingon; tlhIngan-Hol, Toki Pona, Turkish, Tatar, \n Uighur; Uyghur, Ukrainian, Urdu, Vietnamese, Volapük, Waray, Wu Chinese and Yue Chinese", "## Dataset Structure", "### Data Instances\nEach data instance corresponds to a paraphrase, e.g.:", "### Data Fields\nEach dialogue instance has the following fields:\n- 'paraphrase_set_id': a running number that groups together all sentences that are considered paraphrases of each\n other\n- 'sentence_id': OPUS sentence id\n- 'paraphrase': Sentential paraphrase in a given language for a given paraphrase_set_id\n- 'lists': Contributors can add sentences to list in order to specify the original source of the data\n- 'tags': Indicates morphological or phonological properties of the sentence when available\n- 'language': Language identifier, one of the 73 languages that belong to this dataset.", "### Data Splits\n\nThe dataset is having a single 'train' split, contains a total of 1.9 million sentences, with 200 – 250 000\n sentences per language", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nCreative Commons Attribution 2.0 Generic", "### Contributions\n\nThanks to @pacman100 for adding this dataset." ]
[ 616, 9, 120, 56, 239, 200, 299, 6, 22, 151, 37, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 12, 17 ]
[ "passage: ", "passage: TAGS\n#task_categories-text2text-generation #task_categories-translation #task_categories-text-classification #task_ids-semantic-similarity-classification #annotations_creators-machine-generated #language_creators-crowdsourced #multilinguality-multilingual #size_categories-100K<n<1M #size_categories-10K<n<100K #size_categories-1K<n<10K #size_categories-1M<n<10M #size_categories-n<1K #source_datasets-extended|other-tatoeba #language-Afrikaans #language-Arabic #language-Azerbaijani #language-Belarusian #language-ber #language-Bulgarian #language-Bengali #language-Breton #language-Catalan #language-Chavacano #language-Mandarin Chinese #language-Czech #language-Danish #language-German #language-Modern Greek (1453-) #language-English #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Finnish #language-French #language-Galician #language-Gronings #language-Hebrew #language-Hindi #language-Croatian #language-Hungarian #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Indonesian #language-Interlingue #language-Ido #language-Icelandic #language-Italian #language-Japanese #language-Lojban #language-Kabyle #language-Korean #language-Cornish #language-Latin #language-Lingua Franca Nova #language-Lithuanian #language-Macedonian #language-Marathi #language-Norwegian Bokmål #language-Low German #language-Dutch #language-Old Russian #language-Ottoman Turkish (1500-1928) #language-Iranian Persian #language-Polish #language-Portuguese #language-Rundi #language-Romanian #language-Russian #language-Slovenian #language-Serbian #language-Swedish #language-Turkmen #language-Tagalog #language-Klingon #language-Toki Pona #language-Turkish #language-Tatar #language-Uighur #language-Ukrainian #language-Urdu #language-Vietnamese #language-Volapük #language-Waray (Philippines) #language-Wu Chinese #language-Yue Chinese #license-cc-by-2.0 #paraphrase-generation #region-us \n# Dataset Card for TaPaCo Corpus## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: TaPaCo: A Corpus of Sentential Paraphrases for 73 Languages\n- Paper: TaPaCo: A Corpus of Sentential Paraphrases for 73 Languages\n- Data: URL\n- Point of Contact: Yves Scherrer### Dataset Summary\nA freely available paraphrase corpus for 73 languages extracted from the Tatoeba database. \nTatoeba is a crowdsourcing project mainly geared towards language learners. Its aim is to provide example sentences \nand translations for particular linguistic constructions and words. The paraphrase corpus is created by populating a \ngraph with Tatoeba sentences and equivalence links between sentences “meaning the same thing”. This graph is then \ntraversed to extract sets of paraphrases. Several language-independent filters and pruning steps are applied to \nremove uninteresting sentences. A manual evaluation performed on three languages shows that between half and three \nquarters of inferred paraphrases are correct and that most remaining ones are either correct but trivial, \nor near-paraphrases that neutralize a morphological distinction. The corpus contains a total of 1.9 million \nsentences, with 200 – 250 000 sentences per language. It covers a range of languages for which, to our knowledge,\nno other paraphrase dataset exists.", "passage: ### Supported Tasks and Leaderboards\nParaphrase detection and generation have become popular tasks in NLP\nand are increasingly integrated into a wide variety of common downstream tasks such as machine translation\n, information retrieval, question answering, and semantic parsing. Most of the existing datasets\n cover only a single language – in most cases English – or a small number of languages. Furthermore, some paraphrase\n datasets focus on lexical and phrasal rather than sentential paraphrases, while others are created (semi\n -)automatically using machine translation.\n\nThe number of sentences per language ranges from 200 to 250 000, which makes the dataset\nmore suitable for fine-tuning and evaluation purposes than\nfor training. It is well-suited for multi-reference evaluation\nof paraphrase generation models, as there is generally not a\nsingle correct way of paraphrasing a given input sentence.### Languages\n\nThe dataset contains paraphrases in Afrikaans, Arabic, Azerbaijani, Belarusian, Berber languages, Bulgarian, Bengali\n, Breton, Catalan; Valencian, Chavacano, Mandarin, Czech, Danish, German, Greek, Modern (1453-), English, Esperanto\n, Spanish; Castilian, Estonian, Basque, Finnish, French, Galician, Gronings, Hebrew, Hindi, Croatian, Hungarian\n, Armenian, Interlingua (International Auxiliary Language Association), Indonesian, Interlingue; Occidental, Ido\n, Icelandic, Italian, Japanese, Lojban, Kabyle, Korean, Cornish, Latin, Lingua Franca Nova\\t, Lithuanian, Macedonian\n, Marathi, Bokmål, Norwegian; Norwegian Bokmål, Low German; Low Saxon; German, Low; Saxon, Low, Dutch; Flemish, ]Old\n Russian, Turkish, Ottoman (1500-1928), Iranian Persian, Polish, Portuguese, Rundi, Romanian; Moldavian; Moldovan, \n Russian, Slovenian, Serbian, Swedish, Turkmen, Tagalog, Klingon; tlhIngan-Hol, Toki Pona, Turkish, Tatar, \n Uighur; Uyghur, Ukrainian, Urdu, Vietnamese, Volapük, Waray, Wu Chinese and Yue Chinese## Dataset Structure### Data Instances\nEach data instance corresponds to a paraphrase, e.g.:### Data Fields\nEach dialogue instance has the following fields:\n- 'paraphrase_set_id': a running number that groups together all sentences that are considered paraphrases of each\n other\n- 'sentence_id': OPUS sentence id\n- 'paraphrase': Sentential paraphrase in a given language for a given paraphrase_set_id\n- 'lists': Contributors can add sentences to list in order to specify the original source of the data\n- 'tags': Indicates morphological or phonological properties of the sentence when available\n- 'language': Language identifier, one of the 73 languages that belong to this dataset." ]
dd881f51c276754635ecb1d00bf5ba44e13a93af
# Dataset Card for Tashkeela ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Tashkeela](https://sourceforge.net/projects/tashkeela/) - **Repository:** [Tashkeela](https://sourceforge.net/projects/tashkeela/) - **Paper:** [Tashkeela: Novel corpus of Arabic vocalized texts, data for auto-diacritization systems](https://www.sciencedirect.com/science/article/pii/S2352340917300112) - **Point of Contact:** [Taha Zerrouki](mailto:t_zerrouki@esi.dz) ### Dataset Summary It contains 75 million of fully vocalized words mainly 97 books from classical and modern Arabic language. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The dataset is based on Arabic. ## Dataset Structure ### Data Instances ``` {'book': 'zip://Tashkeela-arabic-diacritized-text-utf8-0.3/texts.txt/msa/al-kalema.org/أشكال-التجارب-في-مَثَل-الزارع.htm.txt::https://sourceforge.net/projects/tashkeela/files/latest/download', 'text': 'الكلمة\n\n\nصفحه اصلی\nاشترك\nالكتاب المقدس\nجميع المقالات\nالترتيب بالموضوع\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nهذا المقال على نسخة PDF\n\n\nأشكال التجارب في مَثَل الزارع\n\n\tقد رأينا في مقال " \nوسائل واشكال التجارب" الأشكال التي من الممكن أن تتخذها التجارب (وخاصة الاختبارات التي تأتي من خلال الآلام والاضطهاد وأشراك إطاعة شهوات الإنسان العتيق، الجسد)، نستطيع أيضاً أن نرى هذه الأقسام عاملة في مثال الزارع. هناك مجموعتين في مثال الزارع أنه برغم من سماعهم واستقبالهم للكلمة، إلا أنهم لم يجلبوا ثماراً. والسؤال هو لماذا؟\n\n1. التجارب في القسم الثاني من مثال الزارع\n\nفيما يخص القسم الثاني من مثال الزارع، تخبرنا عنها متى 13: 20- 21 ولوقا 8: 13 \nمتى 13: 20- 21\n" وَالْمَزْرُوعُ عَلَى الأَمَاكِنِ الْمُحْجِرَةِ هُوَ الَّذِي يَسْمَعُ الْكَلِمَةَ، وَحَالاً يَقْبَلُهَا بِفَرَحٍ، وَلكِنْ لَيْسَ لَهُ أَصْلٌ فِي ذَاتِهِ، بَلْ هُوَ إِلَى حِينٍ. فَإِذَا حَدَثَ ضِيقٌ أَوِ اضْطِهَادٌ مِنْ أَجْلِ الْكَلِمَةِ فَحَالاً يَعْثُرُ."\nلوقا 8: 13\n" وَالَّذِينَ عَلَى الصَّخْرِ هُمُ الَّذِينَ مَتَى سَمِعُوا يَقْبَلُونَ الْكَلِمَةَ بِفَرَحٍ، وَهؤُلاَءِ لَيْسَ لَهُمْ أَصْلٌ، فَيُؤْمِنُونَ إِلَى حِينٍ، وَفِي وَقْتِ التَّجْرِبَةِ يَرْتَدُّونَ."\n\nكما نرى، الناس في هذا القسم سمعوا الكلمة وحالاً قبلوها بفرح! بمعنى آخر، لقد كانوا متحمسين جداً تجاه الكلمة. ثم جاءت التجارب والاختبارات في شكل ضيق واضطهاد من أجل الكلمة، أي أنه بسبب الكلمة، اضطهد هؤلاء الناس. وعندئذ توقفوا. عوضاً عن أن يحفظوا ويتمسكوا بالكلمة التي قد حدث واستقبلوها بفرح، تراجعوا وسقطوا بعيداً، إن كنت مؤمناً صغيراً مليء بالحماسة تجاه الله، وبالرغم من أنه قد يبدو أنه لا يوجد شيطان من حولك، فهذا لن يستمر إلى الأبد. فالتجارب والاختبارات آتية. ستحتاج إلى أن تحفظ وتتمسك بالإيمان وبالكلمة التي قد حدث واستقبلتها بفرح. كما تقول لنا الكلمة:\nعبرانيين 10: 35- 39\n" فَلاَ تَطْرَحُوا ثِقَتَكُمُ الَّتِي لَهَا مُجَازَاةٌ عَظِيمَةٌ. لأَنَّكُمْ تَحْتَاجُونَ إِلَى الصَّبْرِ، حَتَّى إِذَا صَنَعْتُمْ مَشِيئَةَ اللهِ تَنَالُونَ الْمَوْعِدَ. لأَنَّهُ بَعْدَ قَلِيل جِدًّا «سَيَأْتِي الآتِي وَلاَ يُبْطِئُ. أَمَّا الْبَارُّ فَبِالإِيمَانِ يَحْيَا، وَإِنِ ارْتَدَّ لاَ تُسَرُّ بِهِ نَفْسِي». وَأَمَّا نَحْنُ فَلَسْنَا مِنَ الارْتِدَادِ لِلْهَلاَكِ، بَلْ مِنَ الإِيمَانِ لاقْتِنَاءِ النَّفْسِ."\n\nوالضيق قد يأخذ أشكالاً عديدة. رأيت أناساً يسقطون، تاركين الإيمان لأن آبائهم أو أقاربهم وأصدقائهم قد عارضوهم ورفضوهم بسبب إيمانهم. بالطبع قد يأخذ الاضطهاد أشكالاً أكثر من ذلك أيضاً، مثل أن تلقى في سجن أو أن تعذب لأجل إيمانك. قد يسبب الموت كذلك، كما حدث مع اسطفانوس ويعقوب أخو يوحنا. وتقول الكلمة من أجلك ومن أجل كل الذين حوكموا:\nرومية 16: 19- 20\n" لأَنَّ طَاعَتَكُمْ ذَاعَتْ إِلَى الْجَمِيعِ، فَأَفْرَحُ أَنَا بِكُمْ، وَأُرِيدُ أَنْ تَكُونُوا حُكَمَاءَ لِلْخَيْرِ وَبُسَطَاءَ لِلشَّرِّ. وَإِلهُ السَّلاَمِ سَيَسْحَقُ الشَّيْطَانَ تَحْتَ أَرْجُلِكُمْ سَرِيعًا."\nو بطرس الأولى 5: 8- 10\n" اُصْحُوا وَاسْهَرُوا. لأَنَّ إِبْلِيسَ خَصْمَكُمْ كَأَسَدٍ زَائِرٍ، يَجُولُ مُلْتَمِسًا مَنْ يَبْتَلِعُهُ هُوَ. فَقَاوِمُوهُ، رَاسِخِينَ فِي الإِيمَانِ، عَالِمِينَ أَنَّ نَفْسَ هذِهِ الآلاَمِ تُجْرَى عَلَى إِخْوَتِكُمُ الَّذِينَ فِي الْعَالَمِ. وَإِلهُ كُلِّ نِعْمَةٍ الَّذِي دَعَانَا إِلَى مَجْدِهِ الأَبَدِيِّ فِي الْمَسِيحِ يَسُوعَ، بَعْدَمَا تَأَلَّمْتُمْ يَسِيرًا، هُوَ يُكَمِّلُكُمْ، وَيُثَبِّتُكُمْ، وَيُقَوِّيكُمْ، وَيُمَكِّنُكُمْ."\n\nتمسك بالإيمان حتى النهاية. ضع حياتك ووضعك بين يدي الله وكن مستعداً لمواجهة أي شيء قد يحدث، أجل وحتى السخرية والعذاب. الله معك، سيقويك وسيعينك تماماً مثلما فعل مع يسوع في بستان جسثيماني. وتماماً مثلما فعل مع بولس في السجن عندما اضطهد من قِبَل اليهود (أعمال الرسل 23: 11). وكما قال بولس في كورنثوس الثانية 1: 7:" عَالِمِينَ أَنَّكُمْ كَمَا أَنْتُمْ شُرَكَاءُ فِي الآلاَمِ، كَذلِكَ فِي التَّعْزِيَةِ أَيْضًا." فالعزاء الآتي من الله يوازن أي سخرية أو أي عذاب قد يأتي إلينا من أي إنسان.\n\n2. التجارب في القسم الثالث من مثال الزارع\n\nبخصوص القسم الثالث من مثال الزارع، فنقرأ عنه في مرقس 4: 18- 19\n\n" وَهؤُلاَءِ هُمُ الَّذِينَ زُرِعُوا بَيْنَ الشَّوْكِ: هؤُلاَءِ هُمُ الَّذِينَ يَسْمَعُونَ الْكَلِمَةَ، وَهُمُومُ هذَا الْعَالَمِ وَغُرُورُ الْغِنَى وَشَهَوَاتُ سَائِرِ الأَشْيَاءِ تَدْخُلُ وَتَخْنُقُ الْكَلِمَةَ فَتَصِيرُ بِلاَ ثَمَرٍ."\nو لوقا 8: 14\n" وَالَّذِي سَقَطَ بَيْنَ الشَّوْكِ هُمُ الَّذِينَ يَسْمَعُونَ، ثُمَّ يَذْهَبُونَ فَيَخْتَنِقُونَ مِنْ هُمُومِ الْحَيَاةِ وَغِنَاهَا وَلَذَّاتِهَا، وَلاَ يُنْضِجُونَ ثَمَرًا."\n\nهؤلاء قد سمعوا الكلمة وفهموها ولكنهم صاروا بلا ثمر، وما هو السبب؟ السبب هو لأنهم تركوا أبواب قلوبهم مفتوحة لأشواك " وَهُمُومُ هذَا الْعَالَمِ وَغُرُورُ الْغِنَى وَشَهَوَاتُ سَائِرِ الأَشْيَاءِ" (مرقس 4: 19)، والتي تدخل فتخنق الكلمة، كما رأينا يعقوب دائماً ما يقول:\nيعقوب 1: 13- 15\n" لاَ يَقُلْ أَحَدٌ إِذَا جُرِّبَ: «إِنِّي أُجَرَّبُ مِنْ قِبَلِ اللهِ»، لأَنَّ اللهَ غَيْرُ مُجَرَّبٍ بِالشُّرُورِ، وَهُوَ لاَ يُجَرِّبُ أَحَدًا. وَلكِنَّ كُلَّ وَاحِدٍ يُجَرَّبُ إِذَا انْجَذَبَ وَانْخَدَعَ مِنْ شَهْوَتِهِ. ثُمَّ الشَّهْوَةُ إِذَا حَبِلَتْ تَلِدُ خَطِيَّةً، وَالْخَطِيَّةُ إِذَا كَمَلَتْ تُنْتِجُ مَوْتًا."\nوتيموثاوس الأولى 6: 9 تقول لنا\n" وَأَمَّا الَّذِينَ يُرِيدُونَ أَنْ يَكُونُوا أَغْنِيَاءَ، فَيَسْقُطُونَ فِي تَجْرِبَةٍ وَفَخٍّ وَشَهَوَاتٍ كَثِيرَةٍ غَبِيَّةٍ وَمُضِرَّةٍ، تُغَرِّقُ النَّاسَ فِي الْعَطَبِ وَالْهَلاَكِ."\n\nيجب أن نلاحظ شيئاً هنا: أن تأثير هموم الحياة هو نفس التأثير الذي لتجارب الغنى وشهوات الأشياء الأخرى. فهموم الحياة أيضاً لا تجلب الثمار، إذاً فإن اردت أن تكون مسيحياً مثمراً، أي مسيحي حقيقي وليس فقط مسيحي اسمي، فيجب عليك أن تزيل أشواك الهموم والغنى وملذات الحياة وأن تمنعهم من العودة مرة أخرى. تحتاج إلى أن تفعل شيئاً، تحتاج إلى أن تتغير والله سيعينك في هذا إن كنت حقاً تريده. التجارب في القسم الثالث من مثال الزارع لا تأتي من خلال الاضطهاد والآلام عن طريق الشيطان. ولكن هنا تأخذ التجارب صوراً أكثر مكراً والتي مع هذا تتطلب مقاومتنا. الاهتمام بما يهتم به هذا العالم ("هموم هذا العالم")، الرغبة في الغنى أو اشتهاء الأشياء الأخرى هي أمور خطيرة جداً. إنها أشواك يجب إزالتها. كما رأينا بولس يقول:\nرومية 13: 14\n" بَلِ الْبَسُوا الرَّبَّ يَسُوعَ الْمَسِيحَ، وَلاَ تَصْنَعُوا تَدْبِيرًا لِلْجَسَدِ لأَجْلِ الشَّهَوَاتِ."\n\n" لاَ تَصْنَعُوا تَدْبِيرًا لِلْجَسَدِ" والتي تعني أنه يجب علينا أن لا نهتم بالجسد وشهواته. ولكن عوضاً عن ذلك ينبغي لنا أن نطعم أنفسنا بلبن الكلمة الصافي الذي ننمو بواستطه (بطرس الأولى 2: 2).\n\n\nتاسوس كيولاشوجلو'} ``` ### Data Fields - `book` (str): Book filename. - `text` (str): Text of the book. ### Data Splits The dataset is not split. ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization The Modern Standard Arabic texts crawled from the Internet. #### Who are the source language producers? Websites. ### Annotations The dataset does not contain any additional annotations. #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [GNU General Public License, version 2 (GPLv2)](https://opensource.org/licenses/GPL-2.0). ### Citation Information The dataset was published on this [paper](https://www.sciencedirect.com/science/article/pii/S2352340917300112#!): ``` @article{zerrouki2017tashkeela, title={Tashkeela: Novel corpus of Arabic vocalized texts, data for auto-diacritization systems}, author={Zerrouki, Taha and Balla, Amar}, journal={Data in brief}, volume={11}, pages={147}, year={2017}, publisher={Elsevier} } ``` ### Contributions Thanks to [@zaidalyafeai](https://github.com/zaidalyafeai) for adding this dataset.
tashkeela
[ "task_categories:text-generation", "task_categories:fill-mask", "task_ids:language-modeling", "task_ids:masked-language-modeling", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:monolingual", "size_categories:n<1K", "source_datasets:original", "language:ar", "license:gpl-2.0", "diacritics-prediction", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["found"], "language": ["ar"], "license": ["gpl-2.0"], "multilinguality": ["monolingual"], "size_categories": ["n<1K"], "source_datasets": ["original"], "task_categories": ["text-generation", "fill-mask"], "task_ids": ["language-modeling", "masked-language-modeling"], "pretty_name": "Tashkeela", "tags": ["diacritics-prediction"], "dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "book", "dtype": "string"}], "config_name": "plain_text", "splits": [{"name": "train", "num_bytes": 1081110249, "num_examples": 97}], "download_size": 183393530, "dataset_size": 1081110249}}
2024-01-18T11:16:43+00:00
[]
[ "ar" ]
TAGS #task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-n<1K #source_datasets-original #language-Arabic #license-gpl-2.0 #diacritics-prediction #region-us
# Dataset Card for Tashkeela ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: Tashkeela - Repository: Tashkeela - Paper: Tashkeela: Novel corpus of Arabic vocalized texts, data for auto-diacritization systems - Point of Contact: Taha Zerrouki ### Dataset Summary It contains 75 million of fully vocalized words mainly 97 books from classical and modern Arabic language. ### Supported Tasks and Leaderboards ### Languages The dataset is based on Arabic. ## Dataset Structure ### Data Instances ### Data Fields - 'book' (str): Book filename. - 'text' (str): Text of the book. ### Data Splits The dataset is not split. ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization The Modern Standard Arabic texts crawled from the Internet. #### Who are the source language producers? Websites. ### Annotations The dataset does not contain any additional annotations. #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information GNU General Public License, version 2 (GPLv2). The dataset was published on this paper: ### Contributions Thanks to @zaidalyafeai for adding this dataset.
[ "# Dataset Card for Tashkeela", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Tashkeela\n- Repository: Tashkeela\n- Paper: Tashkeela: Novel corpus of Arabic vocalized texts, data for auto-diacritization systems\n- Point of Contact: Taha Zerrouki", "### Dataset Summary\n\nIt contains 75 million of fully vocalized words mainly\n97 books from classical and modern Arabic language.", "### Supported Tasks and Leaderboards", "### Languages\n\nThe dataset is based on Arabic.", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- 'book' (str): Book filename.\n- 'text' (str): Text of the book.", "### Data Splits\n\nThe dataset is not split.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe Modern Standard Arabic texts crawled from the Internet.", "#### Who are the source language producers?\n\nWebsites.", "### Annotations\n\nThe dataset does not contain any additional annotations.", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nGNU General Public License, version 2 (GPLv2).\n\n\n\nThe dataset was published on this paper:", "### Contributions\n\nThanks to @zaidalyafeai for adding this dataset." ]
[ "TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-n<1K #source_datasets-original #language-Arabic #license-gpl-2.0 #diacritics-prediction #region-us \n", "# Dataset Card for Tashkeela", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Tashkeela\n- Repository: Tashkeela\n- Paper: Tashkeela: Novel corpus of Arabic vocalized texts, data for auto-diacritization systems\n- Point of Contact: Taha Zerrouki", "### Dataset Summary\n\nIt contains 75 million of fully vocalized words mainly\n97 books from classical and modern Arabic language.", "### Supported Tasks and Leaderboards", "### Languages\n\nThe dataset is based on Arabic.", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- 'book' (str): Book filename.\n- 'text' (str): Text of the book.", "### Data Splits\n\nThe dataset is not split.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe Modern Standard Arabic texts crawled from the Internet.", "#### Who are the source language producers?\n\nWebsites.", "### Annotations\n\nThe dataset does not contain any additional annotations.", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nGNU General Public License, version 2 (GPLv2).\n\n\n\nThe dataset was published on this paper:", "### Contributions\n\nThanks to @zaidalyafeai for adding this dataset." ]
[ 119, 9, 120, 55, 28, 10, 12, 6, 6, 28, 12, 5, 7, 4, 22, 13, 17, 5, 9, 8, 8, 7, 8, 7, 5, 6, 28, 20 ]
[ "passage: TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-n<1K #source_datasets-original #language-Arabic #license-gpl-2.0 #diacritics-prediction #region-us \n# Dataset Card for Tashkeela## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: Tashkeela\n- Repository: Tashkeela\n- Paper: Tashkeela: Novel corpus of Arabic vocalized texts, data for auto-diacritization systems\n- Point of Contact: Taha Zerrouki### Dataset Summary\n\nIt contains 75 million of fully vocalized words mainly\n97 books from classical and modern Arabic language.### Supported Tasks and Leaderboards### Languages\n\nThe dataset is based on Arabic.## Dataset Structure### Data Instances### Data Fields\n\n- 'book' (str): Book filename.\n- 'text' (str): Text of the book.### Data Splits\n\nThe dataset is not split.## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization\n\nThe Modern Standard Arabic texts crawled from the Internet.#### Who are the source language producers?\n\nWebsites.### Annotations\n\nThe dataset does not contain any additional annotations.#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data" ]
fc239783e6ac1e54298bed211c787cc223f647de
# Dataset Card for Taskmaster-1 ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Taskmaster-1](https://research.google/tools/datasets/taskmaster-1/) - **Repository:** [GitHub](https://github.com/google-research-datasets/Taskmaster/tree/master/TM-1-2019) - **Paper:** [Taskmaster-1: Toward a Realistic and Diverse Dialog Dataset](https://arxiv.org/abs/1909.05358) - **Leaderboard:** N/A - **Point of Contact:** [Taskmaster Googlegroup](taskmaster-datasets@googlegroups.com) ### Dataset Summary Taskmaster-1 is a goal-oriented conversational dataset. It includes 13,215 task-based dialogs comprising six domains. Two procedures were used to create this collection, each with unique advantages. The first involves a two-person, spoken "Wizard of Oz" (WOz) approach in which trained agents and crowdsourced workers interact to complete the task while the second is "self-dialog" in which crowdsourced workers write the entire dialog themselves. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The dataset is in English language. ## Dataset Structure ### Data Instances A typical example looks like this ``` { "conversation_id":"dlg-336c8165-068e-4b4b-803d-18ef0676f668", "instruction_id":"restaurant-table-2", "utterances":[ { "index":0, "segments":[ ], "speaker":"USER", "text":"Hi, I'm looking for a place that sells spicy wet hotdogs, can you think of any?" }, { "index":1, "segments":[ { "annotations":[ { "name":"restaurant_reservation.name.restaurant.reject" } ], "end_index":37, "start_index":16, "text":"Spicy Wet Hotdogs LLC" } ], "speaker":"ASSISTANT", "text":"You might enjoy Spicy Wet Hotdogs LLC." }, { "index":2, "segments":[ ], "speaker":"USER", "text":"That sounds really good, can you make me a reservation?" }, { "index":3, "segments":[ ], "speaker":"ASSISTANT", "text":"Certainly, when would you like a reservation?" }, { "index":4, "segments":[ { "annotations":[ { "name":"restaurant_reservation.num.guests" }, { "name":"restaurant_reservation.num.guests" } ], "end_index":20, "start_index":18, "text":"50" } ], "speaker":"USER", "text":"I have a party of 50 who want a really sloppy dog on Saturday at noon." } ] } ``` ### Data Fields Each conversation in the data file has the following structure: - `conversation_id`: A universally unique identifier with the prefix 'dlg-'. The ID has no meaning. - `utterances`: A list of utterances that make up the conversation. - `instruction_id`: A reference to the file(s) containing the user (and, if applicable, agent) instructions for this conversation. Each utterance has the following fields: - `index`: A 0-based index indicating the order of the utterances in the conversation. - `speaker`: Either USER or ASSISTANT, indicating which role generated this utterance. - `text`: The raw text of the utterance. In case of self dialogs (one_person_dialogs), this is written by the crowdsourced worker. In case of the WOz dialogs, 'ASSISTANT' turns are written and 'USER' turns are transcribed from the spoken recordings of crowdsourced workers. - `segments`: A list of various text spans with semantic annotations. Each segment has the following fields: - `start_index`: The position of the start of the annotation in the utterance text. - `end_index`: The position of the end of the annotation in the utterance text. - `text`: The raw text that has been annotated. - `annotations`: A list of annotation details for this segment. Each annotation has a single field: - `name`: The annotation name. ### Data Splits - one_person_dialogs The data in `one_person_dialogs` config is split into `train`, `dev` and `test` splits. | | train | validation | test | |--------------|-------:|------------:|------:| | N. Instances | 6168 | 770 | 770 | - woz_dialogs The data in `woz_dialogs` config has no default splits. | | train | |--------------|-------:| | N. Instances | 5507 | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information The dataset is licensed under `Creative Commons Attribution 4.0 License` ### Citation Information [More Information Needed] ``` @inproceedings{48484, title = {Taskmaster-1: Toward a Realistic and Diverse Dialog Dataset}, author = {Bill Byrne and Karthik Krishnamoorthi and Chinnadhurai Sankar and Arvind Neelakantan and Daniel Duckworth and Semih Yavuz and Ben Goodrich and Amit Dubey and Kyu-Young Kim and Andy Cedilnik}, year = {2019} } ``` ### Contributions Thanks to [@patil-suraj](https://github.com/patil-suraj) for adding this dataset.
taskmaster1
[ "task_categories:text-generation", "task_categories:fill-mask", "task_ids:dialogue-modeling", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:cc-by-4.0", "arxiv:1909.05358", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text-generation", "fill-mask"], "task_ids": ["dialogue-modeling"], "paperswithcode_id": "taskmaster-1", "pretty_name": "Taskmaster-1", "dataset_info": [{"config_name": "one_person_dialogs", "features": [{"name": "conversation_id", "dtype": "string"}, {"name": "instruction_id", "dtype": "string"}, {"name": "utterances", "list": [{"name": "index", "dtype": "int32"}, {"name": "speaker", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "segments", "list": [{"name": "start_index", "dtype": "int32"}, {"name": "end_index", "dtype": "int32"}, {"name": "text", "dtype": "string"}, {"name": "annotations", "list": [{"name": "name", "dtype": "string"}]}]}]}], "splits": [{"name": "train", "num_bytes": 18037058, "num_examples": 6168}, {"name": "validation", "num_bytes": 2239656, "num_examples": 770}, {"name": "test", "num_bytes": 2224163, "num_examples": 770}], "download_size": 103276427, "dataset_size": 22500877}, {"config_name": "woz_dialogs", "features": [{"name": "conversation_id", "dtype": "string"}, {"name": "instruction_id", "dtype": "string"}, {"name": "utterances", "list": [{"name": "index", "dtype": "int32"}, {"name": "speaker", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "segments", "list": [{"name": "start_index", "dtype": "int32"}, {"name": "end_index", "dtype": "int32"}, {"name": "text", "dtype": "string"}, {"name": "annotations", "list": [{"name": "name", "dtype": "string"}]}]}]}], "splits": [{"name": "train", "num_bytes": 13028593, "num_examples": 5507}], "download_size": 103276427, "dataset_size": 13028593}]}
2024-01-18T11:16:45+00:00
[ "1909.05358" ]
[ "en" ]
TAGS #task_categories-text-generation #task_categories-fill-mask #task_ids-dialogue-modeling #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-cc-by-4.0 #arxiv-1909.05358 #region-us
Dataset Card for Taskmaster-1 ============================= Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: Taskmaster-1 * Repository: GitHub * Paper: Taskmaster-1: Toward a Realistic and Diverse Dialog Dataset * Leaderboard: N/A * Point of Contact: Taskmaster Googlegroup ### Dataset Summary Taskmaster-1 is a goal-oriented conversational dataset. It includes 13,215 task-based dialogs comprising six domains. Two procedures were used to create this collection, each with unique advantages. The first involves a two-person, spoken "Wizard of Oz" (WOz) approach in which trained agents and crowdsourced workers interact to complete the task while the second is "self-dialog" in which crowdsourced workers write the entire dialog themselves. ### Supported Tasks and Leaderboards ### Languages The dataset is in English language. Dataset Structure ----------------- ### Data Instances A typical example looks like this ### Data Fields Each conversation in the data file has the following structure: * 'conversation\_id': A universally unique identifier with the prefix 'dlg-'. The ID has no meaning. * 'utterances': A list of utterances that make up the conversation. * 'instruction\_id': A reference to the file(s) containing the user (and, if applicable, agent) instructions for this conversation. Each utterance has the following fields: * 'index': A 0-based index indicating the order of the utterances in the conversation. * 'speaker': Either USER or ASSISTANT, indicating which role generated this utterance. * 'text': The raw text of the utterance. In case of self dialogs (one\_person\_dialogs), this is written by the crowdsourced worker. In case of the WOz dialogs, 'ASSISTANT' turns are written and 'USER' turns are transcribed from the spoken recordings of crowdsourced workers. * 'segments': A list of various text spans with semantic annotations. Each segment has the following fields: * 'start\_index': The position of the start of the annotation in the utterance text. * 'end\_index': The position of the end of the annotation in the utterance text. * 'text': The raw text that has been annotated. * 'annotations': A list of annotation details for this segment. Each annotation has a single field: * 'name': The annotation name. ### Data Splits * one\_person\_dialogs The data in 'one\_person\_dialogs' config is split into 'train', 'dev' and 'test' splits. * woz\_dialogs The data in 'woz\_dialogs' config has no default splits. Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information The dataset is licensed under 'Creative Commons Attribution 4.0 License' ### Contributions Thanks to @patil-suraj for adding this dataset.
[ "### Dataset Summary\n\n\nTaskmaster-1 is a goal-oriented conversational dataset. It includes 13,215 task-based\ndialogs comprising six domains. Two procedures were used to create this collection,\neach with unique advantages. The first involves a two-person, spoken \"Wizard of Oz\" (WOz) approach\nin which trained agents and crowdsourced workers interact to complete the task while the second is\n\"self-dialog\" in which crowdsourced workers write the entire dialog themselves.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe dataset is in English language.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nA typical example looks like this", "### Data Fields\n\n\nEach conversation in the data file has the following structure:\n\n\n* 'conversation\\_id': A universally unique identifier with the prefix 'dlg-'. The ID has no meaning.\n* 'utterances': A list of utterances that make up the conversation.\n* 'instruction\\_id': A reference to the file(s) containing the user (and, if applicable, agent) instructions for this conversation.\n\n\nEach utterance has the following fields:\n\n\n* 'index': A 0-based index indicating the order of the utterances in the conversation.\n* 'speaker': Either USER or ASSISTANT, indicating which role generated this utterance.\n* 'text': The raw text of the utterance. In case of self dialogs (one\\_person\\_dialogs), this is written by the crowdsourced worker. In case of the WOz dialogs, 'ASSISTANT' turns are written and 'USER' turns are transcribed from the spoken recordings of crowdsourced workers.\n* 'segments': A list of various text spans with semantic annotations.\n\n\nEach segment has the following fields:\n\n\n* 'start\\_index': The position of the start of the annotation in the utterance text.\n* 'end\\_index': The position of the end of the annotation in the utterance text.\n* 'text': The raw text that has been annotated.\n* 'annotations': A list of annotation details for this segment.\n\n\nEach annotation has a single field:\n\n\n* 'name': The annotation name.", "### Data Splits\n\n\n* one\\_person\\_dialogs\n\n\nThe data in 'one\\_person\\_dialogs' config is split into 'train', 'dev' and 'test' splits.\n\n\n\n* woz\\_dialogs\n\n\nThe data in 'woz\\_dialogs' config has no default splits.\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nThe dataset is licensed under 'Creative Commons Attribution 4.0 License'", "### Contributions\n\n\nThanks to @patil-suraj for adding this dataset." ]
[ "TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-dialogue-modeling #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-cc-by-4.0 #arxiv-1909.05358 #region-us \n", "### Dataset Summary\n\n\nTaskmaster-1 is a goal-oriented conversational dataset. It includes 13,215 task-based\ndialogs comprising six domains. Two procedures were used to create this collection,\neach with unique advantages. The first involves a two-person, spoken \"Wizard of Oz\" (WOz) approach\nin which trained agents and crowdsourced workers interact to complete the task while the second is\n\"self-dialog\" in which crowdsourced workers write the entire dialog themselves.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe dataset is in English language.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nA typical example looks like this", "### Data Fields\n\n\nEach conversation in the data file has the following structure:\n\n\n* 'conversation\\_id': A universally unique identifier with the prefix 'dlg-'. The ID has no meaning.\n* 'utterances': A list of utterances that make up the conversation.\n* 'instruction\\_id': A reference to the file(s) containing the user (and, if applicable, agent) instructions for this conversation.\n\n\nEach utterance has the following fields:\n\n\n* 'index': A 0-based index indicating the order of the utterances in the conversation.\n* 'speaker': Either USER or ASSISTANT, indicating which role generated this utterance.\n* 'text': The raw text of the utterance. In case of self dialogs (one\\_person\\_dialogs), this is written by the crowdsourced worker. In case of the WOz dialogs, 'ASSISTANT' turns are written and 'USER' turns are transcribed from the spoken recordings of crowdsourced workers.\n* 'segments': A list of various text spans with semantic annotations.\n\n\nEach segment has the following fields:\n\n\n* 'start\\_index': The position of the start of the annotation in the utterance text.\n* 'end\\_index': The position of the end of the annotation in the utterance text.\n* 'text': The raw text that has been annotated.\n* 'annotations': A list of annotation details for this segment.\n\n\nEach annotation has a single field:\n\n\n* 'name': The annotation name.", "### Data Splits\n\n\n* one\\_person\\_dialogs\n\n\nThe data in 'one\\_person\\_dialogs' config is split into 'train', 'dev' and 'test' splits.\n\n\n\n* woz\\_dialogs\n\n\nThe data in 'woz\\_dialogs' config has no default splits.\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nThe dataset is licensed under 'Creative Commons Attribution 4.0 License'", "### Contributions\n\n\nThanks to @patil-suraj for adding this dataset." ]
[ 113, 113, 10, 19, 12, 371, 83, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 20, 19 ]
[ "passage: TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-dialogue-modeling #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-cc-by-4.0 #arxiv-1909.05358 #region-us \n### Dataset Summary\n\n\nTaskmaster-1 is a goal-oriented conversational dataset. It includes 13,215 task-based\ndialogs comprising six domains. Two procedures were used to create this collection,\neach with unique advantages. The first involves a two-person, spoken \"Wizard of Oz\" (WOz) approach\nin which trained agents and crowdsourced workers interact to complete the task while the second is\n\"self-dialog\" in which crowdsourced workers write the entire dialog themselves.### Supported Tasks and Leaderboards### Languages\n\n\nThe dataset is in English language.\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nA typical example looks like this" ]
e1e7b4aae68fae49c9dcd010745ecfe80bb16d75
# Dataset Card for Taskmaster-2 ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Taskmaster-1](https://research.google/tools/datasets/taskmaster-1/) - **Repository:** [GitHub](https://github.com/google-research-datasets/Taskmaster/tree/master/TM-2-2020) - **Paper:** [Taskmaster-1: Toward a Realistic and Diverse Dialog Dataset](https://arxiv.org/abs/1909.05358) - **Leaderboard:** N/A - **Point of Contact:** [Taskmaster Googlegroup](taskmaster-datasets@googlegroups.com) ### Dataset Summary Taskmaster is dataset for goal oriented conversations. The Taskmaster-2 dataset consists of 17,289 dialogs in the seven domains which include restaurants, food ordering, movies, hotels, flights, music and sports. Unlike Taskmaster-1, which includes both written "self-dialogs" and spoken two-person dialogs, Taskmaster-2 consists entirely of spoken two-person dialogs. In addition, while Taskmaster-1 is almost exclusively task-based, Taskmaster-2 contains a good number of search- and recommendation-oriented dialogs. All dialogs in this release were created using a Wizard of Oz (WOz) methodology in which crowdsourced workers played the role of a 'user' and trained call center operators played the role of the 'assistant'. In this way, users were led to believe they were interacting with an automated system that “spoke” using text-to-speech (TTS) even though it was in fact a human behind the scenes. As a result, users could express themselves however they chose in the context of an automated interface. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The dataset is in English language. ## Dataset Structure ### Data Instances A typical example looks like this ``` { "conversation_id": "dlg-0047a087-6a3c-4f27-b0e6-268f53a2e013", "instruction_id": "flight-6", "utterances": [ { "index": 0, "segments": [], "speaker": "USER", "text": "Hi, I'm looking for a flight. I need to visit a friend." }, { "index": 1, "segments": [], "speaker": "ASSISTANT", "text": "Hello, how can I help you?" }, { "index": 2, "segments": [], "speaker": "ASSISTANT", "text": "Sure, I can help you with that." }, { "index": 3, "segments": [], "speaker": "ASSISTANT", "text": "On what dates?" }, { "index": 4, "segments": [ { "annotations": [ { "name": "flight_search.date.depart_origin" } ], "end_index": 37, "start_index": 27, "text": "March 20th" }, { "annotations": [ { "name": "flight_search.date.return" } ], "end_index": 45, "start_index": 41, "text": "22nd" } ], "speaker": "USER", "text": "I'm looking to travel from March 20th to 22nd." } ] } ``` ### Data Fields Each conversation in the data file has the following structure: - `conversation_id`: A universally unique identifier with the prefix 'dlg-'. The ID has no meaning. - `utterances`: A list of utterances that make up the conversation. - `instruction_id`: A reference to the file(s) containing the user (and, if applicable, agent) instructions for this conversation. Each utterance has the following fields: - `index`: A 0-based index indicating the order of the utterances in the conversation. - `speaker`: Either USER or ASSISTANT, indicating which role generated this utterance. - `text`: The raw text of the utterance. In case of self dialogs (one_person_dialogs), this is written by the crowdsourced worker. In case of the WOz dialogs, 'ASSISTANT' turns are written and 'USER' turns are transcribed from the spoken recordings of crowdsourced workers. - `segments`: A list of various text spans with semantic annotations. Each segment has the following fields: - `start_index`: The position of the start of the annotation in the utterance text. - `end_index`: The position of the end of the annotation in the utterance text. - `text`: The raw text that has been annotated. - `annotations`: A list of annotation details for this segment. Each annotation has a single field: - `name`: The annotation name. ### Data Splits There are no deafults splits for all the config. The below table lists the number of examples in each config. | Config | Train | |-------------------|--------| | flights | 2481 | | food-orderings | 1050 | | hotels | 2355 | | movies | 3047 | | music | 1602 | | restaurant-search | 3276 | | sports | 3478 | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information The dataset is licensed under `Creative Commons Attribution 4.0 License` ### Citation Information [More Information Needed] ``` @inproceedings{48484, title = {Taskmaster-1: Toward a Realistic and Diverse Dialog Dataset}, author = {Bill Byrne and Karthik Krishnamoorthi and Chinnadhurai Sankar and Arvind Neelakantan and Daniel Duckworth and Semih Yavuz and Ben Goodrich and Amit Dubey and Kyu-Young Kim and Andy Cedilnik}, year = {2019} } ``` ### Contributions Thanks to [@patil-suraj](https://github.com/patil-suraj) for adding this dataset.
taskmaster2
[ "task_categories:text-generation", "task_categories:fill-mask", "task_ids:dialogue-modeling", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:cc-by-4.0", "arxiv:1909.05358", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text-generation", "fill-mask"], "task_ids": ["dialogue-modeling"], "paperswithcode_id": "taskmaster-2", "pretty_name": "Taskmaster-2", "dataset_info": [{"config_name": "flights", "features": [{"name": "conversation_id", "dtype": "string"}, {"name": "instruction_id", "dtype": "string"}, {"name": "utterances", "list": [{"name": "index", "dtype": "int32"}, {"name": "speaker", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "segments", "list": [{"name": "start_index", "dtype": "int32"}, {"name": "end_index", "dtype": "int32"}, {"name": "text", "dtype": "string"}, {"name": "annotations", "list": [{"name": "name", "dtype": "string"}]}]}]}], "splits": [{"name": "train", "num_bytes": 7073487, "num_examples": 2481}], "download_size": 23029880, "dataset_size": 7073487}, {"config_name": "food-ordering", "features": [{"name": "conversation_id", "dtype": "string"}, {"name": "instruction_id", "dtype": "string"}, {"name": "utterances", "list": [{"name": "index", "dtype": "int32"}, {"name": "speaker", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "segments", "list": [{"name": "start_index", "dtype": "int32"}, {"name": "end_index", "dtype": "int32"}, {"name": "text", "dtype": "string"}, {"name": "annotations", "list": [{"name": "name", "dtype": "string"}]}]}]}], "splits": [{"name": "train", "num_bytes": 1734825, "num_examples": 1050}], "download_size": 5376675, "dataset_size": 1734825}, {"config_name": "hotels", "features": [{"name": "conversation_id", "dtype": "string"}, {"name": "instruction_id", "dtype": "string"}, {"name": "utterances", "list": [{"name": "index", "dtype": "int32"}, {"name": "speaker", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "segments", "list": [{"name": "start_index", "dtype": "int32"}, {"name": "end_index", "dtype": "int32"}, {"name": "text", "dtype": "string"}, {"name": "annotations", "list": [{"name": "name", "dtype": "string"}]}]}]}], "splits": [{"name": "train", "num_bytes": 7436667, "num_examples": 2357}], "download_size": 22507266, "dataset_size": 7436667}, {"config_name": "movies", "features": [{"name": "conversation_id", "dtype": "string"}, {"name": "instruction_id", "dtype": "string"}, {"name": "utterances", "list": [{"name": "index", "dtype": "int32"}, {"name": "speaker", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "segments", "list": [{"name": "start_index", "dtype": "int32"}, {"name": "end_index", "dtype": "int32"}, {"name": "text", "dtype": "string"}, {"name": "annotations", "list": [{"name": "name", "dtype": "string"}]}]}]}], "splits": [{"name": "train", "num_bytes": 7112301, "num_examples": 3056}], "download_size": 21189893, "dataset_size": 7112301}, {"config_name": "music", "features": [{"name": "conversation_id", "dtype": "string"}, {"name": "instruction_id", "dtype": "string"}, {"name": "utterances", "list": [{"name": "index", "dtype": "int32"}, {"name": "speaker", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "segments", "list": [{"name": "start_index", "dtype": "int32"}, {"name": "end_index", "dtype": "int32"}, {"name": "text", "dtype": "string"}, {"name": "annotations", "list": [{"name": "name", "dtype": "string"}]}]}]}], "splits": [{"name": "train", "num_bytes": 2814030, "num_examples": 1603}], "download_size": 8981720, "dataset_size": 2814030}, {"config_name": "restaurant-search", "features": [{"name": "conversation_id", "dtype": "string"}, {"name": "instruction_id", "dtype": "string"}, {"name": "utterances", "list": [{"name": "index", "dtype": "int32"}, {"name": "speaker", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "segments", "list": [{"name": "start_index", "dtype": "int32"}, {"name": "end_index", "dtype": "int32"}, {"name": "text", "dtype": "string"}, {"name": "annotations", "list": [{"name": "name", "dtype": "string"}]}]}]}], "splits": [{"name": "train", "num_bytes": 7341998, "num_examples": 3276}], "download_size": 21472680, "dataset_size": 7341998}, {"config_name": "sports", "features": [{"name": "conversation_id", "dtype": "string"}, {"name": "instruction_id", "dtype": "string"}, {"name": "utterances", "list": [{"name": "index", "dtype": "int32"}, {"name": "speaker", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "segments", "list": [{"name": "start_index", "dtype": "int32"}, {"name": "end_index", "dtype": "int32"}, {"name": "text", "dtype": "string"}, {"name": "annotations", "list": [{"name": "name", "dtype": "string"}]}]}]}], "splits": [{"name": "train", "num_bytes": 5738818, "num_examples": 3481}], "download_size": 19549440, "dataset_size": 5738818}]}
2024-01-18T11:16:46+00:00
[ "1909.05358" ]
[ "en" ]
TAGS #task_categories-text-generation #task_categories-fill-mask #task_ids-dialogue-modeling #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-cc-by-4.0 #arxiv-1909.05358 #region-us
Dataset Card for Taskmaster-2 ============================= Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: Taskmaster-1 * Repository: GitHub * Paper: Taskmaster-1: Toward a Realistic and Diverse Dialog Dataset * Leaderboard: N/A * Point of Contact: Taskmaster Googlegroup ### Dataset Summary Taskmaster is dataset for goal oriented conversations. The Taskmaster-2 dataset consists of 17,289 dialogs in the seven domains which include restaurants, food ordering, movies, hotels, flights, music and sports. Unlike Taskmaster-1, which includes both written "self-dialogs" and spoken two-person dialogs, Taskmaster-2 consists entirely of spoken two-person dialogs. In addition, while Taskmaster-1 is almost exclusively task-based, Taskmaster-2 contains a good number of search- and recommendation-oriented dialogs. All dialogs in this release were created using a Wizard of Oz (WOz) methodology in which crowdsourced workers played the role of a 'user' and trained call center operators played the role of the 'assistant'. In this way, users were led to believe they were interacting with an automated system that “spoke” using text-to-speech (TTS) even though it was in fact a human behind the scenes. As a result, users could express themselves however they chose in the context of an automated interface. ### Supported Tasks and Leaderboards ### Languages The dataset is in English language. Dataset Structure ----------------- ### Data Instances A typical example looks like this ### Data Fields Each conversation in the data file has the following structure: * 'conversation\_id': A universally unique identifier with the prefix 'dlg-'. The ID has no meaning. * 'utterances': A list of utterances that make up the conversation. * 'instruction\_id': A reference to the file(s) containing the user (and, if applicable, agent) instructions for this conversation. Each utterance has the following fields: * 'index': A 0-based index indicating the order of the utterances in the conversation. * 'speaker': Either USER or ASSISTANT, indicating which role generated this utterance. * 'text': The raw text of the utterance. In case of self dialogs (one\_person\_dialogs), this is written by the crowdsourced worker. In case of the WOz dialogs, 'ASSISTANT' turns are written and 'USER' turns are transcribed from the spoken recordings of crowdsourced workers. * 'segments': A list of various text spans with semantic annotations. Each segment has the following fields: * 'start\_index': The position of the start of the annotation in the utterance text. * 'end\_index': The position of the end of the annotation in the utterance text. * 'text': The raw text that has been annotated. * 'annotations': A list of annotation details for this segment. Each annotation has a single field: * 'name': The annotation name. ### Data Splits There are no deafults splits for all the config. The below table lists the number of examples in each config. Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information The dataset is licensed under 'Creative Commons Attribution 4.0 License' ### Contributions Thanks to @patil-suraj for adding this dataset.
[ "### Dataset Summary\n\n\nTaskmaster is dataset for goal oriented conversations. The Taskmaster-2 dataset consists of 17,289 dialogs\nin the seven domains which include restaurants, food ordering, movies, hotels, flights, music and sports.\nUnlike Taskmaster-1, which includes both written \"self-dialogs\" and spoken two-person dialogs,\nTaskmaster-2 consists entirely of spoken two-person dialogs. In addition, while Taskmaster-1 is\nalmost exclusively task-based, Taskmaster-2 contains a good number of search- and recommendation-oriented dialogs.\nAll dialogs in this release were created using a Wizard of Oz (WOz) methodology in which crowdsourced\nworkers played the role of a 'user' and trained call center operators played the role of the 'assistant'.\nIn this way, users were led to believe they were interacting with an automated system that “spoke”\nusing text-to-speech (TTS) even though it was in fact a human behind the scenes.\nAs a result, users could express themselves however they chose in the context of an automated interface.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe dataset is in English language.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nA typical example looks like this", "### Data Fields\n\n\nEach conversation in the data file has the following structure:\n\n\n* 'conversation\\_id': A universally unique identifier with the prefix 'dlg-'. The ID has no meaning.\n* 'utterances': A list of utterances that make up the conversation.\n* 'instruction\\_id': A reference to the file(s) containing the user (and, if applicable, agent) instructions for this conversation.\n\n\nEach utterance has the following fields:\n\n\n* 'index': A 0-based index indicating the order of the utterances in the conversation.\n* 'speaker': Either USER or ASSISTANT, indicating which role generated this utterance.\n* 'text': The raw text of the utterance. In case of self dialogs (one\\_person\\_dialogs), this is written by the crowdsourced worker. In case of the WOz dialogs, 'ASSISTANT' turns are written and 'USER' turns are transcribed from the spoken recordings of crowdsourced workers.\n* 'segments': A list of various text spans with semantic annotations.\n\n\nEach segment has the following fields:\n\n\n* 'start\\_index': The position of the start of the annotation in the utterance text.\n* 'end\\_index': The position of the end of the annotation in the utterance text.\n* 'text': The raw text that has been annotated.\n* 'annotations': A list of annotation details for this segment.\n\n\nEach annotation has a single field:\n\n\n* 'name': The annotation name.", "### Data Splits\n\n\nThere are no deafults splits for all the config. The below table lists the number of examples in each config.\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nThe dataset is licensed under 'Creative Commons Attribution 4.0 License'", "### Contributions\n\n\nThanks to @patil-suraj for adding this dataset." ]
[ "TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-dialogue-modeling #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-cc-by-4.0 #arxiv-1909.05358 #region-us \n", "### Dataset Summary\n\n\nTaskmaster is dataset for goal oriented conversations. The Taskmaster-2 dataset consists of 17,289 dialogs\nin the seven domains which include restaurants, food ordering, movies, hotels, flights, music and sports.\nUnlike Taskmaster-1, which includes both written \"self-dialogs\" and spoken two-person dialogs,\nTaskmaster-2 consists entirely of spoken two-person dialogs. In addition, while Taskmaster-1 is\nalmost exclusively task-based, Taskmaster-2 contains a good number of search- and recommendation-oriented dialogs.\nAll dialogs in this release were created using a Wizard of Oz (WOz) methodology in which crowdsourced\nworkers played the role of a 'user' and trained call center operators played the role of the 'assistant'.\nIn this way, users were led to believe they were interacting with an automated system that “spoke”\nusing text-to-speech (TTS) even though it was in fact a human behind the scenes.\nAs a result, users could express themselves however they chose in the context of an automated interface.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe dataset is in English language.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nA typical example looks like this", "### Data Fields\n\n\nEach conversation in the data file has the following structure:\n\n\n* 'conversation\\_id': A universally unique identifier with the prefix 'dlg-'. The ID has no meaning.\n* 'utterances': A list of utterances that make up the conversation.\n* 'instruction\\_id': A reference to the file(s) containing the user (and, if applicable, agent) instructions for this conversation.\n\n\nEach utterance has the following fields:\n\n\n* 'index': A 0-based index indicating the order of the utterances in the conversation.\n* 'speaker': Either USER or ASSISTANT, indicating which role generated this utterance.\n* 'text': The raw text of the utterance. In case of self dialogs (one\\_person\\_dialogs), this is written by the crowdsourced worker. In case of the WOz dialogs, 'ASSISTANT' turns are written and 'USER' turns are transcribed from the spoken recordings of crowdsourced workers.\n* 'segments': A list of various text spans with semantic annotations.\n\n\nEach segment has the following fields:\n\n\n* 'start\\_index': The position of the start of the annotation in the utterance text.\n* 'end\\_index': The position of the end of the annotation in the utterance text.\n* 'text': The raw text that has been annotated.\n* 'annotations': A list of annotation details for this segment.\n\n\nEach annotation has a single field:\n\n\n* 'name': The annotation name.", "### Data Splits\n\n\nThere are no deafults splits for all the config. The below table lists the number of examples in each config.\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nThe dataset is licensed under 'Creative Commons Attribution 4.0 License'", "### Contributions\n\n\nThanks to @patil-suraj for adding this dataset." ]
[ 113, 254, 10, 19, 12, 371, 41, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 20, 19 ]
[ "passage: TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-dialogue-modeling #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-cc-by-4.0 #arxiv-1909.05358 #region-us \n### Dataset Summary\n\n\nTaskmaster is dataset for goal oriented conversations. The Taskmaster-2 dataset consists of 17,289 dialogs\nin the seven domains which include restaurants, food ordering, movies, hotels, flights, music and sports.\nUnlike Taskmaster-1, which includes both written \"self-dialogs\" and spoken two-person dialogs,\nTaskmaster-2 consists entirely of spoken two-person dialogs. In addition, while Taskmaster-1 is\nalmost exclusively task-based, Taskmaster-2 contains a good number of search- and recommendation-oriented dialogs.\nAll dialogs in this release were created using a Wizard of Oz (WOz) methodology in which crowdsourced\nworkers played the role of a 'user' and trained call center operators played the role of the 'assistant'.\nIn this way, users were led to believe they were interacting with an automated system that “spoke”\nusing text-to-speech (TTS) even though it was in fact a human behind the scenes.\nAs a result, users could express themselves however they chose in the context of an automated interface.### Supported Tasks and Leaderboards### Languages\n\n\nThe dataset is in English language.\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nA typical example looks like this" ]
b8d443036a8c7b4998be4630fb0ecea5ddd473fc
# Dataset Card for taskmaster3 ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Taskmaster](https://research.google/tools/datasets/taskmaster-1/) - **Repository:** [GitHub](https://github.com/google-research-datasets/Taskmaster/tree/master/TM-3-2020) - **Paper:** [Taskmaster-1: Toward a Realistic and Diverse Dialog Dataset](https://arxiv.org/abs/1909.05358) - **Leaderboard:** N/A - **Point of Contact:** [Taskmaster Googlegroup](taskmaster-datasets@googlegroups.com) ### Dataset Summary Taskmaster is dataset for goal oriented conversations. The Taskmaster-3 dataset consists of 23,757 movie ticketing dialogs. By "movie ticketing" we mean conversations where the customer's goal is to purchase tickets after deciding on theater, time, movie name, number of tickets, and date, or opt out of the transaction. This collection was created using the "self-dialog" method. This means a single, crowd-sourced worker is paid to create a conversation writing turns for both speakers, i.e. the customer and the ticketing agent. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The dataset is in English language. ## Dataset Structure ### Data Instances A typical example looks like this ``` { "conversation_id": "dlg-ddee80da-9ffa-4773-9ce7-f73f727cb79c", "instructions": "SCENARIO: Pretend you’re *using a digital assistant to purchase tickets for a movie currently showing in theaters*. ...", "scenario": "4 exchanges with 1 error and predefined variables", "utterances": [ { "apis": [], "index": 0, "segments": [ { "annotations": [ { "name": "num.tickets" } ], "end_index": 21, "start_index": 20, "text": "2" }, { "annotations": [ { "name": "name.movie" } ], "end_index": 42, "start_index": 37, "text": "Mulan" } ], "speaker": "user", "text": "I would like to buy 2 tickets to see Mulan." }, { "index": 6, "segments": [], "speaker": "user", "text": "Yes.", "apis": [ { "args": [ { "arg_name": "name.movie", "arg_value": "Mulan" }, { "arg_name": "name.theater", "arg_value": "Mountain AMC 16" } ], "index": 6, "name": "book_tickets", "response": [ { "response_name": "status", "response_value": "success" } ] } ] } ], "vertical": "Movie Tickets" } ``` ### Data Fields Each conversation in the data file has the following structure: - `conversation_id`: A universally unique identifier with the prefix 'dlg-'. The ID has no meaning. - `utterances`: A list of utterances that make up the conversation. - `instructions`: Instructions for the crowdsourced worker used in creating the conversation. - `vertical`: In this dataset the vertical for all dialogs is "Movie Tickets". - `scenario`: This is the title of the instructions for each dialog. Each utterance has the following fields: - `index`: A 0-based index indicating the order of the utterances in the conversation. - `speaker`: Either USER or ASSISTANT, indicating which role generated this utterance. - `text`: The raw text of the utterance. In case of self dialogs (one_person_dialogs), this is written by the crowdsourced worker. In case of the WOz dialogs, 'ASSISTANT' turns are written and 'USER' turns are transcribed from the spoken recordings of crowdsourced workers. - `segments`: A list of various text spans with semantic annotations. - `apis`: An array of API invocations made during the utterance. Each API has the following structure: - `name`: The name of the API invoked (e.g. find_movies). - `index`: The index of the parent utterance. - `args`: A `list` of `dict` with keys `arg_name` and `arg_value` which represent the name of the argument and the value for the argument respectively. - `response`: A `list` of `dict`s with keys `response_name` and `response_value` which represent the name of the response and the value for the response respectively. Each segment has the following fields: - `start_index`: The position of the start of the annotation in the utterance text. - `end_index`: The position of the end of the annotation in the utterance text. - `text`: The raw text that has been annotated. - `annotations`: A list of annotation details for this segment. Each annotation has a single field: - `name`: The annotation name. ### Data Splits There are no deafults splits for all the config. The below table lists the number of examples in each config. | | Train | |-------------------|--------| | n_instances | 23757 | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information The dataset is licensed under `Creative Commons Attribution 4.0 License` ### Citation Information [More Information Needed] ``` @inproceedings{48484, title = {Taskmaster-1: Toward a Realistic and Diverse Dialog Dataset}, author = {Bill Byrne and Karthik Krishnamoorthi and Chinnadhurai Sankar and Arvind Neelakantan and Daniel Duckworth and Semih Yavuz and Ben Goodrich and Amit Dubey and Kyu-Young Kim and Andy Cedilnik}, year = {2019} } ``` ### Contributions Thanks to [@patil-suraj](https://github.com/patil-suraj) for adding this dataset.
taskmaster3
[ "task_categories:text-generation", "task_categories:fill-mask", "task_ids:dialogue-modeling", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:cc-by-4.0", "arxiv:1909.05358", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-generation", "fill-mask"], "task_ids": ["dialogue-modeling"], "pretty_name": "taskmaster3", "dataset_info": {"features": [{"name": "conversation_id", "dtype": "string"}, {"name": "vertical", "dtype": "string"}, {"name": "instructions", "dtype": "string"}, {"name": "scenario", "dtype": "string"}, {"name": "utterances", "list": [{"name": "index", "dtype": "int32"}, {"name": "speaker", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "apis", "list": [{"name": "name", "dtype": "string"}, {"name": "index", "dtype": "int32"}, {"name": "args", "list": [{"name": "arg_name", "dtype": "string"}, {"name": "arg_value", "dtype": "string"}]}, {"name": "response", "list": [{"name": "response_name", "dtype": "string"}, {"name": "response_value", "dtype": "string"}]}]}, {"name": "segments", "list": [{"name": "start_index", "dtype": "int32"}, {"name": "end_index", "dtype": "int32"}, {"name": "text", "dtype": "string"}, {"name": "annotations", "list": [{"name": "name", "dtype": "string"}]}]}]}], "splits": [{"name": "train", "num_bytes": 143609327, "num_examples": 23757}], "download_size": 313402141, "dataset_size": 143609327}}
2024-01-18T11:16:47+00:00
[ "1909.05358" ]
[ "en" ]
TAGS #task_categories-text-generation #task_categories-fill-mask #task_ids-dialogue-modeling #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-4.0 #arxiv-1909.05358 #region-us
Dataset Card for taskmaster3 ============================ Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: Taskmaster * Repository: GitHub * Paper: Taskmaster-1: Toward a Realistic and Diverse Dialog Dataset * Leaderboard: N/A * Point of Contact: Taskmaster Googlegroup ### Dataset Summary Taskmaster is dataset for goal oriented conversations. The Taskmaster-3 dataset consists of 23,757 movie ticketing dialogs. By "movie ticketing" we mean conversations where the customer's goal is to purchase tickets after deciding on theater, time, movie name, number of tickets, and date, or opt out of the transaction. This collection was created using the "self-dialog" method. This means a single, crowd-sourced worker is paid to create a conversation writing turns for both speakers, i.e. the customer and the ticketing agent. ### Supported Tasks and Leaderboards ### Languages The dataset is in English language. Dataset Structure ----------------- ### Data Instances A typical example looks like this ### Data Fields Each conversation in the data file has the following structure: * 'conversation\_id': A universally unique identifier with the prefix 'dlg-'. The ID has no meaning. * 'utterances': A list of utterances that make up the conversation. * 'instructions': Instructions for the crowdsourced worker used in creating the conversation. * 'vertical': In this dataset the vertical for all dialogs is "Movie Tickets". * 'scenario': This is the title of the instructions for each dialog. Each utterance has the following fields: * 'index': A 0-based index indicating the order of the utterances in the conversation. * 'speaker': Either USER or ASSISTANT, indicating which role generated this utterance. * 'text': The raw text of the utterance. In case of self dialogs (one\_person\_dialogs), this is written by the crowdsourced worker. In case of the WOz dialogs, 'ASSISTANT' turns are written and 'USER' turns are transcribed from the spoken recordings of crowdsourced workers. * 'segments': A list of various text spans with semantic annotations. * 'apis': An array of API invocations made during the utterance. Each API has the following structure: * 'name': The name of the API invoked (e.g. find\_movies). * 'index': The index of the parent utterance. * 'args': A 'list' of 'dict' with keys 'arg\_name' and 'arg\_value' which represent the name of the argument and the value for the argument respectively. * 'response': A 'list' of 'dict's with keys 'response\_name' and 'response\_value' which represent the name of the response and the value for the response respectively. Each segment has the following fields: * 'start\_index': The position of the start of the annotation in the utterance text. * 'end\_index': The position of the end of the annotation in the utterance text. * 'text': The raw text that has been annotated. * 'annotations': A list of annotation details for this segment. Each annotation has a single field: * 'name': The annotation name. ### Data Splits There are no deafults splits for all the config. The below table lists the number of examples in each config. Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information The dataset is licensed under 'Creative Commons Attribution 4.0 License' ### Contributions Thanks to @patil-suraj for adding this dataset.
[ "### Dataset Summary\n\n\nTaskmaster is dataset for goal oriented conversations. The Taskmaster-3 dataset consists of 23,757 movie ticketing dialogs.\nBy \"movie ticketing\" we mean conversations where the customer's goal is to purchase tickets after deciding\non theater, time, movie name, number of tickets, and date, or opt out of the transaction. This collection\nwas created using the \"self-dialog\" method. This means a single, crowd-sourced worker is\npaid to create a conversation writing turns for both speakers, i.e. the customer and the ticketing agent.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe dataset is in English language.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nA typical example looks like this", "### Data Fields\n\n\nEach conversation in the data file has the following structure:\n\n\n* 'conversation\\_id': A universally unique identifier with the prefix 'dlg-'. The ID has no meaning.\n* 'utterances': A list of utterances that make up the conversation.\n* 'instructions': Instructions for the crowdsourced worker used in creating the conversation.\n* 'vertical': In this dataset the vertical for all dialogs is \"Movie Tickets\".\n* 'scenario': This is the title of the instructions for each dialog.\n\n\nEach utterance has the following fields:\n\n\n* 'index': A 0-based index indicating the order of the utterances in the conversation.\n* 'speaker': Either USER or ASSISTANT, indicating which role generated this utterance.\n* 'text': The raw text of the utterance. In case of self dialogs (one\\_person\\_dialogs), this is written by the crowdsourced worker. In case of the WOz dialogs, 'ASSISTANT' turns are written and 'USER' turns are transcribed from the spoken recordings of crowdsourced workers.\n* 'segments': A list of various text spans with semantic annotations.\n* 'apis': An array of API invocations made during the utterance.\n\n\nEach API has the following structure:\n\n\n* 'name': The name of the API invoked (e.g. find\\_movies).\n* 'index': The index of the parent utterance.\n* 'args': A 'list' of 'dict' with keys 'arg\\_name' and 'arg\\_value' which represent the name of the argument and the value for the argument respectively.\n* 'response': A 'list' of 'dict's with keys 'response\\_name' and 'response\\_value' which represent the name of the response and the value for the response respectively.\n\n\nEach segment has the following fields:\n\n\n* 'start\\_index': The position of the start of the annotation in the utterance text.\n* 'end\\_index': The position of the end of the annotation in the utterance text.\n* 'text': The raw text that has been annotated.\n* 'annotations': A list of annotation details for this segment.\n\n\nEach annotation has a single field:\n\n\n* 'name': The annotation name.", "### Data Splits\n\n\nThere are no deafults splits for all the config. The below table lists the number of examples in each config.\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nThe dataset is licensed under 'Creative Commons Attribution 4.0 License'", "### Contributions\n\n\nThanks to @patil-suraj for adding this dataset." ]
[ "TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-dialogue-modeling #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-4.0 #arxiv-1909.05358 #region-us \n", "### Dataset Summary\n\n\nTaskmaster is dataset for goal oriented conversations. The Taskmaster-3 dataset consists of 23,757 movie ticketing dialogs.\nBy \"movie ticketing\" we mean conversations where the customer's goal is to purchase tickets after deciding\non theater, time, movie name, number of tickets, and date, or opt out of the transaction. This collection\nwas created using the \"self-dialog\" method. This means a single, crowd-sourced worker is\npaid to create a conversation writing turns for both speakers, i.e. the customer and the ticketing agent.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe dataset is in English language.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nA typical example looks like this", "### Data Fields\n\n\nEach conversation in the data file has the following structure:\n\n\n* 'conversation\\_id': A universally unique identifier with the prefix 'dlg-'. The ID has no meaning.\n* 'utterances': A list of utterances that make up the conversation.\n* 'instructions': Instructions for the crowdsourced worker used in creating the conversation.\n* 'vertical': In this dataset the vertical for all dialogs is \"Movie Tickets\".\n* 'scenario': This is the title of the instructions for each dialog.\n\n\nEach utterance has the following fields:\n\n\n* 'index': A 0-based index indicating the order of the utterances in the conversation.\n* 'speaker': Either USER or ASSISTANT, indicating which role generated this utterance.\n* 'text': The raw text of the utterance. In case of self dialogs (one\\_person\\_dialogs), this is written by the crowdsourced worker. In case of the WOz dialogs, 'ASSISTANT' turns are written and 'USER' turns are transcribed from the spoken recordings of crowdsourced workers.\n* 'segments': A list of various text spans with semantic annotations.\n* 'apis': An array of API invocations made during the utterance.\n\n\nEach API has the following structure:\n\n\n* 'name': The name of the API invoked (e.g. find\\_movies).\n* 'index': The index of the parent utterance.\n* 'args': A 'list' of 'dict' with keys 'arg\\_name' and 'arg\\_value' which represent the name of the argument and the value for the argument respectively.\n* 'response': A 'list' of 'dict's with keys 'response\\_name' and 'response\\_value' which represent the name of the response and the value for the response respectively.\n\n\nEach segment has the following fields:\n\n\n* 'start\\_index': The position of the start of the annotation in the utterance text.\n* 'end\\_index': The position of the end of the annotation in the utterance text.\n* 'text': The raw text that has been annotated.\n* 'annotations': A list of annotation details for this segment.\n\n\nEach annotation has a single field:\n\n\n* 'name': The annotation name.", "### Data Splits\n\n\nThere are no deafults splits for all the config. The below table lists the number of examples in each config.\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nThe dataset is licensed under 'Creative Commons Attribution 4.0 License'", "### Contributions\n\n\nThanks to @patil-suraj for adding this dataset." ]
[ 113, 134, 10, 19, 12, 561, 41, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 20, 19 ]
[ "passage: TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-dialogue-modeling #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-4.0 #arxiv-1909.05358 #region-us \n### Dataset Summary\n\n\nTaskmaster is dataset for goal oriented conversations. The Taskmaster-3 dataset consists of 23,757 movie ticketing dialogs.\nBy \"movie ticketing\" we mean conversations where the customer's goal is to purchase tickets after deciding\non theater, time, movie name, number of tickets, and date, or opt out of the transaction. This collection\nwas created using the \"self-dialog\" method. This means a single, crowd-sourced worker is\npaid to create a conversation writing turns for both speakers, i.e. the customer and the ticketing agent.### Supported Tasks and Leaderboards### Languages\n\n\nThe dataset is in English language.\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nA typical example looks like this" ]
00476f0f7e251c934e14f6e88c42a15e1b67c5a5
# Dataset Card for Tatoeba ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** http://opus.nlpl.eu/Tatoeba.php - **Repository:** None - **Paper:** http://www.lrec-conf.org/proceedings/lrec2012/pdf/463_Paper.pdf - **Leaderboard:** [More Information Needed] - **Point of Contact:** [More Information Needed] ### Dataset Summary Tatoeba is a collection of sentences and translations. To load a language pair which isn't part of the config, all you need to do is specify the language code as pairs. You can find the valid pairs in Homepage section of Dataset Description: http://opus.nlpl.eu/Tatoeba.php E.g. `dataset = load_dataset("tatoeba", lang1="en", lang2="he")` The default date is v2021-07-22, but you can also change the date with `dataset = load_dataset("tatoeba", lang1="en", lang2="he", date="v2020-11-09")` ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The languages in the dataset are: - ab - acm - ady - af - afb - afh - aii - ain - ajp - akl - aln - am - an - ang - aoz - apc - ar - arq - ary - arz - as - ast - avk - awa - ayl - az - ba - bal - bar - be - ber - bg - bho - bjn - bm - bn - bo - br - brx - bs - bua - bvy - bzt - ca - cay - cbk - ce - ceb - ch - chg - chn - cho - chr - cjy - ckb - ckt - cmn - co - code - cpi - crh - crk - cs - csb - cv - cy - da - de - dng - drt - dsb - dtp - dv - dws - ee - egl - el - emx - en - enm - eo - es - et - eu - ext - fi - fj - fkv - fo - fr - frm - fro - frr - fuc - fur - fuv - fy - ga - gag - gan - gbm - gcf - gd - gil - gl - gn - gom - gos - got - grc - gsw - gu - gv - ha - hak - haw - hbo - he - hi - hif - hil - hnj - hoc - hr - hrx - hsb - hsn - ht - hu - hy - ia - iba - id - ie - ig - ii - ike - ilo - io - is - it - izh - ja - jam - jbo - jdt - jpa - jv - ka - kaa - kab - kam - kek - kha - kjh - kk - kl - km - kmr - kn - ko - koi - kpv - krc - krl - ksh - ku - kum - kw - kxi - ky - kzj: Coastal Kadazan (deprecated tag; preferred value: Kadazan Dusun; Central Dusun (`dtp`)) - la - laa - lad - lb - ldn - lfn - lg - lij - liv - lkt - lld - lmo - ln - lo - lt - ltg - lut - lv - lzh - lzz - mad - mai - max - mdf - mfe - mg - mgm - mh - mhr - mi - mic - min - mk - ml - mn - mni - mnw - moh - mr - mt - mvv - mwl - mww - my - myv - na - nah - nan - nb - nch - nds - ngt - ngu - niu - nl - nlv - nn - nog - non - nov - npi - nst - nus - nv - ny - nys - oar - oc - ofs - ood - or - orv - os - osp - ota - otk - pa - pag - pal - pam - pap - pau - pcd - pdc - pes - phn - pi - pl - pms - pnb - ppl - prg - ps - pt - qu - quc - qya - rap - rif - rm - rn - ro - rom - ru - rue - rw - sa - sah - sc - scn - sco - sd - sdh - se - sg - sgs - shs - shy - si - sjn - sl - sm - sma - sn - so - sq - sr - stq - su - sux - sv - swg - swh - syc - ta - te - tet - tg - th - thv - ti - tig - tk - tl - tlh - tly - tmr - tmw - tn - to - toi - tok - tpi - tpw - tr - ts - tt - tts - tvl - ty - tyv - tzl - udm - ug - uk - umb - ur - uz - vec - vep - vi - vo - vro - wa - war - wo - wuu - xal - xh - xqa - yi - yo - yue - zlm - zsm - zu - zza ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@abhishekkrthakur](https://github.com/abhishekkrthakur) for adding this dataset.
tatoeba
[ "task_categories:translation", "annotations_creators:found", "language_creators:found", "multilinguality:multilingual", "size_categories:10K<n<100K", "source_datasets:original", "language:ab", "language:acm", "language:ady", "language:af", "language:afb", "language:afh", "language:aii", "language:ain", "language:ajp", "language:akl", "language:aln", "language:am", "language:an", "language:ang", "language:aoz", "language:apc", "language:ar", "language:arq", "language:ary", "language:arz", "language:as", "language:ast", "language:avk", "language:awa", "language:ayl", "language:az", "language:ba", "language:bal", "language:bar", "language:be", "language:ber", "language:bg", "language:bho", "language:bjn", "language:bm", "language:bn", "language:bo", "language:br", "language:brx", "language:bs", "language:bua", "language:bvy", "language:bzt", "language:ca", "language:cay", "language:cbk", "language:ce", "language:ceb", "language:ch", "language:chg", "language:chn", "language:cho", "language:chr", "language:cjy", "language:ckb", "language:ckt", "language:cmn", "language:co", "language:code", "language:cpi", "language:crh", "language:crk", "language:cs", "language:csb", "language:cv", "language:cy", "language:da", "language:de", "language:dng", "language:drt", "language:dsb", "language:dtp", "language:dv", "language:dws", "language:ee", "language:egl", "language:el", "language:emx", "language:en", "language:enm", "language:eo", "language:es", "language:et", "language:eu", "language:ext", "language:fi", "language:fj", "language:fkv", "language:fo", "language:fr", "language:frm", "language:fro", "language:frr", "language:fuc", "language:fur", "language:fuv", "language:fy", "language:ga", "language:gag", "language:gan", "language:gbm", "language:gcf", "language:gd", "language:gil", "language:gl", "language:gn", "language:gom", "language:gos", "language:got", "language:grc", "language:gsw", "language:gu", "language:gv", "language:ha", "language:hak", "language:haw", "language:hbo", "language:he", "language:hi", "language:hif", "language:hil", "language:hnj", "language:hoc", "language:hr", "language:hrx", "language:hsb", "language:hsn", "language:ht", "language:hu", "language:hy", "language:ia", "language:iba", "language:id", "language:ie", "language:ig", "language:ii", "language:ike", "language:ilo", "language:io", "language:is", "language:it", "language:izh", "language:ja", "language:jam", "language:jbo", "language:jdt", "language:jpa", "language:jv", "language:ka", "language:kaa", "language:kab", "language:kam", "language:kek", "language:kha", "language:kjh", "language:kk", "language:kl", "language:km", "language:kmr", "language:kn", "language:ko", "language:koi", "language:kpv", "language:krc", "language:krl", "language:ksh", "language:ku", "language:kum", "language:kw", "language:kxi", "language:ky", "language:la", "language:laa", "language:lad", "language:lb", "language:ldn", "language:lfn", "language:lg", "language:lij", "language:liv", "language:lkt", "language:lld", "language:lmo", "language:ln", "language:lo", "language:lt", "language:ltg", "language:lut", "language:lv", "language:lzh", "language:lzz", "language:mad", "language:mai", "language:max", "language:mdf", "language:mfe", "language:mg", "language:mgm", "language:mh", "language:mhr", "language:mi", "language:mic", "language:min", "language:mk", "language:ml", "language:mn", "language:mni", "language:mnw", "language:moh", "language:mr", "language:mt", "language:mvv", "language:mwl", "language:mww", "language:my", "language:myv", "language:na", "language:nah", "language:nan", "language:nb", "language:nch", "language:nds", "language:ngt", "language:ngu", "language:niu", "language:nl", "language:nlv", "language:nn", "language:nog", "language:non", "language:nov", "language:npi", "language:nst", "language:nus", "language:nv", "language:ny", "language:nys", "language:oar", "language:oc", "language:ofs", "language:ood", "language:or", "language:orv", "language:os", "language:osp", "language:ota", "language:otk", "language:pa", "language:pag", "language:pal", "language:pam", "language:pap", "language:pau", "language:pcd", "language:pdc", "language:pes", "language:phn", "language:pi", "language:pl", "language:pms", "language:pnb", "language:ppl", "language:prg", "language:ps", "language:pt", "language:qu", "language:quc", "language:qya", "language:rap", "language:rif", "language:rm", "language:rn", "language:ro", "language:rom", "language:ru", "language:rue", "language:rw", "language:sa", "language:sah", "language:sc", "language:scn", "language:sco", "language:sd", "language:sdh", "language:se", "language:sg", "language:sgs", "language:shs", "language:shy", "language:si", "language:sjn", "language:sl", "language:sm", "language:sma", "language:sn", "language:so", "language:sq", "language:sr", "language:stq", "language:su", "language:sux", "language:sv", "language:swg", "language:swh", "language:syc", "language:ta", "language:te", "language:tet", "language:tg", "language:th", "language:thv", "language:ti", "language:tig", "language:tk", "language:tl", "language:tlh", "language:tly", "language:tmr", "language:tmw", "language:tn", "language:to", "language:toi", "language:tok", "language:tpi", "language:tpw", "language:tr", "language:ts", "language:tt", "language:tts", "language:tvl", "language:ty", "language:tyv", "language:tzl", "language:udm", "language:ug", "language:uk", "language:umb", "language:ur", "language:uz", "language:vec", "language:vep", "language:vi", "language:vo", "language:vro", "language:wa", "language:war", "language:wo", "language:wuu", "language:xal", "language:xh", "language:xqa", "language:yi", "language:yo", "language:yue", "language:zlm", "language:zsm", "language:zu", "language:zza", "license:cc-by-2.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["ab", "acm", "ady", "af", "afb", "afh", "aii", "ain", "ajp", "akl", "aln", "am", "an", "ang", "aoz", "apc", "ar", "arq", "ary", "arz", "as", "ast", "avk", "awa", "ayl", "az", "ba", "bal", "bar", "be", "ber", "bg", "bho", "bjn", "bm", "bn", "bo", "br", "brx", "bs", "bua", "bvy", "bzt", "ca", "cay", "cbk", "ce", "ceb", "ch", "chg", "chn", "cho", "chr", "cjy", "ckb", "ckt", "cmn", "co", "code", "cpi", "crh", "crk", "cs", "csb", "cv", "cy", "da", "de", "dng", "drt", "dsb", "dtp", "dv", "dws", "ee", "egl", "el", "emx", "en", "enm", "eo", "es", "et", "eu", "ext", "fi", "fj", "fkv", "fo", "fr", "frm", "fro", "frr", "fuc", "fur", "fuv", "fy", "ga", "gag", "gan", "gbm", "gcf", "gd", "gil", "gl", "gn", "gom", "gos", "got", "grc", "gsw", "gu", "gv", "ha", "hak", "haw", "hbo", "he", "hi", "hif", "hil", "hnj", "hoc", "hr", "hrx", "hsb", "hsn", "ht", "hu", "hy", "ia", "iba", "id", "ie", "ig", "ii", "ike", "ilo", "io", "is", "it", "izh", "ja", "jam", "jbo", "jdt", "jpa", "jv", "ka", "kaa", "kab", "kam", "kek", "kha", "kjh", "kk", "kl", "km", "kmr", "kn", "ko", "koi", "kpv", "krc", "krl", "ksh", "ku", "kum", "kw", "kxi", "ky", "la", "laa", "lad", "lb", "ldn", "lfn", "lg", "lij", "liv", "lkt", "lld", "lmo", "ln", "lo", "lt", "ltg", "lut", "lv", "lzh", "lzz", "mad", "mai", "max", "mdf", "mfe", "mg", "mgm", "mh", "mhr", "mi", "mic", "min", "mk", "ml", "mn", "mni", "mnw", "moh", "mr", "mt", "mvv", "mwl", "mww", "my", "myv", "na", "nah", "nan", "nb", "nch", "nds", "ngt", "ngu", "niu", "nl", "nlv", "nn", "nog", "non", "nov", "npi", "nst", "nus", "nv", "ny", "nys", "oar", "oc", "ofs", "ood", "or", "orv", "os", "osp", "ota", "otk", "pa", "pag", "pal", "pam", "pap", "pau", "pcd", "pdc", "pes", "phn", "pi", "pl", "pms", "pnb", "ppl", "prg", "ps", "pt", "qu", "quc", "qya", "rap", "rif", "rm", "rn", "ro", "rom", "ru", "rue", "rw", "sa", "sah", "sc", "scn", "sco", "sd", "sdh", "se", "sg", "sgs", "shs", "shy", "si", "sjn", "sl", "sm", "sma", "sn", "so", "sq", "sr", "stq", "su", "sux", "sv", "swg", "swh", "syc", "ta", "te", "tet", "tg", "th", "thv", "ti", "tig", "tk", "tl", "tlh", "tly", "tmr", "tmw", "tn", "to", "toi", "tok", "tpi", "tpw", "tr", "ts", "tt", "tts", "tvl", "ty", "tyv", "tzl", "udm", "ug", "uk", "umb", "ur", "uz", "vec", "vep", "vi", "vo", "vro", "wa", "war", "wo", "wuu", "xal", "xh", "xqa", "yi", "yo", "yue", "zlm", "zsm", "zu", "zza"], "license": ["cc-by-2.0"], "multilinguality": ["multilingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["translation"], "task_ids": [], "paperswithcode_id": "tatoeba", "pretty_name": "Tatoeba", "dataset_info": [{"config_name": "en-mr", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "mr"]}}}], "splits": [{"name": "train", "num_bytes": 6190484, "num_examples": 53462}], "download_size": 1436200, "dataset_size": 6190484}, {"config_name": "eo-nl", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["eo", "nl"]}}}], "splits": [{"name": "train", "num_bytes": 8150048, "num_examples": 93650}], "download_size": 3020382, "dataset_size": 8150048}, {"config_name": "es-pt", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["es", "pt"]}}}], "splits": [{"name": "train", "num_bytes": 6180464, "num_examples": 67782}], "download_size": 2340361, "dataset_size": 6180464}, {"config_name": "fr-ru", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["fr", "ru"]}}}], "splits": [{"name": "train", "num_bytes": 19775390, "num_examples": 195161}], "download_size": 5509784, "dataset_size": 19775390}, {"config_name": "es-gl", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["es", "gl"]}}}], "splits": [{"name": "train", "num_bytes": 287683, "num_examples": 3135}], "download_size": 128506, "dataset_size": 287683}]}
2024-01-18T11:16:48+00:00
[]
[ "ab", "acm", "ady", "af", "afb", "afh", "aii", "ain", "ajp", "akl", "aln", "am", "an", "ang", "aoz", "apc", "ar", "arq", "ary", "arz", "as", "ast", "avk", "awa", "ayl", "az", "ba", "bal", "bar", "be", "ber", "bg", "bho", "bjn", "bm", "bn", "bo", "br", "brx", "bs", "bua", "bvy", "bzt", "ca", "cay", "cbk", "ce", "ceb", "ch", "chg", "chn", "cho", "chr", "cjy", "ckb", "ckt", "cmn", "co", "code", "cpi", "crh", "crk", "cs", "csb", "cv", "cy", "da", "de", "dng", "drt", "dsb", "dtp", "dv", "dws", "ee", "egl", "el", "emx", "en", "enm", "eo", "es", "et", "eu", "ext", "fi", "fj", "fkv", "fo", "fr", "frm", "fro", "frr", "fuc", "fur", "fuv", "fy", "ga", "gag", "gan", "gbm", "gcf", "gd", "gil", "gl", "gn", "gom", "gos", "got", "grc", "gsw", "gu", "gv", "ha", "hak", "haw", "hbo", "he", "hi", "hif", "hil", "hnj", "hoc", "hr", "hrx", "hsb", "hsn", "ht", "hu", "hy", "ia", "iba", "id", "ie", "ig", "ii", "ike", "ilo", "io", "is", "it", "izh", "ja", "jam", "jbo", "jdt", "jpa", "jv", "ka", "kaa", "kab", "kam", "kek", "kha", "kjh", "kk", "kl", "km", "kmr", "kn", "ko", "koi", "kpv", "krc", "krl", "ksh", "ku", "kum", "kw", "kxi", "ky", "la", "laa", "lad", "lb", "ldn", "lfn", "lg", "lij", "liv", "lkt", "lld", "lmo", "ln", "lo", "lt", "ltg", "lut", "lv", "lzh", "lzz", "mad", "mai", "max", "mdf", "mfe", "mg", "mgm", "mh", "mhr", "mi", "mic", "min", "mk", "ml", "mn", "mni", "mnw", "moh", "mr", "mt", "mvv", "mwl", "mww", "my", "myv", "na", "nah", "nan", "nb", "nch", "nds", "ngt", "ngu", "niu", "nl", "nlv", "nn", "nog", "non", "nov", "npi", "nst", "nus", "nv", "ny", "nys", "oar", "oc", "ofs", "ood", "or", "orv", "os", "osp", "ota", "otk", "pa", "pag", "pal", "pam", "pap", "pau", "pcd", "pdc", "pes", "phn", "pi", "pl", "pms", "pnb", "ppl", "prg", "ps", "pt", "qu", "quc", "qya", "rap", "rif", "rm", "rn", "ro", "rom", "ru", "rue", "rw", "sa", "sah", "sc", "scn", "sco", "sd", "sdh", "se", "sg", "sgs", "shs", "shy", "si", "sjn", "sl", "sm", "sma", "sn", "so", "sq", "sr", "stq", "su", "sux", "sv", "swg", "swh", "syc", "ta", "te", "tet", "tg", "th", "thv", "ti", "tig", "tk", "tl", "tlh", "tly", "tmr", "tmw", "tn", "to", "toi", "tok", "tpi", "tpw", "tr", "ts", "tt", "tts", "tvl", "ty", "tyv", "tzl", "udm", "ug", "uk", "umb", "ur", "uz", "vec", "vep", "vi", "vo", "vro", "wa", "war", "wo", "wuu", "xal", "xh", "xqa", "yi", "yo", "yue", "zlm", "zsm", "zu", "zza" ]
TAGS #task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-10K<n<100K #source_datasets-original #language-Abkhazian #language-Mesopotamian Arabic #language-Adyghe #language-Afrikaans #language-Gulf Arabic #language-Afrihili #language-Assyrian Neo-Aramaic #language-Ainu (Japan) #language-South Levantine Arabic #language-Aklanon #language-Gheg Albanian #language-Amharic #language-Aragonese #language-Old English (ca. 450-1100) #language-Uab Meto #language-Levantine Arabic #language-Arabic #language-Algerian Arabic #language-Moroccan Arabic #language-Egyptian Arabic #language-Assamese #language-Asturian #language-Kotava #language-Awadhi #language-Libyan Arabic #language-Azerbaijani #language-Bashkir #language-Baluchi #language-Bavarian #language-Belarusian #language-ber #language-Bulgarian #language-Bhojpuri #language-Banjar #language-Bambara #language-Bengali #language-Tibetan #language-Breton #language-Bodo (India) #language-Bosnian #language-Buriat #language-Baybayanon #language-Brithenig #language-Catalan #language-Cayuga #language-Chavacano #language-Chechen #language-Cebuano #language-Chamorro #language-Chagatai #language-Chinook jargon #language-Choctaw #language-Cherokee #language-Jinyu Chinese #language-Central Kurdish #language-Chukot #language-Mandarin Chinese #language-Corsican #language-code #language-Chinese Pidgin English #language-Crimean Tatar #language-Plains Cree #language-Czech #language-Kashubian #language-Chuvash #language-Welsh #language-Danish #language-German #language-Dungan #language-Drents #language-Lower Sorbian #language-Kadazan Dusun #language-Dhivehi #language-Dutton World Speedwords #language-Ewe #language-Emilian #language-Modern Greek (1453-) #language-Erromintxela #language-English #language-Middle English (1100-1500) #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Extremaduran #language-Finnish #language-Fijian #language-Kven Finnish #language-Faroese #language-French #language-Middle French (ca. 1400-1600) #language-Old French (842-ca. 1400) #language-Northern Frisian #language-Pulaar #language-Friulian #language-Nigerian Fulfulde #language-Western Frisian #language-Irish #language-Gagauz #language-Gan Chinese #language-Garhwali #language-Guadeloupean Creole French #language-Scottish Gaelic #language-Gilbertese #language-Galician #language-Guarani #language-Goan Konkani #language-Gronings #language-Gothic #language-Ancient Greek (to 1453) #language-Swiss German #language-Gujarati #language-Manx #language-Hausa #language-Hakka Chinese #language-Hawaiian #language-Ancient Hebrew #language-Hebrew #language-Hindi #language-Fiji Hindi #language-Hiligaynon #language-Hmong Njua #language-Ho #language-Croatian #language-Hunsrik #language-Upper Sorbian #language-Xiang Chinese #language-Haitian #language-Hungarian #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Iban #language-Indonesian #language-Interlingue #language-Igbo #language-Sichuan Yi #language-Eastern Canadian Inuktitut #language-Iloko #language-Ido #language-Icelandic #language-Italian #language-Ingrian #language-Japanese #language-Jamaican Creole English #language-Lojban #language-Judeo-Tat #language-Jewish Palestinian Aramaic #language-Javanese #language-Georgian #language-Kara-Kalpak #language-Kabyle #language-Kamba (Kenya) #language-Kekchí #language-Khasi #language-Khakas #language-Kazakh #language-Kalaallisut #language-Khmer #language-Northern Kurdish #language-Kannada #language-Korean #language-Komi-Permyak #language-Komi-Zyrian #language-Karachay-Balkar #language-Karelian #language-Kölsch #language-Kurdish #language-Kumyk #language-Cornish #language-Keningau Murut #language-Kirghiz #language-Latin #language-Southern Subanen #language-Ladino #language-Luxembourgish #language-Láadan #language-Lingua Franca Nova #language-Ganda #language-Ligurian #language-Liv #language-Lakota #language-Ladin #language-Lombard #language-Lingala #language-Lao #language-Lithuanian #language-Latgalian #language-Lushootseed #language-Latvian #language-Literary Chinese #language-Laz #language-Madurese #language-Maithili #language-North Moluccan Malay #language-Moksha #language-Morisyen #language-Malagasy #language-Mambae #language-Marshallese #language-Eastern Mari #language-Maori #language-Mi'kmaq #language-Minangkabau #language-Macedonian #language-Malayalam #language-Mongolian #language-Manipuri #language-Mon #language-Mohawk #language-Marathi #language-Maltese #language-Tagal Murut #language-Mirandese #language-Hmong Daw #language-Burmese #language-Erzya #language-Nauru #language-nah #language-Min Nan Chinese #language-Norwegian Bokmål #language-Central Huasteca Nahuatl #language-Low German #language-Kriang #language-Guerrero Nahuatl #language-Niuean #language-Dutch #language-Orizaba Nahuatl #language-Norwegian Nynorsk #language-Nogai #language-Old Norse #language-Novial #language-Nepali (individual language) #language-Tase Naga #language-Nuer #language-Navajo #language-Nyanja #language-Nyungar #language-Old Aramaic (up to 700 BCE) #language-Occitan (post 1500) #language-Old Frisian #language-Tohono O'odham #language-Oriya (macrolanguage) #language-Old Russian #language-Ossetian #language-Old Spanish #language-Ottoman Turkish (1500-1928) #language-Old Turkish #language-Panjabi #language-Pangasinan #language-Pahlavi #language-Pampanga #language-Papiamento #language-Palauan #language-Picard #language-Pennsylvania German #language-Iranian Persian #language-Phoenician #language-Pali #language-Polish #language-Piemontese #language-Western Panjabi #language-Pipil #language-Prussian #language-Pushto #language-Portuguese #language-Quechua #language-K'iche' #language-Quenya #language-Rapanui #language-Tarifit #language-Romansh #language-Rundi #language-Romanian #language-Romany #language-Russian #language-Rusyn #language-Kinyarwanda #language-Sanskrit #language-Yakut #language-Sardinian #language-Sicilian #language-Scots #language-Sindhi #language-Southern Kurdish #language-Northern Sami #language-Sango #language-Samogitian #language-Shuswap #language-Tachawit #language-Sinhala #language-Sindarin #language-Slovenian #language-Samoan #language-Southern Sami #language-Shona #language-Somali #language-Albanian #language-Serbian #language-Saterfriesisch #language-Sundanese #language-Sumerian #language-Swedish #language-Swabian #language-Swahili (individual language) #language-Classical Syriac #language-Tamil #language-Telugu #language-Tetum #language-Tajik #language-Thai #language-Tahaggart Tamahaq #language-Tigrinya #language-Tigre #language-Turkmen #language-Tagalog #language-Klingon #language-Talysh #language-Jewish Babylonian Aramaic (ca. 200-1200 CE) #language-Temuan #language-Tswana #language-Tonga (Tonga Islands) #language-Tonga (Zambia) #language-Toki Pona #language-Tok Pisin #language-tpw #language-Turkish #language-Tsonga #language-Tatar #language-Northeastern Thai #language-Tuvalu #language-Tahitian #language-Tuvinian #language-Talossan #language-Udmurt #language-Uighur #language-Ukrainian #language-Umbundu #language-Urdu #language-Uzbek #language-Venetian #language-Veps #language-Vietnamese #language-Volapük #language-Võro #language-Walloon #language-Waray (Philippines) #language-Wolof #language-Wu Chinese #language-Kalmyk #language-Xhosa #language-Karakhanid #language-Yiddish #language-Yoruba #language-Yue Chinese #language-Malay (individual language) #language-Standard Malay #language-Zulu #language-Zaza #license-cc-by-2.0 #region-us
# Dataset Card for Tatoeba ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: None - Paper: URL - Leaderboard: - Point of Contact: ### Dataset Summary Tatoeba is a collection of sentences and translations. To load a language pair which isn't part of the config, all you need to do is specify the language code as pairs. You can find the valid pairs in Homepage section of Dataset Description: URL E.g. 'dataset = load_dataset("tatoeba", lang1="en", lang2="he")' The default date is v2021-07-22, but you can also change the date with 'dataset = load_dataset("tatoeba", lang1="en", lang2="he", date="v2020-11-09")' ### Supported Tasks and Leaderboards ### Languages The languages in the dataset are: - ab - acm - ady - af - afb - afh - aii - ain - ajp - akl - aln - am - an - ang - aoz - apc - ar - arq - ary - arz - as - ast - avk - awa - ayl - az - ba - bal - bar - be - ber - bg - bho - bjn - bm - bn - bo - br - brx - bs - bua - bvy - bzt - ca - cay - cbk - ce - ceb - ch - chg - chn - cho - chr - cjy - ckb - ckt - cmn - co - code - cpi - crh - crk - cs - csb - cv - cy - da - de - dng - drt - dsb - dtp - dv - dws - ee - egl - el - emx - en - enm - eo - es - et - eu - ext - fi - fj - fkv - fo - fr - frm - fro - frr - fuc - fur - fuv - fy - ga - gag - gan - gbm - gcf - gd - gil - gl - gn - gom - gos - got - grc - gsw - gu - gv - ha - hak - haw - hbo - he - hi - hif - hil - hnj - hoc - hr - hrx - hsb - hsn - ht - hu - hy - ia - iba - id - ie - ig - ii - ike - ilo - io - is - it - izh - ja - jam - jbo - jdt - jpa - jv - ka - kaa - kab - kam - kek - kha - kjh - kk - kl - km - kmr - kn - ko - koi - kpv - krc - krl - ksh - ku - kum - kw - kxi - ky - kzj: Coastal Kadazan (deprecated tag; preferred value: Kadazan Dusun; Central Dusun ('dtp')) - la - laa - lad - lb - ldn - lfn - lg - lij - liv - lkt - lld - lmo - ln - lo - lt - ltg - lut - lv - lzh - lzz - mad - mai - max - mdf - mfe - mg - mgm - mh - mhr - mi - mic - min - mk - ml - mn - mni - mnw - moh - mr - mt - mvv - mwl - mww - my - myv - na - nah - nan - nb - nch - nds - ngt - ngu - niu - nl - nlv - nn - nog - non - nov - npi - nst - nus - nv - ny - nys - oar - oc - ofs - ood - or - orv - os - osp - ota - otk - pa - pag - pal - pam - pap - pau - pcd - pdc - pes - phn - pi - pl - pms - pnb - ppl - prg - ps - pt - qu - quc - qya - rap - rif - rm - rn - ro - rom - ru - rue - rw - sa - sah - sc - scn - sco - sd - sdh - se - sg - sgs - shs - shy - si - sjn - sl - sm - sma - sn - so - sq - sr - stq - su - sux - sv - swg - swh - syc - ta - te - tet - tg - th - thv - ti - tig - tk - tl - tlh - tly - tmr - tmw - tn - to - toi - tok - tpi - tpw - tr - ts - tt - tts - tvl - ty - tyv - tzl - udm - ug - uk - umb - ur - uz - vec - vep - vi - vo - vro - wa - war - wo - wuu - xal - xh - xqa - yi - yo - yue - zlm - zsm - zu - zza ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @abhishekkrthakur for adding this dataset.
[ "# Dataset Card for Tatoeba", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: URL\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\nTatoeba is a collection of sentences and translations.\n\nTo load a language pair which isn't part of the config, all you need to do is specify the language code as pairs.\nYou can find the valid pairs in Homepage section of Dataset Description: URL\nE.g.\n\n'dataset = load_dataset(\"tatoeba\", lang1=\"en\", lang2=\"he\")'\n\nThe default date is v2021-07-22, but you can also change the date with\n\n'dataset = load_dataset(\"tatoeba\", lang1=\"en\", lang2=\"he\", date=\"v2020-11-09\")'", "### Supported Tasks and Leaderboards", "### Languages\n\nThe languages in the dataset are:\n- ab\n- acm\n- ady\n- af\n- afb\n- afh\n- aii\n- ain\n- ajp\n- akl\n- aln\n- am\n- an\n- ang\n- aoz\n- apc\n- ar\n- arq\n- ary\n- arz\n- as\n- ast\n- avk\n- awa\n- ayl\n- az\n- ba\n- bal\n- bar\n- be\n- ber\n- bg\n- bho\n- bjn\n- bm\n- bn\n- bo\n- br\n- brx\n- bs\n- bua\n- bvy\n- bzt\n- ca\n- cay\n- cbk\n- ce\n- ceb\n- ch\n- chg\n- chn\n- cho\n- chr\n- cjy\n- ckb\n- ckt\n- cmn\n- co\n- code\n- cpi\n- crh\n- crk\n- cs\n- csb\n- cv\n- cy\n- da\n- de\n- dng\n- drt\n- dsb\n- dtp\n- dv\n- dws\n- ee\n- egl\n- el\n- emx\n- en\n- enm\n- eo\n- es\n- et\n- eu\n- ext\n- fi\n- fj\n- fkv\n- fo\n- fr\n- frm\n- fro\n- frr\n- fuc\n- fur\n- fuv\n- fy\n- ga\n- gag\n- gan\n- gbm\n- gcf\n- gd\n- gil\n- gl\n- gn\n- gom\n- gos\n- got\n- grc\n- gsw\n- gu\n- gv\n- ha\n- hak\n- haw\n- hbo\n- he\n- hi\n- hif\n- hil\n- hnj\n- hoc\n- hr\n- hrx\n- hsb\n- hsn\n- ht\n- hu\n- hy\n- ia\n- iba\n- id\n- ie\n- ig\n- ii\n- ike\n- ilo\n- io\n- is\n- it\n- izh\n- ja\n- jam\n- jbo\n- jdt\n- jpa\n- jv\n- ka\n- kaa\n- kab\n- kam\n- kek\n- kha\n- kjh\n- kk\n- kl\n- km\n- kmr\n- kn\n- ko\n- koi\n- kpv\n- krc\n- krl\n- ksh\n- ku\n- kum\n- kw\n- kxi\n- ky\n- kzj: Coastal Kadazan (deprecated tag; preferred value: Kadazan Dusun; Central Dusun ('dtp'))\n- la\n- laa\n- lad\n- lb\n- ldn\n- lfn\n- lg\n- lij\n- liv\n- lkt\n- lld\n- lmo\n- ln\n- lo\n- lt\n- ltg\n- lut\n- lv\n- lzh\n- lzz\n- mad\n- mai\n- max\n- mdf\n- mfe\n- mg\n- mgm\n- mh\n- mhr\n- mi\n- mic\n- min\n- mk\n- ml\n- mn\n- mni\n- mnw\n- moh\n- mr\n- mt\n- mvv\n- mwl\n- mww\n- my\n- myv\n- na\n- nah\n- nan\n- nb\n- nch\n- nds\n- ngt\n- ngu\n- niu\n- nl\n- nlv\n- nn\n- nog\n- non\n- nov\n- npi\n- nst\n- nus\n- nv\n- ny\n- nys\n- oar\n- oc\n- ofs\n- ood\n- or\n- orv\n- os\n- osp\n- ota\n- otk\n- pa\n- pag\n- pal\n- pam\n- pap\n- pau\n- pcd\n- pdc\n- pes\n- phn\n- pi\n- pl\n- pms\n- pnb\n- ppl\n- prg\n- ps\n- pt\n- qu\n- quc\n- qya\n- rap\n- rif\n- rm\n- rn\n- ro\n- rom\n- ru\n- rue\n- rw\n- sa\n- sah\n- sc\n- scn\n- sco\n- sd\n- sdh\n- se\n- sg\n- sgs\n- shs\n- shy\n- si\n- sjn\n- sl\n- sm\n- sma\n- sn\n- so\n- sq\n- sr\n- stq\n- su\n- sux\n- sv\n- swg\n- swh\n- syc\n- ta\n- te\n- tet\n- tg\n- th\n- thv\n- ti\n- tig\n- tk\n- tl\n- tlh\n- tly\n- tmr\n- tmw\n- tn\n- to\n- toi\n- tok\n- tpi\n- tpw\n- tr\n- ts\n- tt\n- tts\n- tvl\n- ty\n- tyv\n- tzl\n- udm\n- ug\n- uk\n- umb\n- ur\n- uz\n- vec\n- vep\n- vi\n- vo\n- vro\n- wa\n- war\n- wo\n- wuu\n- xal\n- xh\n- xqa\n- yi\n- yo\n- yue\n- zlm\n- zsm\n- zu\n- zza", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ "TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-10K<n<100K #source_datasets-original #language-Abkhazian #language-Mesopotamian Arabic #language-Adyghe #language-Afrikaans #language-Gulf Arabic #language-Afrihili #language-Assyrian Neo-Aramaic #language-Ainu (Japan) #language-South Levantine Arabic #language-Aklanon #language-Gheg Albanian #language-Amharic #language-Aragonese #language-Old English (ca. 450-1100) #language-Uab Meto #language-Levantine Arabic #language-Arabic #language-Algerian Arabic #language-Moroccan Arabic #language-Egyptian Arabic #language-Assamese #language-Asturian #language-Kotava #language-Awadhi #language-Libyan Arabic #language-Azerbaijani #language-Bashkir #language-Baluchi #language-Bavarian #language-Belarusian #language-ber #language-Bulgarian #language-Bhojpuri #language-Banjar #language-Bambara #language-Bengali #language-Tibetan #language-Breton #language-Bodo (India) #language-Bosnian #language-Buriat #language-Baybayanon #language-Brithenig #language-Catalan #language-Cayuga #language-Chavacano #language-Chechen #language-Cebuano #language-Chamorro #language-Chagatai #language-Chinook jargon #language-Choctaw #language-Cherokee #language-Jinyu Chinese #language-Central Kurdish #language-Chukot #language-Mandarin Chinese #language-Corsican #language-code #language-Chinese Pidgin English #language-Crimean Tatar #language-Plains Cree #language-Czech #language-Kashubian #language-Chuvash #language-Welsh #language-Danish #language-German #language-Dungan #language-Drents #language-Lower Sorbian #language-Kadazan Dusun #language-Dhivehi #language-Dutton World Speedwords #language-Ewe #language-Emilian #language-Modern Greek (1453-) #language-Erromintxela #language-English #language-Middle English (1100-1500) #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Extremaduran #language-Finnish #language-Fijian #language-Kven Finnish #language-Faroese #language-French #language-Middle French (ca. 1400-1600) #language-Old French (842-ca. 1400) #language-Northern Frisian #language-Pulaar #language-Friulian #language-Nigerian Fulfulde #language-Western Frisian #language-Irish #language-Gagauz #language-Gan Chinese #language-Garhwali #language-Guadeloupean Creole French #language-Scottish Gaelic #language-Gilbertese #language-Galician #language-Guarani #language-Goan Konkani #language-Gronings #language-Gothic #language-Ancient Greek (to 1453) #language-Swiss German #language-Gujarati #language-Manx #language-Hausa #language-Hakka Chinese #language-Hawaiian #language-Ancient Hebrew #language-Hebrew #language-Hindi #language-Fiji Hindi #language-Hiligaynon #language-Hmong Njua #language-Ho #language-Croatian #language-Hunsrik #language-Upper Sorbian #language-Xiang Chinese #language-Haitian #language-Hungarian #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Iban #language-Indonesian #language-Interlingue #language-Igbo #language-Sichuan Yi #language-Eastern Canadian Inuktitut #language-Iloko #language-Ido #language-Icelandic #language-Italian #language-Ingrian #language-Japanese #language-Jamaican Creole English #language-Lojban #language-Judeo-Tat #language-Jewish Palestinian Aramaic #language-Javanese #language-Georgian #language-Kara-Kalpak #language-Kabyle #language-Kamba (Kenya) #language-Kekchí #language-Khasi #language-Khakas #language-Kazakh #language-Kalaallisut #language-Khmer #language-Northern Kurdish #language-Kannada #language-Korean #language-Komi-Permyak #language-Komi-Zyrian #language-Karachay-Balkar #language-Karelian #language-Kölsch #language-Kurdish #language-Kumyk #language-Cornish #language-Keningau Murut #language-Kirghiz #language-Latin #language-Southern Subanen #language-Ladino #language-Luxembourgish #language-Láadan #language-Lingua Franca Nova #language-Ganda #language-Ligurian #language-Liv #language-Lakota #language-Ladin #language-Lombard #language-Lingala #language-Lao #language-Lithuanian #language-Latgalian #language-Lushootseed #language-Latvian #language-Literary Chinese #language-Laz #language-Madurese #language-Maithili #language-North Moluccan Malay #language-Moksha #language-Morisyen #language-Malagasy #language-Mambae #language-Marshallese #language-Eastern Mari #language-Maori #language-Mi'kmaq #language-Minangkabau #language-Macedonian #language-Malayalam #language-Mongolian #language-Manipuri #language-Mon #language-Mohawk #language-Marathi #language-Maltese #language-Tagal Murut #language-Mirandese #language-Hmong Daw #language-Burmese #language-Erzya #language-Nauru #language-nah #language-Min Nan Chinese #language-Norwegian Bokmål #language-Central Huasteca Nahuatl #language-Low German #language-Kriang #language-Guerrero Nahuatl #language-Niuean #language-Dutch #language-Orizaba Nahuatl #language-Norwegian Nynorsk #language-Nogai #language-Old Norse #language-Novial #language-Nepali (individual language) #language-Tase Naga #language-Nuer #language-Navajo #language-Nyanja #language-Nyungar #language-Old Aramaic (up to 700 BCE) #language-Occitan (post 1500) #language-Old Frisian #language-Tohono O'odham #language-Oriya (macrolanguage) #language-Old Russian #language-Ossetian #language-Old Spanish #language-Ottoman Turkish (1500-1928) #language-Old Turkish #language-Panjabi #language-Pangasinan #language-Pahlavi #language-Pampanga #language-Papiamento #language-Palauan #language-Picard #language-Pennsylvania German #language-Iranian Persian #language-Phoenician #language-Pali #language-Polish #language-Piemontese #language-Western Panjabi #language-Pipil #language-Prussian #language-Pushto #language-Portuguese #language-Quechua #language-K'iche' #language-Quenya #language-Rapanui #language-Tarifit #language-Romansh #language-Rundi #language-Romanian #language-Romany #language-Russian #language-Rusyn #language-Kinyarwanda #language-Sanskrit #language-Yakut #language-Sardinian #language-Sicilian #language-Scots #language-Sindhi #language-Southern Kurdish #language-Northern Sami #language-Sango #language-Samogitian #language-Shuswap #language-Tachawit #language-Sinhala #language-Sindarin #language-Slovenian #language-Samoan #language-Southern Sami #language-Shona #language-Somali #language-Albanian #language-Serbian #language-Saterfriesisch #language-Sundanese #language-Sumerian #language-Swedish #language-Swabian #language-Swahili (individual language) #language-Classical Syriac #language-Tamil #language-Telugu #language-Tetum #language-Tajik #language-Thai #language-Tahaggart Tamahaq #language-Tigrinya #language-Tigre #language-Turkmen #language-Tagalog #language-Klingon #language-Talysh #language-Jewish Babylonian Aramaic (ca. 200-1200 CE) #language-Temuan #language-Tswana #language-Tonga (Tonga Islands) #language-Tonga (Zambia) #language-Toki Pona #language-Tok Pisin #language-tpw #language-Turkish #language-Tsonga #language-Tatar #language-Northeastern Thai #language-Tuvalu #language-Tahitian #language-Tuvinian #language-Talossan #language-Udmurt #language-Uighur #language-Ukrainian #language-Umbundu #language-Urdu #language-Uzbek #language-Venetian #language-Veps #language-Vietnamese #language-Volapük #language-Võro #language-Walloon #language-Waray (Philippines) #language-Wolof #language-Wu Chinese #language-Kalmyk #language-Xhosa #language-Karakhanid #language-Yiddish #language-Yoruba #language-Yue Chinese #language-Malay (individual language) #language-Standard Malay #language-Zulu #language-Zaza #license-cc-by-2.0 #region-us \n", "# Dataset Card for Tatoeba", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: URL\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\nTatoeba is a collection of sentences and translations.\n\nTo load a language pair which isn't part of the config, all you need to do is specify the language code as pairs.\nYou can find the valid pairs in Homepage section of Dataset Description: URL\nE.g.\n\n'dataset = load_dataset(\"tatoeba\", lang1=\"en\", lang2=\"he\")'\n\nThe default date is v2021-07-22, but you can also change the date with\n\n'dataset = load_dataset(\"tatoeba\", lang1=\"en\", lang2=\"he\", date=\"v2020-11-09\")'", "### Supported Tasks and Leaderboards", "### Languages\n\nThe languages in the dataset are:\n- ab\n- acm\n- ady\n- af\n- afb\n- afh\n- aii\n- ain\n- ajp\n- akl\n- aln\n- am\n- an\n- ang\n- aoz\n- apc\n- ar\n- arq\n- ary\n- arz\n- as\n- ast\n- avk\n- awa\n- ayl\n- az\n- ba\n- bal\n- bar\n- be\n- ber\n- bg\n- bho\n- bjn\n- bm\n- bn\n- bo\n- br\n- brx\n- bs\n- bua\n- bvy\n- bzt\n- ca\n- cay\n- cbk\n- ce\n- ceb\n- ch\n- chg\n- chn\n- cho\n- chr\n- cjy\n- ckb\n- ckt\n- cmn\n- co\n- code\n- cpi\n- crh\n- crk\n- cs\n- csb\n- cv\n- cy\n- da\n- de\n- dng\n- drt\n- dsb\n- dtp\n- dv\n- dws\n- ee\n- egl\n- el\n- emx\n- en\n- enm\n- eo\n- es\n- et\n- eu\n- ext\n- fi\n- fj\n- fkv\n- fo\n- fr\n- frm\n- fro\n- frr\n- fuc\n- fur\n- fuv\n- fy\n- ga\n- gag\n- gan\n- gbm\n- gcf\n- gd\n- gil\n- gl\n- gn\n- gom\n- gos\n- got\n- grc\n- gsw\n- gu\n- gv\n- ha\n- hak\n- haw\n- hbo\n- he\n- hi\n- hif\n- hil\n- hnj\n- hoc\n- hr\n- hrx\n- hsb\n- hsn\n- ht\n- hu\n- hy\n- ia\n- iba\n- id\n- ie\n- ig\n- ii\n- ike\n- ilo\n- io\n- is\n- it\n- izh\n- ja\n- jam\n- jbo\n- jdt\n- jpa\n- jv\n- ka\n- kaa\n- kab\n- kam\n- kek\n- kha\n- kjh\n- kk\n- kl\n- km\n- kmr\n- kn\n- ko\n- koi\n- kpv\n- krc\n- krl\n- ksh\n- ku\n- kum\n- kw\n- kxi\n- ky\n- kzj: Coastal Kadazan (deprecated tag; preferred value: Kadazan Dusun; Central Dusun ('dtp'))\n- la\n- laa\n- lad\n- lb\n- ldn\n- lfn\n- lg\n- lij\n- liv\n- lkt\n- lld\n- lmo\n- ln\n- lo\n- lt\n- ltg\n- lut\n- lv\n- lzh\n- lzz\n- mad\n- mai\n- max\n- mdf\n- mfe\n- mg\n- mgm\n- mh\n- mhr\n- mi\n- mic\n- min\n- mk\n- ml\n- mn\n- mni\n- mnw\n- moh\n- mr\n- mt\n- mvv\n- mwl\n- mww\n- my\n- myv\n- na\n- nah\n- nan\n- nb\n- nch\n- nds\n- ngt\n- ngu\n- niu\n- nl\n- nlv\n- nn\n- nog\n- non\n- nov\n- npi\n- nst\n- nus\n- nv\n- ny\n- nys\n- oar\n- oc\n- ofs\n- ood\n- or\n- orv\n- os\n- osp\n- ota\n- otk\n- pa\n- pag\n- pal\n- pam\n- pap\n- pau\n- pcd\n- pdc\n- pes\n- phn\n- pi\n- pl\n- pms\n- pnb\n- ppl\n- prg\n- ps\n- pt\n- qu\n- quc\n- qya\n- rap\n- rif\n- rm\n- rn\n- ro\n- rom\n- ru\n- rue\n- rw\n- sa\n- sah\n- sc\n- scn\n- sco\n- sd\n- sdh\n- se\n- sg\n- sgs\n- shs\n- shy\n- si\n- sjn\n- sl\n- sm\n- sma\n- sn\n- so\n- sq\n- sr\n- stq\n- su\n- sux\n- sv\n- swg\n- swh\n- syc\n- ta\n- te\n- tet\n- tg\n- th\n- thv\n- ti\n- tig\n- tk\n- tl\n- tlh\n- tly\n- tmr\n- tmw\n- tn\n- to\n- toi\n- tok\n- tpi\n- tpw\n- tr\n- ts\n- tt\n- tts\n- tvl\n- ty\n- tyv\n- tzl\n- udm\n- ug\n- uk\n- umb\n- ur\n- uz\n- vec\n- vep\n- vi\n- vo\n- vro\n- wa\n- war\n- wo\n- wuu\n- xal\n- xh\n- xqa\n- yi\n- yo\n- yue\n- zlm\n- zsm\n- zu\n- zza", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ 2384, 8, 120, 28, 143, 10, 974, 6, 6, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 20 ]
[ "passage: ", "passage: TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-10K<n<100K #source_datasets-original #language-Abkhazian #language-Mesopotamian Arabic #language-Adyghe #language-Afrikaans #language-Gulf Arabic #language-Afrihili #language-Assyrian Neo-Aramaic #language-Ainu (Japan) #language-South Levantine Arabic #language-Aklanon #language-Gheg Albanian #language-Amharic #language-Aragonese #language-Old English (ca. 450-1100) #language-Uab Meto #language-Levantine Arabic #language-Arabic #language-Algerian Arabic #language-Moroccan Arabic #language-Egyptian Arabic #language-Assamese #language-Asturian #language-Kotava #language-Awadhi #language-Libyan Arabic #language-Azerbaijani #language-Bashkir #language-Baluchi #language-Bavarian #language-Belarusian #language-ber #language-Bulgarian #language-Bhojpuri #language-Banjar #language-Bambara #language-Bengali #language-Tibetan #language-Breton #language-Bodo (India) #language-Bosnian #language-Buriat #language-Baybayanon #language-Brithenig #language-Catalan #language-Cayuga #language-Chavacano #language-Chechen #language-Cebuano #language-Chamorro #language-Chagatai #language-Chinook jargon #language-Choctaw #language-Cherokee #language-Jinyu Chinese #language-Central Kurdish #language-Chukot #language-Mandarin Chinese #language-Corsican #language-code #language-Chinese Pidgin English #language-Crimean Tatar #language-Plains Cree #language-Czech #language-Kashubian #language-Chuvash #language-Welsh #language-Danish #language-German #language-Dungan #language-Drents #language-Lower Sorbian #language-Kadazan Dusun #language-Dhivehi #language-Dutton World Speedwords #language-Ewe #language-Emilian #language-Modern Greek (1453-) #language-Erromintxela #language-English #language-Middle English (1100-1500) #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Extremaduran #language-Finnish #language-Fijian #language-Kven Finnish #language-Faroese #language-French #language-Middle French (ca. 1400-1600) #language-Old French (842-ca. 1400) #language-Northern Frisian #language-Pulaar #language-Friulian #language-Nigerian Fulfulde #language-Western Frisian #language-Irish #language-Gagauz #language-Gan Chinese #language-Garhwali #language-Guadeloupean Creole French #language-Scottish Gaelic #language-Gilbertese #language-Galician #language-Guarani #language-Goan Konkani #language-Gronings #language-Gothic #language-Ancient Greek (to 1453) #language-Swiss German #language-Gujarati #language-Manx #language-Hausa #language-Hakka Chinese #language-Hawaiian #language-Ancient Hebrew #language-Hebrew #language-Hindi #language-Fiji Hindi #language-Hiligaynon #language-Hmong Njua #language-Ho #language-Croatian #language-Hunsrik #language-Upper Sorbian #language-Xiang Chinese #language-Haitian #language-Hungarian #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Iban #language-Indonesian #language-Interlingue #language-Igbo #language-Sichuan Yi #language-Eastern Canadian Inuktitut #language-Iloko #language-Ido #language-Icelandic #language-Italian #language-Ingrian #language-Japanese #language-Jamaican Creole English #language-Lojban #language-Judeo-Tat #language-Jewish Palestinian Aramaic #language-Javanese #language-Georgian #language-Kara-Kalpak #language-Kabyle #language-Kamba (Kenya) #language-Kekchí #language-Khasi #language-Khakas #language-Kazakh #language-Kalaallisut #language-Khmer #language-Northern Kurdish #language-Kannada #language-Korean #language-Komi-Permyak #language-Komi-Zyrian #language-Karachay-Balkar #language-Karelian #language-Kölsch #language-Kurdish #language-Kumyk #language-Cornish #language-Keningau Murut #language-Kirghiz #language-Latin #language-Southern Subanen #language-Ladino #language-Luxembourgish #language-Láadan #language-Lingua Franca Nova #language-Ganda #language-Ligurian #language-Liv #language-Lakota #language-Ladin #language-Lombard #language-Lingala #language-Lao #language-Lithuanian #language-Latgalian #language-Lushootseed #language-Latvian #language-Literary Chinese #language-Laz #language-Madurese #language-Maithili #language-North Moluccan Malay #language-Moksha #language-Morisyen #language-Malagasy #language-Mambae #language-Marshallese #language-Eastern Mari #language-Maori #language-Mi'kmaq #language-Minangkabau #language-Macedonian #language-Malayalam #language-Mongolian #language-Manipuri #language-Mon #language-Mohawk #language-Marathi #language-Maltese #language-Tagal Murut #language-Mirandese #language-Hmong Daw #language-Burmese #language-Erzya #language-Nauru #language-nah #language-Min Nan Chinese #language-Norwegian Bokmål #language-Central Huasteca Nahuatl #language-Low German #language-Kriang #language-Guerrero Nahuatl #language-Niuean #language-Dutch #language-Orizaba Nahuatl #language-Norwegian Nynorsk #language-Nogai #language-Old Norse #language-Novial #language-Nepali (individual language) #language-Tase Naga #language-Nuer #language-Navajo #language-Nyanja #language-Nyungar #language-Old Aramaic (up to 700 BCE) #language-Occitan (post 1500) #language-Old Frisian #language-Tohono O'odham #language-Oriya (macrolanguage) #language-Old Russian #language-Ossetian #language-Old Spanish #language-Ottoman Turkish (1500-1928) #language-Old Turkish #language-Panjabi #language-Pangasinan #language-Pahlavi #language-Pampanga #language-Papiamento #language-Palauan #language-Picard #language-Pennsylvania German #language-Iranian Persian #language-Phoenician #language-Pali #language-Polish #language-Piemontese #language-Western Panjabi #language-Pipil #language-Prussian #language-Pushto #language-Portuguese #language-Quechua #language-K'iche' #language-Quenya #language-Rapanui #language-Tarifit #language-Romansh #language-Rundi #language-Romanian #language-Romany #language-Russian #language-Rusyn #language-Kinyarwanda #language-Sanskrit #language-Yakut #language-Sardinian #language-Sicilian #language-Scots #language-Sindhi #language-Southern Kurdish #language-Northern Sami #language-Sango #language-Samogitian #language-Shuswap #language-Tachawit #language-Sinhala #language-Sindarin #language-Slovenian #language-Samoan #language-Southern Sami #language-Shona #language-Somali #language-Albanian #language-Serbian #language-Saterfriesisch #language-Sundanese #language-Sumerian #language-Swedish #language-Swabian #language-Swahili (individual language) #language-Classical Syriac #language-Tamil #language-Telugu #language-Tetum #language-Tajik #language-Thai #language-Tahaggart Tamahaq #language-Tigrinya #language-Tigre #language-Turkmen #language-Tagalog #language-Klingon #language-Talysh #language-Jewish Babylonian Aramaic (ca. 200-1200 CE) #language-Temuan #language-Tswana #language-Tonga (Tonga Islands) #language-Tonga (Zambia) #language-Toki Pona #language-Tok Pisin #language-tpw #language-Turkish #language-Tsonga #language-Tatar #language-Northeastern Thai #language-Tuvalu #language-Tahitian #language-Tuvinian #language-Talossan #language-Udmurt #language-Uighur #language-Ukrainian #language-Umbundu #language-Urdu #language-Uzbek #language-Venetian #language-Veps #language-Vietnamese #language-Volapük #language-Võro #language-Walloon #language-Waray (Philippines) #language-Wolof #language-Wu Chinese #language-Kalmyk #language-Xhosa #language-Karakhanid #language-Yiddish #language-Yoruba #language-Yue Chinese #language-Malay (individual language) #language-Standard Malay #language-Zulu #language-Zaza #license-cc-by-2.0 #region-us \n# Dataset Card for Tatoeba## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: URL\n- Leaderboard: \n- Point of Contact:### Dataset Summary\n\nTatoeba is a collection of sentences and translations.\n\nTo load a language pair which isn't part of the config, all you need to do is specify the language code as pairs.\nYou can find the valid pairs in Homepage section of Dataset Description: URL\nE.g.\n\n'dataset = load_dataset(\"tatoeba\", lang1=\"en\", lang2=\"he\")'\n\nThe default date is v2021-07-22, but you can also change the date with\n\n'dataset = load_dataset(\"tatoeba\", lang1=\"en\", lang2=\"he\", date=\"v2020-11-09\")'### Supported Tasks and Leaderboards" ]
1989b4f9db21675b95daea5db40e23a08650ee2c
# Dataset Card for "ted_hrlr" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** - **Repository:** https://github.com/neulab/word-embeddings-for-nmt - **Paper:** [When and Why Are Pre-Trained Word Embeddings Useful for Neural Machine Translation?](https://aclanthology.org/N18-2084/) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 1.83 GB - **Size of the generated dataset:** 281.66 MB - **Total amount of disk used:** 2.12 GB ### Dataset Summary Data sets derived from TED talk transcripts for comparing similar language pairs where one is high resource and the other is low resource. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### az_to_en - **Size of downloaded dataset files:** 131.01 MB - **Size of the generated dataset:** 1.53 MB - **Total amount of disk used:** 132.54 MB An example of 'train' looks as follows. ``` { "translation": { "az": "zəhmət olmasa , sizə xitab edən sözlər eşidəndə əlinizi qaldırın .", "en": "please raise your hand if something applies to you ." } } ``` #### aztr_to_en - **Size of downloaded dataset files:** 131.01 MB - **Size of the generated dataset:** 40.14 MB - **Total amount of disk used:** 171.15 MB An example of 'train' looks as follows. ``` { "translation": { "az_tr": "zəhmət olmasa , sizə xitab edən sözlər eşidəndə əlinizi qaldırın .", "en": "please raise your hand if something applies to you ." } } ``` #### be_to_en - **Size of downloaded dataset files:** 131.01 MB - **Size of the generated dataset:** 1.43 MB - **Total amount of disk used:** 132.42 MB An example of 'train' looks as follows. ``` { "translation": { "be": "zəhmət olmasa , sizə xitab edən sözlər eşidəndə əlinizi qaldırın .", "en": "please raise your hand if something applies to you ." } } ``` #### beru_to_en - **Size of downloaded dataset files:** 131.01 MB - **Size of the generated dataset:** 60.20 MB - **Total amount of disk used:** 191.21 MB An example of 'validation' looks as follows. ``` This example was too long and was cropped: { "translation": "{\"be_ru\": \"11 yaşımdaydım . səhərin birində , evimizdəki sevinc səslərinə oyandığım indiki kimi yadımdadır .\", \"en\": \"when i was..." } ``` #### es_to_pt - **Size of downloaded dataset files:** 131.01 MB - **Size of the generated dataset:** 9.13 MB - **Total amount of disk used:** 140.14 MB An example of 'validation' looks as follows. ``` This example was too long and was cropped: { "translation": "{\"es\": \"11 yaşımdaydım . səhərin birində , evimizdəki sevinc səslərinə oyandığım indiki kimi yadımdadır .\", \"pt\": \"when i was 11..." } ``` ### Data Fields The data fields are the same among all splits. #### az_to_en - `translation`: a multilingual `string` variable, with possible languages including `az`, `en`. #### aztr_to_en - `translation`: a multilingual `string` variable, with possible languages including `az_tr`, `en`. #### be_to_en - `translation`: a multilingual `string` variable, with possible languages including `be`, `en`. #### beru_to_en - `translation`: a multilingual `string` variable, with possible languages including `be_ru`, `en`. #### es_to_pt - `translation`: a multilingual `string` variable, with possible languages including `es`, `pt`. ### Data Splits | name |train |validation|test| |----------|-----:|---------:|---:| |az_to_en | 5947| 672| 904| |aztr_to_en|188397| 672| 904| |be_to_en | 4510| 249| 665| |beru_to_en|212615| 249| 665| |es_to_pt | 44939| 1017|1764| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @inproceedings{qi-etal-2018-pre, title = "When and Why Are Pre-Trained Word Embeddings Useful for Neural Machine Translation?", author = "Qi, Ye and Sachan, Devendra and Felix, Matthieu and Padmanabhan, Sarguna and Neubig, Graham", booktitle = "Proceedings of the 2018 Conference of the North {A}merican Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers)", month = jun, year = "2018", address = "New Orleans, Louisiana", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/N18-2084", doi = "10.18653/v1/N18-2084", pages = "529--535", } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@lewtun](https://github.com/lewtun), [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset.
ted_hrlr
[ "task_categories:translation", "annotations_creators:crowdsourced", "language_creators:expert-generated", "multilinguality:translation", "size_categories:1M<n<10M", "source_datasets:extended|ted_talks_iwslt", "language:az", "language:be", "language:en", "language:es", "language:fr", "language:gl", "language:he", "language:it", "language:pt", "language:ru", "language:tr", "license:cc-by-nc-nd-4.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["expert-generated"], "language": ["az", "be", "en", "es", "fr", "gl", "he", "it", "pt", "ru", "tr"], "license": ["cc-by-nc-nd-4.0"], "multilinguality": ["translation"], "size_categories": ["1M<n<10M"], "source_datasets": ["extended|ted_talks_iwslt"], "task_categories": ["translation"], "task_ids": [], "pretty_name": "TEDHrlr", "dataset_info": [{"config_name": "az_to_en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["az", "en"]}}}], "splits": [{"name": "test", "num_bytes": 186540, "num_examples": 904}, {"name": "train", "num_bytes": 1226853, "num_examples": 5947}, {"name": "validation", "num_bytes": 122709, "num_examples": 672}], "download_size": 131005909, "dataset_size": 1536102}, {"config_name": "aztr_to_en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["az_tr", "en"]}}}], "splits": [{"name": "test", "num_bytes": 186540, "num_examples": 904}, {"name": "train", "num_bytes": 39834469, "num_examples": 188397}, {"name": "validation", "num_bytes": 122709, "num_examples": 672}], "download_size": 131005909, "dataset_size": 40143718}, {"config_name": "be_to_en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["be", "en"]}}}], "splits": [{"name": "test", "num_bytes": 186606, "num_examples": 665}, {"name": "train", "num_bytes": 1176899, "num_examples": 4510}, {"name": "validation", "num_bytes": 59328, "num_examples": 249}], "download_size": 131005909, "dataset_size": 1422833}, {"config_name": "beru_to_en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["be_ru", "en"]}}}], "splits": [{"name": "test", "num_bytes": 186606, "num_examples": 665}, {"name": "train", "num_bytes": 59953616, "num_examples": 212615}, {"name": "validation", "num_bytes": 59328, "num_examples": 249}], "download_size": 131005909, "dataset_size": 60199550}, {"config_name": "es_to_pt", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["es", "pt"]}}}], "splits": [{"name": "test", "num_bytes": 343640, "num_examples": 1764}, {"name": "train", "num_bytes": 8611393, "num_examples": 44939}, {"name": "validation", "num_bytes": 181535, "num_examples": 1017}], "download_size": 131005909, "dataset_size": 9136568}, {"config_name": "fr_to_pt", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["fr", "pt"]}}}], "splits": [{"name": "test", "num_bytes": 311650, "num_examples": 1495}, {"name": "train", "num_bytes": 8755387, "num_examples": 43874}, {"name": "validation", "num_bytes": 212317, "num_examples": 1132}], "download_size": 131005909, "dataset_size": 9279354}, {"config_name": "gl_to_en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["gl", "en"]}}}], "splits": [{"name": "test", "num_bytes": 193213, "num_examples": 1008}, {"name": "train", "num_bytes": 1961363, "num_examples": 10018}, {"name": "validation", "num_bytes": 137929, "num_examples": 683}], "download_size": 131005909, "dataset_size": 2292505}, {"config_name": "glpt_to_en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["gl_pt", "en"]}}}], "splits": [{"name": "test", "num_bytes": 193213, "num_examples": 1008}, {"name": "train", "num_bytes": 11734254, "num_examples": 61803}, {"name": "validation", "num_bytes": 137929, "num_examples": 683}], "download_size": 131005909, "dataset_size": 12065396}, {"config_name": "he_to_pt", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["he", "pt"]}}}], "splits": [{"name": "test", "num_bytes": 361378, "num_examples": 1624}, {"name": "train", "num_bytes": 10627615, "num_examples": 48512}, {"name": "validation", "num_bytes": 230725, "num_examples": 1146}], "download_size": 131005909, "dataset_size": 11219718}, {"config_name": "it_to_pt", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["it", "pt"]}}}], "splits": [{"name": "test", "num_bytes": 324726, "num_examples": 1670}, {"name": "train", "num_bytes": 8905825, "num_examples": 46260}, {"name": "validation", "num_bytes": 210375, "num_examples": 1163}], "download_size": 131005909, "dataset_size": 9440926}, {"config_name": "pt_to_en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["pt", "en"]}}}], "splits": [{"name": "test", "num_bytes": 347803, "num_examples": 1804}, {"name": "train", "num_bytes": 9772911, "num_examples": 51786}, {"name": "validation", "num_bytes": 207960, "num_examples": 1194}], "download_size": 131005909, "dataset_size": 10328674}, {"config_name": "ru_to_en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ru", "en"]}}}], "splits": [{"name": "test", "num_bytes": 1459576, "num_examples": 5477}, {"name": "train", "num_bytes": 58778442, "num_examples": 208107}, {"name": "validation", "num_bytes": 1318357, "num_examples": 4806}], "download_size": 131005909, "dataset_size": 61556375}, {"config_name": "ru_to_pt", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ru", "pt"]}}}], "splits": [{"name": "test", "num_bytes": 409062, "num_examples": 1589}, {"name": "train", "num_bytes": 11882860, "num_examples": 47279}, {"name": "validation", "num_bytes": 276866, "num_examples": 1185}], "download_size": 131005909, "dataset_size": 12568788}, {"config_name": "tr_to_en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["tr", "en"]}}}], "splits": [{"name": "test", "num_bytes": 1026406, "num_examples": 5030}, {"name": "train", "num_bytes": 38607636, "num_examples": 182451}, {"name": "validation", "num_bytes": 832358, "num_examples": 4046}], "download_size": 131005909, "dataset_size": 40466400}]}
2024-01-18T11:16:49+00:00
[]
[ "az", "be", "en", "es", "fr", "gl", "he", "it", "pt", "ru", "tr" ]
TAGS #task_categories-translation #annotations_creators-crowdsourced #language_creators-expert-generated #multilinguality-translation #size_categories-1M<n<10M #source_datasets-extended|ted_talks_iwslt #language-Azerbaijani #language-Belarusian #language-English #language-Spanish #language-French #language-Galician #language-Hebrew #language-Italian #language-Portuguese #language-Russian #language-Turkish #license-cc-by-nc-nd-4.0 #region-us
Dataset Card for "ted\_hrlr" ============================ Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: * Repository: URL * Paper: When and Why Are Pre-Trained Word Embeddings Useful for Neural Machine Translation? * Point of Contact: * Size of downloaded dataset files: 1.83 GB * Size of the generated dataset: 281.66 MB * Total amount of disk used: 2.12 GB ### Dataset Summary Data sets derived from TED talk transcripts for comparing similar language pairs where one is high resource and the other is low resource. ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- ### Data Instances #### az\_to\_en * Size of downloaded dataset files: 131.01 MB * Size of the generated dataset: 1.53 MB * Total amount of disk used: 132.54 MB An example of 'train' looks as follows. #### aztr\_to\_en * Size of downloaded dataset files: 131.01 MB * Size of the generated dataset: 40.14 MB * Total amount of disk used: 171.15 MB An example of 'train' looks as follows. #### be\_to\_en * Size of downloaded dataset files: 131.01 MB * Size of the generated dataset: 1.43 MB * Total amount of disk used: 132.42 MB An example of 'train' looks as follows. #### beru\_to\_en * Size of downloaded dataset files: 131.01 MB * Size of the generated dataset: 60.20 MB * Total amount of disk used: 191.21 MB An example of 'validation' looks as follows. #### es\_to\_pt * Size of downloaded dataset files: 131.01 MB * Size of the generated dataset: 9.13 MB * Total amount of disk used: 140.14 MB An example of 'validation' looks as follows. ### Data Fields The data fields are the same among all splits. #### az\_to\_en * 'translation': a multilingual 'string' variable, with possible languages including 'az', 'en'. #### aztr\_to\_en * 'translation': a multilingual 'string' variable, with possible languages including 'az\_tr', 'en'. #### be\_to\_en * 'translation': a multilingual 'string' variable, with possible languages including 'be', 'en'. #### beru\_to\_en * 'translation': a multilingual 'string' variable, with possible languages including 'be\_ru', 'en'. #### es\_to\_pt * 'translation': a multilingual 'string' variable, with possible languages including 'es', 'pt'. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @thomwolf, @lewtun, @patrickvonplaten for adding this dataset.
[ "### Dataset Summary\n\n\nData sets derived from TED talk transcripts for comparing similar language pairs\nwhere one is high resource and the other is low resource.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### az\\_to\\_en\n\n\n* Size of downloaded dataset files: 131.01 MB\n* Size of the generated dataset: 1.53 MB\n* Total amount of disk used: 132.54 MB\n\n\nAn example of 'train' looks as follows.", "#### aztr\\_to\\_en\n\n\n* Size of downloaded dataset files: 131.01 MB\n* Size of the generated dataset: 40.14 MB\n* Total amount of disk used: 171.15 MB\n\n\nAn example of 'train' looks as follows.", "#### be\\_to\\_en\n\n\n* Size of downloaded dataset files: 131.01 MB\n* Size of the generated dataset: 1.43 MB\n* Total amount of disk used: 132.42 MB\n\n\nAn example of 'train' looks as follows.", "#### beru\\_to\\_en\n\n\n* Size of downloaded dataset files: 131.01 MB\n* Size of the generated dataset: 60.20 MB\n* Total amount of disk used: 191.21 MB\n\n\nAn example of 'validation' looks as follows.", "#### es\\_to\\_pt\n\n\n* Size of downloaded dataset files: 131.01 MB\n* Size of the generated dataset: 9.13 MB\n* Total amount of disk used: 140.14 MB\n\n\nAn example of 'validation' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### az\\_to\\_en\n\n\n* 'translation': a multilingual 'string' variable, with possible languages including 'az', 'en'.", "#### aztr\\_to\\_en\n\n\n* 'translation': a multilingual 'string' variable, with possible languages including 'az\\_tr', 'en'.", "#### be\\_to\\_en\n\n\n* 'translation': a multilingual 'string' variable, with possible languages including 'be', 'en'.", "#### beru\\_to\\_en\n\n\n* 'translation': a multilingual 'string' variable, with possible languages including 'be\\_ru', 'en'.", "#### es\\_to\\_pt\n\n\n* 'translation': a multilingual 'string' variable, with possible languages including 'es', 'pt'.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @thomwolf, @lewtun, @patrickvonplaten for adding this dataset." ]
[ "TAGS\n#task_categories-translation #annotations_creators-crowdsourced #language_creators-expert-generated #multilinguality-translation #size_categories-1M<n<10M #source_datasets-extended|ted_talks_iwslt #language-Azerbaijani #language-Belarusian #language-English #language-Spanish #language-French #language-Galician #language-Hebrew #language-Italian #language-Portuguese #language-Russian #language-Turkish #license-cc-by-nc-nd-4.0 #region-us \n", "### Dataset Summary\n\n\nData sets derived from TED talk transcripts for comparing similar language pairs\nwhere one is high resource and the other is low resource.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### az\\_to\\_en\n\n\n* Size of downloaded dataset files: 131.01 MB\n* Size of the generated dataset: 1.53 MB\n* Total amount of disk used: 132.54 MB\n\n\nAn example of 'train' looks as follows.", "#### aztr\\_to\\_en\n\n\n* Size of downloaded dataset files: 131.01 MB\n* Size of the generated dataset: 40.14 MB\n* Total amount of disk used: 171.15 MB\n\n\nAn example of 'train' looks as follows.", "#### be\\_to\\_en\n\n\n* Size of downloaded dataset files: 131.01 MB\n* Size of the generated dataset: 1.43 MB\n* Total amount of disk used: 132.42 MB\n\n\nAn example of 'train' looks as follows.", "#### beru\\_to\\_en\n\n\n* Size of downloaded dataset files: 131.01 MB\n* Size of the generated dataset: 60.20 MB\n* Total amount of disk used: 191.21 MB\n\n\nAn example of 'validation' looks as follows.", "#### es\\_to\\_pt\n\n\n* Size of downloaded dataset files: 131.01 MB\n* Size of the generated dataset: 9.13 MB\n* Total amount of disk used: 140.14 MB\n\n\nAn example of 'validation' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### az\\_to\\_en\n\n\n* 'translation': a multilingual 'string' variable, with possible languages including 'az', 'en'.", "#### aztr\\_to\\_en\n\n\n* 'translation': a multilingual 'string' variable, with possible languages including 'az\\_tr', 'en'.", "#### be\\_to\\_en\n\n\n* 'translation': a multilingual 'string' variable, with possible languages including 'be', 'en'.", "#### beru\\_to\\_en\n\n\n* 'translation': a multilingual 'string' variable, with possible languages including 'be\\_ru', 'en'.", "#### es\\_to\\_pt\n\n\n* 'translation': a multilingual 'string' variable, with possible languages including 'es', 'pt'.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @thomwolf, @lewtun, @patrickvonplaten for adding this dataset." ]
[ 151, 37, 10, 11, 6, 57, 59, 57, 59, 58, 17, 37, 41, 37, 40, 37, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 6, 28 ]
[ "passage: TAGS\n#task_categories-translation #annotations_creators-crowdsourced #language_creators-expert-generated #multilinguality-translation #size_categories-1M<n<10M #source_datasets-extended|ted_talks_iwslt #language-Azerbaijani #language-Belarusian #language-English #language-Spanish #language-French #language-Galician #language-Hebrew #language-Italian #language-Portuguese #language-Russian #language-Turkish #license-cc-by-nc-nd-4.0 #region-us \n### Dataset Summary\n\n\nData sets derived from TED talk transcripts for comparing similar language pairs\nwhere one is high resource and the other is low resource.### Supported Tasks and Leaderboards### Languages\n\n\nDataset Structure\n-----------------### Data Instances#### az\\_to\\_en\n\n\n* Size of downloaded dataset files: 131.01 MB\n* Size of the generated dataset: 1.53 MB\n* Total amount of disk used: 132.54 MB\n\n\nAn example of 'train' looks as follows.#### aztr\\_to\\_en\n\n\n* Size of downloaded dataset files: 131.01 MB\n* Size of the generated dataset: 40.14 MB\n* Total amount of disk used: 171.15 MB\n\n\nAn example of 'train' looks as follows.#### be\\_to\\_en\n\n\n* Size of downloaded dataset files: 131.01 MB\n* Size of the generated dataset: 1.43 MB\n* Total amount of disk used: 132.42 MB\n\n\nAn example of 'train' looks as follows.#### beru\\_to\\_en\n\n\n* Size of downloaded dataset files: 131.01 MB\n* Size of the generated dataset: 60.20 MB\n* Total amount of disk used: 191.21 MB\n\n\nAn example of 'validation' looks as follows.#### es\\_to\\_pt\n\n\n* Size of downloaded dataset files: 131.01 MB\n* Size of the generated dataset: 9.13 MB\n* Total amount of disk used: 140.14 MB\n\n\nAn example of 'validation' looks as follows." ]
d0051d67070edbc92ebd7c0197b47ff888e4afc3
# Dataset Card for TedIwlst2013 ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** http://opus.nlpl.eu/TED2013.php - **Repository:** None - **Paper:** hhttp://www.lrec-conf.org/proceedings/lrec2012/pdf/463_Paper.pdf - **Leaderboard:** None - **Point of Contact:** [More Information Needed] ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@abhishekkrthakur](https://github.com/abhishekkrthakur) for adding this dataset.
ted_iwlst2013
[ "task_categories:translation", "annotations_creators:found", "language_creators:found", "multilinguality:multilingual", "size_categories:100K<n<1M", "source_datasets:original", "language:ar", "language:de", "language:en", "language:es", "language:fa", "language:fr", "language:it", "language:nl", "language:pl", "language:pt", "language:ro", "language:ru", "language:sl", "language:tr", "language:zh", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["ar", "de", "en", "es", "fa", "fr", "it", "nl", "pl", "pt", "ro", "ru", "sl", "tr", "zh"], "license": ["unknown"], "multilinguality": ["multilingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["translation"], "task_ids": [], "pretty_name": "TedIwlst2013", "config_names": ["ar-en", "de-en", "en-es", "en-fa", "en-fr", "en-it", "en-nl", "en-pl", "en-pt", "en-ro", "en-ru", "en-sl", "en-tr", "en-zh"], "dataset_info": [{"config_name": "ar-en", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["ar", "en"]}}}], "splits": [{"name": "train", "num_bytes": 37413446, "num_examples": 152838}], "download_size": 12065234, "dataset_size": 37413446}, {"config_name": "de-en", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["de", "en"]}}}], "splits": [{"name": "train", "num_bytes": 30295518, "num_examples": 143836}], "download_size": 10931406, "dataset_size": 30295518}, {"config_name": "en-es", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "es"]}}}], "splits": [{"name": "train", "num_bytes": 32522545, "num_examples": 157895}], "download_size": 11642092, "dataset_size": 32522545}, {"config_name": "en-fa", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "fa"]}}}], "splits": [{"name": "train", "num_bytes": 22228781, "num_examples": 80510}], "download_size": 6579696, "dataset_size": 22228781}, {"config_name": "en-fr", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "fr"]}}}], "splits": [{"name": "train", "num_bytes": 34355481, "num_examples": 160420}], "download_size": 12061420, "dataset_size": 34355481}, {"config_name": "en-it", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "it"]}}}], "splits": [{"name": "train", "num_bytes": 32916537, "num_examples": 159391}], "download_size": 11774644, "dataset_size": 32916537}, {"config_name": "en-nl", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "nl"]}}}], "splits": [{"name": "train", "num_bytes": 29679822, "num_examples": 145951}], "download_size": 10712032, "dataset_size": 29679822}, {"config_name": "en-pl", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "pl"]}}}], "splits": [{"name": "train", "num_bytes": 29776339, "num_examples": 149120}], "download_size": 10999482, "dataset_size": 29776339}, {"config_name": "en-pt", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "pt"]}}}], "splits": [{"name": "train", "num_bytes": 32179607, "num_examples": 155995}], "download_size": 11493053, "dataset_size": 32179607}, {"config_name": "en-ro", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "ro"]}}}], "splits": [{"name": "train", "num_bytes": 32958421, "num_examples": 158483}], "download_size": 11936172, "dataset_size": 32958421}, {"config_name": "en-ru", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "ru"]}}}], "splits": [{"name": "train", "num_bytes": 36529465, "num_examples": 133660}], "download_size": 11167700, "dataset_size": 36529465}, {"config_name": "en-sl", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "sl"]}}}], "splits": [{"name": "train", "num_bytes": 2831344, "num_examples": 14960}], "download_size": 1060712, "dataset_size": 2831344}, {"config_name": "en-tr", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "tr"]}}}], "splits": [{"name": "train", "num_bytes": 28016103, "num_examples": 137028}], "download_size": 10038531, "dataset_size": 28016103}, {"config_name": "en-zh", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 30205477, "num_examples": 154579}], "download_size": 11714497, "dataset_size": 30205477}]}
2024-01-18T11:16:53+00:00
[]
[ "ar", "de", "en", "es", "fa", "fr", "it", "nl", "pl", "pt", "ro", "ru", "sl", "tr", "zh" ]
TAGS #task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-100K<n<1M #source_datasets-original #language-Arabic #language-German #language-English #language-Spanish #language-Persian #language-French #language-Italian #language-Dutch #language-Polish #language-Portuguese #language-Romanian #language-Russian #language-Slovenian #language-Turkish #language-Chinese #license-unknown #region-us
# Dataset Card for TedIwlst2013 ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: None - Paper: hhttp://URL - Leaderboard: None - Point of Contact: ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @abhishekkrthakur for adding this dataset.
[ "# Dataset Card for TedIwlst2013", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: hhttp://URL\n- Leaderboard: None\n- Point of Contact:", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ "TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-100K<n<1M #source_datasets-original #language-Arabic #language-German #language-English #language-Spanish #language-Persian #language-French #language-Italian #language-Dutch #language-Polish #language-Portuguese #language-Romanian #language-Russian #language-Slovenian #language-Turkish #language-Chinese #license-unknown #region-us \n", "# Dataset Card for TedIwlst2013", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: hhttp://URL\n- Leaderboard: None\n- Point of Contact:", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ 145, 10, 120, 33, 6, 10, 4, 6, 6, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 20 ]
[ "passage: TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-100K<n<1M #source_datasets-original #language-Arabic #language-German #language-English #language-Spanish #language-Persian #language-French #language-Italian #language-Dutch #language-Polish #language-Portuguese #language-Romanian #language-Russian #language-Slovenian #language-Turkish #language-Chinese #license-unknown #region-us \n# Dataset Card for TedIwlst2013## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: hhttp://URL\n- Leaderboard: None\n- Point of Contact:### Dataset Summary### Supported Tasks and Leaderboards### Languages## Dataset Structure### Data Instances### Data Fields### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators### Licensing Information### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
4216d6f1452102ff962a97f5e4eba8d8d84f66f2
# Dataset Card for "ted_multi" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://github.com/neulab/word-embeddings-for-nmt](https://github.com/neulab/word-embeddings-for-nmt) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 352.23 MB - **Size of the generated dataset:** 791.01 MB - **Total amount of disk used:** 1.14 GB ### Dataset Summary Massively multilingual (60 language) data set derived from TED Talk transcripts. Each record consists of parallel arrays of language and text. Missing and incomplete translations will be filtered out. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### plain_text - **Size of downloaded dataset files:** 352.23 MB - **Size of the generated dataset:** 791.01 MB - **Total amount of disk used:** 1.14 GB An example of 'validation' looks as follows. ``` This example was too long and was cropped: { "talk_name": "shabana_basij_rasikh_dare_to_educate_afghan_girls", "translations": "{\"language\": [\"ar\", \"az\", \"bg\", \"bn\", \"cs\", \"da\", \"de\", \"el\", \"en\", \"es\", \"fa\", \"fr\", \"he\", \"hi\", \"hr\", \"hu\", \"hy\", \"id\", \"it\", ..." } ``` ### Data Fields The data fields are the same among all splits. #### plain_text - `translations`: a multilingual `string` variable, with possible languages including `ar`, `az`, `be`, `bg`, `bn`. - `talk_name`: a `string` feature. ### Data Splits | name |train |validation|test| |----------|-----:|---------:|---:| |plain_text|258098| 6049|7213| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @InProceedings{qi-EtAl:2018:N18-2, author = {Qi, Ye and Sachan, Devendra and Felix, Matthieu and Padmanabhan, Sarguna and Neubig, Graham}, title = {When and Why Are Pre-Trained Word Embeddings Useful for Neural Machine Translation?}, booktitle = {Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers)}, month = {June}, year = {2018}, address = {New Orleans, Louisiana}, publisher = {Association for Computational Linguistics}, pages = {529--535}, abstract = {The performance of Neural Machine Translation (NMT) systems often suffers in low-resource scenarios where sufficiently large-scale parallel corpora cannot be obtained. Pre-trained word embeddings have proven to be invaluable for improving performance in natural language analysis tasks, which often suffer from paucity of data. However, their utility for NMT has not been extensively explored. In this work, we perform five sets of experiments that analyze when we can expect pre-trained word embeddings to help in NMT tasks. We show that such embeddings can be surprisingly effective in some cases -- providing gains of up to 20 BLEU points in the most favorable setting.}, url = {http://www.aclweb.org/anthology/N18-2084} } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@lewtun](https://github.com/lewtun), [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset.
ted_multi
[ "region:us" ]
2022-03-02T23:29:22+00:00
{"pretty_name": "TEDMulti", "dataset_info": {"features": [{"name": "translations", "dtype": {"translation_variable_languages": {"languages": ["ar", "az", "be", "bg", "bn", "bs", "calv", "cs", "da", "de", "el", "en", "eo", "es", "et", "eu", "fa", "fi", "fr", "fr-ca", "gl", "he", "hi", "hr", "hu", "hy", "id", "it", "ja", "ka", "kk", "ko", "ku", "lt", "mk", "mn", "mr", "ms", "my", "nb", "nl", "pl", "pt", "pt-br", "ro", "ru", "sk", "sl", "sq", "sr", "sv", "ta", "th", "tr", "uk", "ur", "vi", "zh", "zh-cn", "zh-tw"], "num_languages": 60}}}, {"name": "talk_name", "dtype": "string"}], "config_name": "plain_text", "splits": [{"name": "test", "num_bytes": 23364983, "num_examples": 7213}, {"name": "train", "num_bytes": 748209995, "num_examples": 258098}, {"name": "validation", "num_bytes": 19435383, "num_examples": 6049}], "download_size": 352222045, "dataset_size": 791010361}}
2024-01-18T11:16:56+00:00
[]
[]
TAGS #region-us
Dataset Card for "ted\_multi" ============================= Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: * Point of Contact: * Size of downloaded dataset files: 352.23 MB * Size of the generated dataset: 791.01 MB * Total amount of disk used: 1.14 GB ### Dataset Summary Massively multilingual (60 language) data set derived from TED Talk transcripts. Each record consists of parallel arrays of language and text. Missing and incomplete translations will be filtered out. ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- ### Data Instances #### plain\_text * Size of downloaded dataset files: 352.23 MB * Size of the generated dataset: 791.01 MB * Total amount of disk used: 1.14 GB An example of 'validation' looks as follows. ### Data Fields The data fields are the same among all splits. #### plain\_text * 'translations': a multilingual 'string' variable, with possible languages including 'ar', 'az', 'be', 'bg', 'bn'. * 'talk\_name': a 'string' feature. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @thomwolf, @lewtun, @patrickvonplaten for adding this dataset.
[ "### Dataset Summary\n\n\nMassively multilingual (60 language) data set derived from TED Talk transcripts.\nEach record consists of parallel arrays of language and text. Missing and\nincomplete translations will be filtered out.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### plain\\_text\n\n\n* Size of downloaded dataset files: 352.23 MB\n* Size of the generated dataset: 791.01 MB\n* Total amount of disk used: 1.14 GB\n\n\nAn example of 'validation' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### plain\\_text\n\n\n* 'translations': a multilingual 'string' variable, with possible languages including 'ar', 'az', 'be', 'bg', 'bn'.\n* 'talk\\_name': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @thomwolf, @lewtun, @patrickvonplaten for adding this dataset." ]
[ "TAGS\n#region-us \n", "### Dataset Summary\n\n\nMassively multilingual (60 language) data set derived from TED Talk transcripts.\nEach record consists of parallel arrays of language and text. Missing and\nincomplete translations will be filtered out.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### plain\\_text\n\n\n* Size of downloaded dataset files: 352.23 MB\n* Size of the generated dataset: 791.01 MB\n* Total amount of disk used: 1.14 GB\n\n\nAn example of 'validation' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### plain\\_text\n\n\n* 'translations': a multilingual 'string' variable, with possible languages including 'ar', 'az', 'be', 'bg', 'bn'.\n* 'talk\\_name': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @thomwolf, @lewtun, @patrickvonplaten for adding this dataset." ]
[ 6, 53, 10, 11, 6, 55, 17, 61, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 6, 28 ]
[ "passage: TAGS\n#region-us \n### Dataset Summary\n\n\nMassively multilingual (60 language) data set derived from TED Talk transcripts.\nEach record consists of parallel arrays of language and text. Missing and\nincomplete translations will be filtered out.### Supported Tasks and Leaderboards### Languages\n\n\nDataset Structure\n-----------------### Data Instances#### plain\\_text\n\n\n* Size of downloaded dataset files: 352.23 MB\n* Size of the generated dataset: 791.01 MB\n* Total amount of disk used: 1.14 GB\n\n\nAn example of 'validation' looks as follows.### Data Fields\n\n\nThe data fields are the same among all splits.#### plain\\_text\n\n\n* 'translations': a multilingual 'string' variable, with possible languages including 'ar', 'az', 'be', 'bg', 'bn'.\n* 'talk\\_name': a 'string' feature.### Data Splits\n\n\n\nDataset Creation\n----------------### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------### Social Impact of Dataset### Discussion of Biases### Other Known Limitations\n\n\nAdditional Information\n----------------------### Dataset Curators### Licensing Information### Contributions\n\n\nThanks to @thomwolf, @lewtun, @patrickvonplaten for adding this dataset." ]
39479634113d3d716305d294dfa5581fa29df496
# Dataset Card for Web Inventory of Transcribed & Translated(WIT) Ted Talks ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://wit3.fbk.eu/home - **Repository:** https://drive.google.com/file/d/1Cz1Un9p8Xn9IpEMMrg2kXSDt0dnjxc4z/view?usp=sharing - **Paper:** https://www.aclweb.org/anthology/2012.eamt-1.60.pdf - **Leaderboard:** [Needs More Information] - **Point of Contact:** [Mauro Cettolo](mailto:cettolo@fbk.eu) [Roldano Cattoni](mailto:cattoni@fbk.eu) ### Dataset Summary The Web Inventory Talk is a collection of the original Ted talks and their translated version. The translations are available in more than 109+ languages, though the distribution is not uniform. To load a language pair which isn't part of the config, all you need to do is specify the language code as pairs. E.g. `dataset = load_dataset("ted_talks_iwslt", language_pair=("it", "pl"), year="2014")` The full list of languages is: 'af', 'am', 'ar', 'arq', 'art-x-bork', 'as', 'ast', 'az', 'be', 'bg', 'bi', 'bn', 'bo', 'bs', 'ca', 'ceb', 'cnh', 'cs', 'da', 'de', 'el', 'en', 'eo', 'es', 'et', 'eu', 'fa', 'fi', 'fil', 'fr', 'fr-ca', 'ga', 'gl', 'gu', 'ha', 'he', 'hi', 'hr', 'ht', 'hu', 'hup', 'hy', 'id', 'ig', 'inh', 'is', 'it', 'ja', 'ka', 'kk', 'km', 'kn', 'ko', 'ku', 'ky', 'la', 'lb', 'lo', 'lt', 'ltg', 'lv', 'mg', 'mk', 'ml', 'mn', 'mr', 'ms', 'mt', 'my', 'nb', 'ne', 'nl', 'nn', 'oc', 'pa', 'pl', 'ps', 'pt', 'pt-br', 'ro', 'ru', 'rup', 'sh', 'si', 'sk', 'sl', 'so', 'sq', 'sr', 'srp', 'sv', 'sw', 'szl', 'ta', 'te', 'tg', 'th', 'tl', 'tlh', 'tr', 'tt', 'ug', 'uk', 'ur', 'uz', 'vi', 'zh', 'zh-cn', 'zh-tw'. The full list of years is: '2014', '2015', '2016'. ### Supported Tasks and Leaderboards machine learning task, language modeling and generation ### Languages Ted talks are mostly held in English (`en`). Almost all of the talks have been translated, by volunteers, into Arabic, Bulgarian, Chinese (simplified), French, Italian, Korean, Portuguese (Brazil) and Spanish. For about 70 other languages, the number of translated talks ranges from several hundreds (e.g. such as other Dutch, German, Hebrew, Romanian) to one (e.g. Hausa, Hupa, Bislama, Ingush, Maltese). The languages in the dataset are: - af - am - ar - arq - art - as - ast - az - be - bg - bi - bn - bo - bs - ca - ceb - cnh - cs - da - de - el - en - eo - es - et - eu - fa - fi - fil - fr - ga - gl - gu - ha - he - hi - hr - ht - hu - hup - hy - id - ig - inh - is - it - ja - ka - kk - km - kn - ko - ku - ky - la - lb - lo - lt - ltg - lv - mg - mk - ml - mn - mr - ms - mt - my - nb - ne - nl - nn - oc - pa - pl - ps - pt - ro - ru - rup - sh - si - sk - sl - so - sq - sr - srp: Serbian (`sr`) - sv - sw - szl - ta - te - tg - th - tl - tlh - tr - tt - ug - uk - ur - uz - vi - zh ## Dataset Structure ### Data Instances One example from the dataset is: ``` {'translation': {'hi': 'जब मार्च २०१४ में इबोला का प्रकोप छाया, पर्डिस सबेटी और उनकी टीम को वाइरस के जीनोम का अनुक्रमण करना था, सीखना था कि यह कैसे परवतिर्त होते हैं और फैलते हैं। सबेटी ने तुरंत ही अपने अनुसंधान को वेब में जारी किया, ताकि दुनिया भर के वाइरस ट्रैकर्स और वैज्ञानिक इस तत्काल लड़ाई में शामिल हो सकें। इस बातचीत में, वह दिखाती हैं कि सबका सहयोग ही कुंजी है वाइरस को रोकने के लिए--और लड़ने के लिए आगे आने वाले हमलों से। सबेटी ने कहा,"हमने खुले तौर पर काम किया, साझा किया और साथ काम किया"। "हमे दुनिया को एक वाइरस के विनाश से नहीं, पर अरबों दिलों और दिमागों की एकता से परिभाषित करना है"।', 'nl': 'Toen Ebola in maart 2014 uitbrak, zijn Pardis Sabeti en haar team aan het werk gegaan om het genoom in kaart te brengen. Zo ontdekten ze hoe het virus zich verspreidde en muteerde. Sabeti zette direct haar onderzoek op het internet, zodat wereldwijd virus-jagers en wetenschappers mee konden werken aan de strijd. In deze talk laat ze zien hoe die openheid geholpen heeft bij het stoppen van het virus en hoe het kan helpen bij de strijd tegen het volgende virus. "We moesten transparant werken, delen en samenwerken". Sabeti zegt:"Laat de wereld niet ten onder gaan aan een virus, maar verlicht worden door miljoenen harten en geesten die samenwerken."'}} ``` The original XML files are formatted like this example: ``` <file id="1"> <head> <url>http://www.ted.com/talks/ryan_holladay_to_hear_this_music_you_have_to_be_there_literally.html</url> <pagesize>66634</pagesize> <dtime>Sun Jan 12 15:17:32 CET 2014</dtime> <content-type>text/html; charset=utf-8</content-type> <encoding>utf-8</encoding> <videourl>http://download.ted.com/talks/RyanHolladay_2013S.mp4</videourl> <videopath>talks/RyanHolladay_2013S.mp4</videopath> <transcription> <seekvideo id="2939">(Music)</seekvideo> <seekvideo id="7555">For any of you who have visited or lived in New York City,</seekvideo> <seekvideo id="11221">these shots might start to look familiar.</seekvideo> <seekvideo id="16116">This is Central Park,</seekvideo> . . . <seekvideo id="361992">for people to interact with</seekvideo> <seekvideo id="363709">and experience music.</seekvideo> <seekvideo id="365451">Thank you.</seekvideo> <seekvideo id="367495">(Applause)</seekvideo> </transcription> <talkid>1903</talkid> <title>Ryan Holladay: To hear this music you have to be there. Literally</title> <description>The music industry ......segments of sounds that only play when a listener is physically nearby. (Filmed at TED@BCG.)</description> <keywords>entertainment,music,technology</keywords> <image>http://images.ted.com/images/ted/d98c17773da6f84e9f915895c270c7ffd2de3778_389x292.jpg</image> <date>2014/01/12</date> <wordnum>885</wordnum> <charnum>5051</charnum> </head> <content>(Music) For any of you who have visited or lived in New York City, these shots might start to look familiar. This is Central Park, ............new ways for people to interact with and experience music. Thank you. (Applause)</content> </file> ``` ### Data Fields The fields of the dataset are: - translation: - <lang1>: text in <lang1> - <lang2>L translated text in <lang2> Information about the original data files: For each language, a single XML file is generated which includes all talks subtitled in that language. Each talk is enclosed in tags `<file id="int">` and `</file>` and includes, among other tags: | Tags | Description | |---|:---| | `<url>`| the address of the original HTML document of the talk | | `<speaker>` | the name of the talk speaker | | `<talkid>` | the numeric talk identifier | | `<transcript>` | talk subtitles split in captions | | `<date>` | the issue date of the talk | | `<content>` | talk subtitles | ### Data Splits The paper doesn't provide any specific train-test-dev splits. However data can be split by available years (2014, 2015, 2016) ## Dataset Creation ### Curation Rationale TED Conference, based in California, has been posting all video recordings of its talks together with subtitles in English and their translations in more than 80 languages. Aside from its cultural and social relevance, this content, which is published under the Creative Commons BYNC-ND license, also represents a precious language resource for the machine translation research community, thanks to its size, variety of topics, and covered languages. ### Source Data #### Initial Data Collection and Normalization The talks were collected from the [Ted Conference website](http://www.ted.com/) #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? Translation has been contributed by volunteers ### Personal and Sensitive Information No personal and sensitive information is provided in the dataset. All talks are publicly available ## Considerations for Using the Data ### Social Impact of Dataset In statistical machine translation, large amount of in-domain parallel data are usually required to properly train translation and reordering models. With more than 900+ Ted talks (as of 2011) and translation in more than 90+ languages. This dataset provides a useful resource for the MT research community. In turn, this enables easy access to a vast treasure trove of human knowledge. ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators The original dataset was curated by: [Mauro Cettolo](mailto:cettolo@fbk.eu) [Roldano Cattoni](mailto:cattoni@fbk.eu) Author: Christian Girardi For issues with the HuggingFace Dataset implementation, reach out: [Aakash Gupta](mailto:aakashg80@gmail.com) ### Licensing Information cc-by-nc-nd-4.0 ### Citation Information ``` @inproceedings{cettolo-etal-2012-wit3, title = "{WIT}3: Web Inventory of Transcribed and Translated Talks", author = "Cettolo, Mauro and Girardi, Christian and Federico, Marcello", booktitle = "Proceedings of the 16th Annual conference of the European Association for Machine Translation", month = may # " 28{--}30", year = "2012", address = "Trento, Italy", publisher = "European Association for Machine Translation", url = "https://www.aclweb.org/anthology/2012.eamt-1.60", pages = "261--268", } ``` ### Contributions Thanks to [@skyprince999](https://github.com/skyprince999) for adding this dataset.
ted_talks_iwslt
[ "task_categories:translation", "annotations_creators:expert-generated", "language_creators:crowdsourced", "language_creators:expert-generated", "multilinguality:translation", "size_categories:1K<n<10K", "size_categories:n<1K", "source_datasets:original", "language:af", "language:am", "language:ar", "language:arq", "language:art", "language:as", "language:ast", "language:az", "language:be", "language:bg", "language:bi", "language:bn", "language:bo", "language:bs", "language:ca", "language:ceb", "language:cnh", "language:cs", "language:da", "language:de", "language:el", "language:en", "language:eo", "language:es", "language:et", "language:eu", "language:fa", "language:fi", "language:fil", "language:fr", "language:ga", "language:gl", "language:gu", "language:ha", "language:he", "language:hi", "language:hr", "language:ht", "language:hu", "language:hup", "language:hy", "language:id", "language:ig", "language:inh", "language:is", "language:it", "language:ja", "language:ka", "language:kk", "language:km", "language:kn", "language:ko", "language:ku", "language:ky", "language:la", "language:lb", "language:lo", "language:lt", "language:ltg", "language:lv", "language:mg", "language:mk", "language:ml", "language:mn", "language:mr", "language:ms", "language:mt", "language:my", "language:nb", "language:ne", "language:nl", "language:nn", "language:oc", "language:pa", "language:pl", "language:ps", "language:pt", "language:ro", "language:ru", "language:rup", "language:sh", "language:si", "language:sk", "language:sl", "language:so", "language:sq", "language:sr", "language:sv", "language:sw", "language:szl", "language:ta", "language:te", "language:tg", "language:th", "language:tl", "language:tlh", "language:tr", "language:tt", "language:ug", "language:uk", "language:ur", "language:uz", "language:vi", "language:zh", "license:cc-by-nc-nd-4.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["crowdsourced", "expert-generated"], "language": ["af", "am", "ar", "arq", "art", "as", "ast", "az", "be", "bg", "bi", "bn", "bo", "bs", "ca", "ceb", "cnh", "cs", "da", "de", "el", "en", "eo", "es", "et", "eu", "fa", "fi", "fil", "fr", "ga", "gl", "gu", "ha", "he", "hi", "hr", "ht", "hu", "hup", "hy", "id", "ig", "inh", "is", "it", "ja", "ka", "kk", "km", "kn", "ko", "ku", "ky", "la", "lb", "lo", "lt", "ltg", "lv", "mg", "mk", "ml", "mn", "mr", "ms", "mt", "my", "nb", "ne", "nl", "nn", "oc", "pa", "pl", "ps", "pt", "ro", "ru", "rup", "sh", "si", "sk", "sl", "so", "sq", "sr", "sv", "sw", "szl", "ta", "te", "tg", "th", "tl", "tlh", "tr", "tt", "ug", "uk", "ur", "uz", "vi", "zh"], "license": ["cc-by-nc-nd-4.0"], "multilinguality": ["translation"], "size_categories": ["1K<n<10K", "n<1K"], "source_datasets": ["original"], "task_categories": ["translation"], "task_ids": [], "pretty_name": "Web Inventory of Transcribed & Translated (WIT) Ted Talks", "config_names": ["de_ja_2014", "de_ja_2015", "de_ja_2016", "eu_ca_2014", "eu_ca_2015", "eu_ca_2016", "fr-ca_hi_2014", "fr-ca_hi_2015", "fr-ca_hi_2016", "nl_en_2014", "nl_en_2015", "nl_en_2016", "nl_hi_2014", "nl_hi_2015", "nl_hi_2016"], "language_bcp47": ["art-x-bork", "fr-CA", "pt-BR", "zh-CN", "zh-TW"], "dataset_info": [{"config_name": "eu_ca_2014", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["eu", "ca"]}}}], "splits": [{"name": "train", "num_bytes": 15192, "num_examples": 44}], "download_size": 1666674366, "dataset_size": 15192}, {"config_name": "eu_ca_2015", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["eu", "ca"]}}}], "splits": [{"name": "train", "num_bytes": 18768, "num_examples": 52}], "download_size": 1666674366, "dataset_size": 18768}, {"config_name": "eu_ca_2016", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["eu", "ca"]}}}], "splits": [{"name": "train", "num_bytes": 19506, "num_examples": 54}], "download_size": 1666674366, "dataset_size": 19506}, {"config_name": "nl_en_2014", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["nl", "en"]}}}], "splits": [{"name": "train", "num_bytes": 1035545, "num_examples": 2966}], "download_size": 1666674366, "dataset_size": 1035545}, {"config_name": "nl_en_2015", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["nl", "en"]}}}], "splits": [{"name": "train", "num_bytes": 1292610, "num_examples": 3550}], "download_size": 1666674366, "dataset_size": 1292610}, {"config_name": "nl_en_2016", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["nl", "en"]}}}], "splits": [{"name": "train", "num_bytes": 1434207, "num_examples": 3852}], "download_size": 1666674366, "dataset_size": 1434207}, {"config_name": "nl_hi_2014", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["nl", "hi"]}}}], "splits": [{"name": "train", "num_bytes": 214870, "num_examples": 367}], "download_size": 1666674366, "dataset_size": 214870}, {"config_name": "nl_hi_2015", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["nl", "hi"]}}}], "splits": [{"name": "train", "num_bytes": 252192, "num_examples": 421}], "download_size": 1666674366, "dataset_size": 252192}, {"config_name": "nl_hi_2016", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["nl", "hi"]}}}], "splits": [{"name": "train", "num_bytes": 310922, "num_examples": 496}], "download_size": 1666674366, "dataset_size": 310922}, {"config_name": "de_ja_2014", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["de", "ja"]}}}], "splits": [{"name": "train", "num_bytes": 1074403, "num_examples": 2536}], "download_size": 1666674366, "dataset_size": 1074403}, {"config_name": "de_ja_2015", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["de", "ja"]}}}], "splits": [{"name": "train", "num_bytes": 1442047, "num_examples": 3247}], "download_size": 1666674366, "dataset_size": 1442047}, {"config_name": "de_ja_2016", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["de", "ja"]}}}], "splits": [{"name": "train", "num_bytes": 1630729, "num_examples": 3590}], "download_size": 1666674366, "dataset_size": 1630729}, {"config_name": "fr-ca_hi_2014", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["fr-ca", "hi"]}}}], "splits": [{"name": "train", "num_bytes": 74472, "num_examples": 127}], "download_size": 1666674366, "dataset_size": 74472}, {"config_name": "fr-ca_hi_2015", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["fr-ca", "hi"]}}}], "splits": [{"name": "train", "num_bytes": 82448, "num_examples": 141}], "download_size": 1666674366, "dataset_size": 82448}, {"config_name": "fr-ca_hi_2016", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["fr-ca", "hi"]}}}], "splits": [{"name": "train", "num_bytes": 93425, "num_examples": 156}], "download_size": 1666674366, "dataset_size": 93425}]}
2024-01-18T11:16:58+00:00
[]
[ "af", "am", "ar", "arq", "art", "as", "ast", "az", "be", "bg", "bi", "bn", "bo", "bs", "ca", "ceb", "cnh", "cs", "da", "de", "el", "en", "eo", "es", "et", "eu", "fa", "fi", "fil", "fr", "ga", "gl", "gu", "ha", "he", "hi", "hr", "ht", "hu", "hup", "hy", "id", "ig", "inh", "is", "it", "ja", "ka", "kk", "km", "kn", "ko", "ku", "ky", "la", "lb", "lo", "lt", "ltg", "lv", "mg", "mk", "ml", "mn", "mr", "ms", "mt", "my", "nb", "ne", "nl", "nn", "oc", "pa", "pl", "ps", "pt", "ro", "ru", "rup", "sh", "si", "sk", "sl", "so", "sq", "sr", "sv", "sw", "szl", "ta", "te", "tg", "th", "tl", "tlh", "tr", "tt", "ug", "uk", "ur", "uz", "vi", "zh" ]
TAGS #task_categories-translation #annotations_creators-expert-generated #language_creators-crowdsourced #language_creators-expert-generated #multilinguality-translation #size_categories-1K<n<10K #size_categories-n<1K #source_datasets-original #language-Afrikaans #language-Amharic #language-Arabic #language-Algerian Arabic #language-art #language-Assamese #language-Asturian #language-Azerbaijani #language-Belarusian #language-Bulgarian #language-Bislama #language-Bengali #language-Tibetan #language-Bosnian #language-Catalan #language-Cebuano #language-Hakha Chin #language-Czech #language-Danish #language-German #language-Modern Greek (1453-) #language-English #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Persian #language-Finnish #language-Filipino #language-French #language-Irish #language-Galician #language-Gujarati #language-Hausa #language-Hebrew #language-Hindi #language-Croatian #language-Haitian #language-Hungarian #language-Hupa #language-Armenian #language-Indonesian #language-Igbo #language-Ingush #language-Icelandic #language-Italian #language-Japanese #language-Georgian #language-Kazakh #language-Khmer #language-Kannada #language-Korean #language-Kurdish #language-Kirghiz #language-Latin #language-Luxembourgish #language-Lao #language-Lithuanian #language-Latgalian #language-Latvian #language-Malagasy #language-Macedonian #language-Malayalam #language-Mongolian #language-Marathi #language-Malay (macrolanguage) #language-Maltese #language-Burmese #language-Norwegian Bokmål #language-Nepali (macrolanguage) #language-Dutch #language-Norwegian Nynorsk #language-Occitan (post 1500) #language-Panjabi #language-Polish #language-Pushto #language-Portuguese #language-Romanian #language-Russian #language-Macedo-Romanian #language-Serbo-Croatian #language-Sinhala #language-Slovak #language-Slovenian #language-Somali #language-Albanian #language-Serbian #language-Swedish #language-Swahili (macrolanguage) #language-Silesian #language-Tamil #language-Telugu #language-Tajik #language-Thai #language-Tagalog #language-Klingon #language-Turkish #language-Tatar #language-Uighur #language-Ukrainian #language-Urdu #language-Uzbek #language-Vietnamese #language-Chinese #license-cc-by-nc-nd-4.0 #region-us
Dataset Card for Web Inventory of Transcribed & Translated(WIT) Ted Talks ========================================================================= Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: URL * Paper: URL * Leaderboard: * Point of Contact: Mauro Cettolo Roldano Cattoni ### Dataset Summary The Web Inventory Talk is a collection of the original Ted talks and their translated version. The translations are available in more than 109+ languages, though the distribution is not uniform. To load a language pair which isn't part of the config, all you need to do is specify the language code as pairs. E.g. 'dataset = load\_dataset("ted\_talks\_iwslt", language\_pair=("it", "pl"), year="2014")' The full list of languages is: 'af', 'am', 'ar', 'arq', 'art-x-bork', 'as', 'ast', 'az', 'be', 'bg', 'bi', 'bn', 'bo', 'bs', 'ca', 'ceb', 'cnh', 'cs', 'da', 'de', 'el', 'en', 'eo', 'es', 'et', 'eu', 'fa', 'fi', 'fil', 'fr', 'fr-ca', 'ga', 'gl', 'gu', 'ha', 'he', 'hi', 'hr', 'ht', 'hu', 'hup', 'hy', 'id', 'ig', 'inh', 'is', 'it', 'ja', 'ka', 'kk', 'km', 'kn', 'ko', 'ku', 'ky', 'la', 'lb', 'lo', 'lt', 'ltg', 'lv', 'mg', 'mk', 'ml', 'mn', 'mr', 'ms', 'mt', 'my', 'nb', 'ne', 'nl', 'nn', 'oc', 'pa', 'pl', 'ps', 'pt', 'pt-br', 'ro', 'ru', 'rup', 'sh', 'si', 'sk', 'sl', 'so', 'sq', 'sr', 'srp', 'sv', 'sw', 'szl', 'ta', 'te', 'tg', 'th', 'tl', 'tlh', 'tr', 'tt', 'ug', 'uk', 'ur', 'uz', 'vi', 'zh', 'zh-cn', 'zh-tw'. The full list of years is: '2014', '2015', '2016'. ### Supported Tasks and Leaderboards machine learning task, language modeling and generation ### Languages Ted talks are mostly held in English ('en'). Almost all of the talks have been translated, by volunteers, into Arabic, Bulgarian, Chinese (simplified), French, Italian, Korean, Portuguese (Brazil) and Spanish. For about 70 other languages, the number of translated talks ranges from several hundreds (e.g. such as other Dutch, German, Hebrew, Romanian) to one (e.g. Hausa, Hupa, Bislama, Ingush, Maltese). The languages in the dataset are: * af * am * ar * arq * art * as * ast * az * be * bg * bi * bn * bo * bs * ca * ceb * cnh * cs * da * de * el * en * eo * es * et * eu * fa * fi * fil * fr * ga * gl * gu * ha * he * hi * hr * ht * hu * hup * hy * id * ig * inh * is * it * ja * ka * kk * km * kn * ko * ku * ky * la * lb * lo * lt * ltg * lv * mg * mk * ml * mn * mr * ms * mt * my * nb * ne * nl * nn * oc * pa * pl * ps * pt * ro * ru * rup * sh * si * sk * sl * so * sq * sr * srp: Serbian ('sr') * sv * sw * szl * ta * te * tg * th * tl * tlh * tr * tt * ug * uk * ur * uz * vi * zh Dataset Structure ----------------- ### Data Instances One example from the dataset is: The original XML files are formatted like this example: ### Data Fields The fields of the dataset are: * translation: + : text in + L translated text in Information about the original data files: For each language, a single XML file is generated which includes all talks subtitled in that language. Each talk is enclosed in tags '' and '' and includes, among other tags: ### Data Splits The paper doesn't provide any specific train-test-dev splits. However data can be split by available years (2014, 2015, 2016) Dataset Creation ---------------- ### Curation Rationale TED Conference, based in California, has been posting all video recordings of its talks together with subtitles in English and their translations in more than 80 languages. Aside from its cultural and social relevance, this content, which is published under the Creative Commons BYNC-ND license, also represents a precious language resource for the machine translation research community, thanks to its size, variety of topics, and covered languages. ### Source Data #### Initial Data Collection and Normalization The talks were collected from the Ted Conference website #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? Translation has been contributed by volunteers ### Personal and Sensitive Information No personal and sensitive information is provided in the dataset. All talks are publicly available Considerations for Using the Data --------------------------------- ### Social Impact of Dataset In statistical machine translation, large amount of in-domain parallel data are usually required to properly train translation and reordering models. With more than 900+ Ted talks (as of 2011) and translation in more than 90+ languages. This dataset provides a useful resource for the MT research community. In turn, this enables easy access to a vast treasure trove of human knowledge. ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators The original dataset was curated by: Mauro Cettolo Roldano Cattoni Author: Christian Girardi For issues with the HuggingFace Dataset implementation, reach out: Aakash Gupta ### Licensing Information cc-by-nc-nd-4.0 ### Contributions Thanks to @skyprince999 for adding this dataset.
[ "### Dataset Summary\n\n\nThe Web Inventory Talk is a collection of the original Ted talks and their translated version. The translations are available in more than 109+ languages, though the distribution is not uniform.\n\n\nTo load a language pair which isn't part of the config, all you need to do is specify the language code as pairs.\nE.g.\n\n\n'dataset = load\\_dataset(\"ted\\_talks\\_iwslt\", language\\_pair=(\"it\", \"pl\"), year=\"2014\")'\n\n\nThe full list of languages is: 'af', 'am', 'ar', 'arq', 'art-x-bork', 'as', 'ast', 'az', 'be', 'bg', 'bi', 'bn', 'bo', 'bs', 'ca', 'ceb', 'cnh', 'cs', 'da', 'de', 'el', 'en', 'eo', 'es', 'et', 'eu', 'fa', 'fi', 'fil', 'fr', 'fr-ca', 'ga', 'gl', 'gu', 'ha', 'he', 'hi', 'hr', 'ht', 'hu', 'hup', 'hy', 'id', 'ig', 'inh', 'is', 'it', 'ja', 'ka', 'kk', 'km', 'kn', 'ko', 'ku', 'ky', 'la', 'lb', 'lo', 'lt', 'ltg', 'lv', 'mg', 'mk', 'ml', 'mn', 'mr', 'ms', 'mt', 'my', 'nb', 'ne', 'nl', 'nn', 'oc', 'pa', 'pl', 'ps', 'pt', 'pt-br', 'ro', 'ru', 'rup', 'sh', 'si', 'sk', 'sl', 'so', 'sq', 'sr', 'srp', 'sv', 'sw', 'szl', 'ta', 'te', 'tg', 'th', 'tl', 'tlh', 'tr', 'tt', 'ug', 'uk', 'ur', 'uz', 'vi', 'zh', 'zh-cn', 'zh-tw'.\n\n\nThe full list of years is: '2014', '2015', '2016'.", "### Supported Tasks and Leaderboards\n\n\nmachine learning task, language modeling and generation", "### Languages\n\n\nTed talks are mostly held in English ('en'). Almost all of the talks have been translated, by volunteers, into Arabic, Bulgarian, Chinese (simplified), French, Italian, Korean, Portuguese (Brazil) and Spanish. For about 70 other languages, the number of translated talks ranges from several hundreds (e.g. such as other Dutch, German, Hebrew, Romanian) to one (e.g. Hausa, Hupa, Bislama, Ingush, Maltese).\n\n\nThe languages in the dataset are:\n\n\n* af\n* am\n* ar\n* arq\n* art\n* as\n* ast\n* az\n* be\n* bg\n* bi\n* bn\n* bo\n* bs\n* ca\n* ceb\n* cnh\n* cs\n* da\n* de\n* el\n* en\n* eo\n* es\n* et\n* eu\n* fa\n* fi\n* fil\n* fr\n* ga\n* gl\n* gu\n* ha\n* he\n* hi\n* hr\n* ht\n* hu\n* hup\n* hy\n* id\n* ig\n* inh\n* is\n* it\n* ja\n* ka\n* kk\n* km\n* kn\n* ko\n* ku\n* ky\n* la\n* lb\n* lo\n* lt\n* ltg\n* lv\n* mg\n* mk\n* ml\n* mn\n* mr\n* ms\n* mt\n* my\n* nb\n* ne\n* nl\n* nn\n* oc\n* pa\n* pl\n* ps\n* pt\n* ro\n* ru\n* rup\n* sh\n* si\n* sk\n* sl\n* so\n* sq\n* sr\n* srp: Serbian ('sr')\n* sv\n* sw\n* szl\n* ta\n* te\n* tg\n* th\n* tl\n* tlh\n* tr\n* tt\n* ug\n* uk\n* ur\n* uz\n* vi\n* zh\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nOne example from the dataset is:\n\n\nThe original XML files are formatted like this example:", "### Data Fields\n\n\nThe fields of the dataset are:\n\n\n* translation:\n\t+ : text in\n\t+ L translated text in\n\n\nInformation about the original data files:\n\n\nFor each language, a single XML file is generated which includes all talks subtitled in\nthat language. Each talk is enclosed in tags '' and '' and includes, among other tags:", "### Data Splits\n\n\nThe paper doesn't provide any specific train-test-dev splits. However data can be split by available years (2014, 2015, 2016)\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nTED Conference, based in California, has been posting all video recordings of its talks together with subtitles in English and their translations in more than 80 languages. Aside from its cultural and social relevance, this content, which is published under the Creative Commons BYNC-ND license, also represents a precious language resource for the machine translation research community, thanks to its size, variety of topics, and covered languages.", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nThe talks were collected from the Ted Conference website", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?\n\n\nTranslation has been contributed by volunteers", "### Personal and Sensitive Information\n\n\nNo personal and sensitive information is provided in the dataset. All talks are publicly available\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset\n\n\nIn statistical machine translation, large amount of in-domain parallel data are usually required to properly train translation and reordering models. With more than 900+ Ted talks (as of 2011) and translation in more than 90+ languages. This dataset provides a useful resource for the MT research community.\n\n\nIn turn, this enables easy access to a vast treasure trove of human knowledge.", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThe original dataset was curated by:\nMauro Cettolo\nRoldano Cattoni\n\n\nAuthor:\nChristian Girardi\n\n\nFor issues with the HuggingFace Dataset implementation, reach out: Aakash Gupta", "### Licensing Information\n\n\ncc-by-nc-nd-4.0", "### Contributions\n\n\nThanks to @skyprince999 for adding this dataset." ]
[ "TAGS\n#task_categories-translation #annotations_creators-expert-generated #language_creators-crowdsourced #language_creators-expert-generated #multilinguality-translation #size_categories-1K<n<10K #size_categories-n<1K #source_datasets-original #language-Afrikaans #language-Amharic #language-Arabic #language-Algerian Arabic #language-art #language-Assamese #language-Asturian #language-Azerbaijani #language-Belarusian #language-Bulgarian #language-Bislama #language-Bengali #language-Tibetan #language-Bosnian #language-Catalan #language-Cebuano #language-Hakha Chin #language-Czech #language-Danish #language-German #language-Modern Greek (1453-) #language-English #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Persian #language-Finnish #language-Filipino #language-French #language-Irish #language-Galician #language-Gujarati #language-Hausa #language-Hebrew #language-Hindi #language-Croatian #language-Haitian #language-Hungarian #language-Hupa #language-Armenian #language-Indonesian #language-Igbo #language-Ingush #language-Icelandic #language-Italian #language-Japanese #language-Georgian #language-Kazakh #language-Khmer #language-Kannada #language-Korean #language-Kurdish #language-Kirghiz #language-Latin #language-Luxembourgish #language-Lao #language-Lithuanian #language-Latgalian #language-Latvian #language-Malagasy #language-Macedonian #language-Malayalam #language-Mongolian #language-Marathi #language-Malay (macrolanguage) #language-Maltese #language-Burmese #language-Norwegian Bokmål #language-Nepali (macrolanguage) #language-Dutch #language-Norwegian Nynorsk #language-Occitan (post 1500) #language-Panjabi #language-Polish #language-Pushto #language-Portuguese #language-Romanian #language-Russian #language-Macedo-Romanian #language-Serbo-Croatian #language-Sinhala #language-Slovak #language-Slovenian #language-Somali #language-Albanian #language-Serbian #language-Swedish #language-Swahili (macrolanguage) #language-Silesian #language-Tamil #language-Telugu #language-Tajik #language-Thai #language-Tagalog #language-Klingon #language-Turkish #language-Tatar #language-Uighur #language-Ukrainian #language-Urdu #language-Uzbek #language-Vietnamese #language-Chinese #license-cc-by-nc-nd-4.0 #region-us \n", "### Dataset Summary\n\n\nThe Web Inventory Talk is a collection of the original Ted talks and their translated version. The translations are available in more than 109+ languages, though the distribution is not uniform.\n\n\nTo load a language pair which isn't part of the config, all you need to do is specify the language code as pairs.\nE.g.\n\n\n'dataset = load\\_dataset(\"ted\\_talks\\_iwslt\", language\\_pair=(\"it\", \"pl\"), year=\"2014\")'\n\n\nThe full list of languages is: 'af', 'am', 'ar', 'arq', 'art-x-bork', 'as', 'ast', 'az', 'be', 'bg', 'bi', 'bn', 'bo', 'bs', 'ca', 'ceb', 'cnh', 'cs', 'da', 'de', 'el', 'en', 'eo', 'es', 'et', 'eu', 'fa', 'fi', 'fil', 'fr', 'fr-ca', 'ga', 'gl', 'gu', 'ha', 'he', 'hi', 'hr', 'ht', 'hu', 'hup', 'hy', 'id', 'ig', 'inh', 'is', 'it', 'ja', 'ka', 'kk', 'km', 'kn', 'ko', 'ku', 'ky', 'la', 'lb', 'lo', 'lt', 'ltg', 'lv', 'mg', 'mk', 'ml', 'mn', 'mr', 'ms', 'mt', 'my', 'nb', 'ne', 'nl', 'nn', 'oc', 'pa', 'pl', 'ps', 'pt', 'pt-br', 'ro', 'ru', 'rup', 'sh', 'si', 'sk', 'sl', 'so', 'sq', 'sr', 'srp', 'sv', 'sw', 'szl', 'ta', 'te', 'tg', 'th', 'tl', 'tlh', 'tr', 'tt', 'ug', 'uk', 'ur', 'uz', 'vi', 'zh', 'zh-cn', 'zh-tw'.\n\n\nThe full list of years is: '2014', '2015', '2016'.", "### Supported Tasks and Leaderboards\n\n\nmachine learning task, language modeling and generation", "### Languages\n\n\nTed talks are mostly held in English ('en'). Almost all of the talks have been translated, by volunteers, into Arabic, Bulgarian, Chinese (simplified), French, Italian, Korean, Portuguese (Brazil) and Spanish. For about 70 other languages, the number of translated talks ranges from several hundreds (e.g. such as other Dutch, German, Hebrew, Romanian) to one (e.g. Hausa, Hupa, Bislama, Ingush, Maltese).\n\n\nThe languages in the dataset are:\n\n\n* af\n* am\n* ar\n* arq\n* art\n* as\n* ast\n* az\n* be\n* bg\n* bi\n* bn\n* bo\n* bs\n* ca\n* ceb\n* cnh\n* cs\n* da\n* de\n* el\n* en\n* eo\n* es\n* et\n* eu\n* fa\n* fi\n* fil\n* fr\n* ga\n* gl\n* gu\n* ha\n* he\n* hi\n* hr\n* ht\n* hu\n* hup\n* hy\n* id\n* ig\n* inh\n* is\n* it\n* ja\n* ka\n* kk\n* km\n* kn\n* ko\n* ku\n* ky\n* la\n* lb\n* lo\n* lt\n* ltg\n* lv\n* mg\n* mk\n* ml\n* mn\n* mr\n* ms\n* mt\n* my\n* nb\n* ne\n* nl\n* nn\n* oc\n* pa\n* pl\n* ps\n* pt\n* ro\n* ru\n* rup\n* sh\n* si\n* sk\n* sl\n* so\n* sq\n* sr\n* srp: Serbian ('sr')\n* sv\n* sw\n* szl\n* ta\n* te\n* tg\n* th\n* tl\n* tlh\n* tr\n* tt\n* ug\n* uk\n* ur\n* uz\n* vi\n* zh\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nOne example from the dataset is:\n\n\nThe original XML files are formatted like this example:", "### Data Fields\n\n\nThe fields of the dataset are:\n\n\n* translation:\n\t+ : text in\n\t+ L translated text in\n\n\nInformation about the original data files:\n\n\nFor each language, a single XML file is generated which includes all talks subtitled in\nthat language. Each talk is enclosed in tags '' and '' and includes, among other tags:", "### Data Splits\n\n\nThe paper doesn't provide any specific train-test-dev splits. However data can be split by available years (2014, 2015, 2016)\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nTED Conference, based in California, has been posting all video recordings of its talks together with subtitles in English and their translations in more than 80 languages. Aside from its cultural and social relevance, this content, which is published under the Creative Commons BYNC-ND license, also represents a precious language resource for the machine translation research community, thanks to its size, variety of topics, and covered languages.", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nThe talks were collected from the Ted Conference website", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?\n\n\nTranslation has been contributed by volunteers", "### Personal and Sensitive Information\n\n\nNo personal and sensitive information is provided in the dataset. All talks are publicly available\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset\n\n\nIn statistical machine translation, large amount of in-domain parallel data are usually required to properly train translation and reordering models. With more than 900+ Ted talks (as of 2011) and translation in more than 90+ languages. This dataset provides a useful resource for the MT research community.\n\n\nIn turn, this enables easy access to a vast treasure trove of human knowledge.", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThe original dataset was curated by:\nMauro Cettolo\nRoldano Cattoni\n\n\nAuthor:\nChristian Girardi\n\n\nFor issues with the HuggingFace Dataset implementation, reach out: Aakash Gupta", "### Licensing Information\n\n\ncc-by-nc-nd-4.0", "### Contributions\n\n\nThanks to @skyprince999 for adding this dataset." ]
[ 712, 613, 19, 388, 25, 76, 40, 100, 4, 21, 10, 5, 5, 17, 37, 88, 8, 14, 46, 16, 18 ]
[ "passage: ", "passage: TAGS\n#task_categories-translation #annotations_creators-expert-generated #language_creators-crowdsourced #language_creators-expert-generated #multilinguality-translation #size_categories-1K<n<10K #size_categories-n<1K #source_datasets-original #language-Afrikaans #language-Amharic #language-Arabic #language-Algerian Arabic #language-art #language-Assamese #language-Asturian #language-Azerbaijani #language-Belarusian #language-Bulgarian #language-Bislama #language-Bengali #language-Tibetan #language-Bosnian #language-Catalan #language-Cebuano #language-Hakha Chin #language-Czech #language-Danish #language-German #language-Modern Greek (1453-) #language-English #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Persian #language-Finnish #language-Filipino #language-French #language-Irish #language-Galician #language-Gujarati #language-Hausa #language-Hebrew #language-Hindi #language-Croatian #language-Haitian #language-Hungarian #language-Hupa #language-Armenian #language-Indonesian #language-Igbo #language-Ingush #language-Icelandic #language-Italian #language-Japanese #language-Georgian #language-Kazakh #language-Khmer #language-Kannada #language-Korean #language-Kurdish #language-Kirghiz #language-Latin #language-Luxembourgish #language-Lao #language-Lithuanian #language-Latgalian #language-Latvian #language-Malagasy #language-Macedonian #language-Malayalam #language-Mongolian #language-Marathi #language-Malay (macrolanguage) #language-Maltese #language-Burmese #language-Norwegian Bokmål #language-Nepali (macrolanguage) #language-Dutch #language-Norwegian Nynorsk #language-Occitan (post 1500) #language-Panjabi #language-Polish #language-Pushto #language-Portuguese #language-Romanian #language-Russian #language-Macedo-Romanian #language-Serbo-Croatian #language-Sinhala #language-Slovak #language-Slovenian #language-Somali #language-Albanian #language-Serbian #language-Swedish #language-Swahili (macrolanguage) #language-Silesian #language-Tamil #language-Telugu #language-Tajik #language-Thai #language-Tagalog #language-Klingon #language-Turkish #language-Tatar #language-Uighur #language-Ukrainian #language-Urdu #language-Uzbek #language-Vietnamese #language-Chinese #license-cc-by-nc-nd-4.0 #region-us \n", "passage: ### Dataset Summary\n\n\nThe Web Inventory Talk is a collection of the original Ted talks and their translated version. The translations are available in more than 109+ languages, though the distribution is not uniform.\n\n\nTo load a language pair which isn't part of the config, all you need to do is specify the language code as pairs.\nE.g.\n\n\n'dataset = load\\_dataset(\"ted\\_talks\\_iwslt\", language\\_pair=(\"it\", \"pl\"), year=\"2014\")'\n\n\nThe full list of languages is: 'af', 'am', 'ar', 'arq', 'art-x-bork', 'as', 'ast', 'az', 'be', 'bg', 'bi', 'bn', 'bo', 'bs', 'ca', 'ceb', 'cnh', 'cs', 'da', 'de', 'el', 'en', 'eo', 'es', 'et', 'eu', 'fa', 'fi', 'fil', 'fr', 'fr-ca', 'ga', 'gl', 'gu', 'ha', 'he', 'hi', 'hr', 'ht', 'hu', 'hup', 'hy', 'id', 'ig', 'inh', 'is', 'it', 'ja', 'ka', 'kk', 'km', 'kn', 'ko', 'ku', 'ky', 'la', 'lb', 'lo', 'lt', 'ltg', 'lv', 'mg', 'mk', 'ml', 'mn', 'mr', 'ms', 'mt', 'my', 'nb', 'ne', 'nl', 'nn', 'oc', 'pa', 'pl', 'ps', 'pt', 'pt-br', 'ro', 'ru', 'rup', 'sh', 'si', 'sk', 'sl', 'so', 'sq', 'sr', 'srp', 'sv', 'sw', 'szl', 'ta', 'te', 'tg', 'th', 'tl', 'tlh', 'tr', 'tt', 'ug', 'uk', 'ur', 'uz', 'vi', 'zh', 'zh-cn', 'zh-tw'.\n\n\nThe full list of years is: '2014', '2015', '2016'.### Supported Tasks and Leaderboards\n\n\nmachine learning task, language modeling and generation### Languages\n\n\nTed talks are mostly held in English ('en'). Almost all of the talks have been translated, by volunteers, into Arabic, Bulgarian, Chinese (simplified), French, Italian, Korean, Portuguese (Brazil) and Spanish. For about 70 other languages, the number of translated talks ranges from several hundreds (e.g. such as other Dutch, German, Hebrew, Romanian) to one (e.g. Hausa, Hupa, Bislama, Ingush, Maltese).\n\n\nThe languages in the dataset are:\n\n\n* af\n* am\n* ar\n* arq\n* art\n* as\n* ast\n* az\n* be\n* bg\n* bi\n* bn\n* bo\n* bs\n* ca\n* ceb\n* cnh\n* cs\n* da\n* de\n* el\n* en\n* eo\n* es\n* et\n* eu\n* fa\n* fi\n* fil\n* fr\n* ga\n* gl\n* gu\n* ha\n* he\n* hi\n* hr\n* ht\n* hu\n* hup\n* hy\n* id\n* ig\n* inh\n* is\n* it\n* ja\n* ka\n* kk\n* km\n* kn\n* ko\n* ku\n* ky\n* la\n* lb\n* lo\n* lt\n* ltg\n* lv\n* mg\n* mk\n* ml\n* mn\n* mr\n* ms\n* mt\n* my\n* nb\n* ne\n* nl\n* nn\n* oc\n* pa\n* pl\n* ps\n* pt\n* ro\n* ru\n* rup\n* sh\n* si\n* sk\n* sl\n* so\n* sq\n* sr\n* srp: Serbian ('sr')\n* sv\n* sw\n* szl\n* ta\n* te\n* tg\n* th\n* tl\n* tlh\n* tr\n* tt\n* ug\n* uk\n* ur\n* uz\n* vi\n* zh\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nOne example from the dataset is:\n\n\nThe original XML files are formatted like this example:" ]
29806b24219457139f241d5f5135437db164d666
# Dataset Card for [telugu_books] ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** [Telugu Books](https://www.kaggle.com/sudalairajkumar/telugu-nlp) - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary This dataset is created by scraping telugu novels from teluguone.com this dataset can be used for nlp tasks like topic modeling, word embeddings, transfer learning etc ### Supported Tasks and Leaderboards [More Information Needed] ### Languages TE - Telugu ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields - Text: Sentence from a novel ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? Anusha Motamarri ### Annotations #### Annotation process Anusha Motamarri #### Who are the annotators? Anusha Motamarri ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@vinaykudari](https://github.com/vinaykudari) for adding this dataset.
telugu_books
[ "task_categories:text-generation", "task_categories:fill-mask", "task_ids:language-modeling", "task_ids:masked-language-modeling", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:n<1K", "source_datasets:original", "language:te", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["expert-generated"], "language": ["te"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["n<1K"], "source_datasets": ["original"], "task_categories": ["text-generation", "fill-mask"], "task_ids": ["language-modeling", "masked-language-modeling"], "pretty_name": "TeluguBooks", "dataset_info": {"features": [{"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 315076011, "num_examples": 25794}], "download_size": 0, "dataset_size": 315076011}}
2024-01-18T11:16:59+00:00
[]
[ "te" ]
TAGS #task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-n<1K #source_datasets-original #language-Telugu #license-unknown #region-us
# Dataset Card for [telugu_books] ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information ## Dataset Description - Homepage: Telugu Books - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary This dataset is created by scraping telugu novels from URL this dataset can be used for nlp tasks like topic modeling, word embeddings, transfer learning etc ### Supported Tasks and Leaderboards ### Languages TE - Telugu ## Dataset Structure ### Data Instances ### Data Fields - Text: Sentence from a novel ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? Anusha Motamarri ### Annotations #### Annotation process Anusha Motamarri #### Who are the annotators? Anusha Motamarri ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @vinaykudari for adding this dataset.
[ "# Dataset Card for [telugu_books]\n\n ## Table of Contents\n - Dataset Description\n - Dataset Summary\n - Supported Tasks\n - Languages\n - Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n - Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n - Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n - Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n\n ## Dataset Description\n\n - Homepage:\n\n Telugu Books\n\n - Repository:\n - Paper:\n - Leaderboard:\n - Point of Contact:\n\n ### Dataset Summary\n\nThis dataset is created by scraping telugu novels from URL this dataset can be used for nlp tasks like topic modeling, word embeddings, transfer learning etc\n\n ### Supported Tasks and Leaderboards\n\n \n\n ### Languages\n\nTE - Telugu\n\n ## Dataset Structure\n\n ### Data Instances\n\n \n\n ### Data Fields\n\n - Text: Sentence from a novel\n\n ### Data Splits\n\n \n\n ## Dataset Creation\n\n ### Curation Rationale\n\n \n\n ### Source Data\n\n #### Initial Data Collection and Normalization\n\n \n\n #### Who are the source language producers?\n\n Anusha Motamarri\n\n ### Annotations\n\n #### Annotation process\n\n Anusha Motamarri\n\n #### Who are the annotators?\n\n Anusha Motamarri\n\n ### Personal and Sensitive Information\n\n \n\n ## Considerations for Using the Data\n\n ### Social Impact of Dataset\n\n \n\n ### Discussion of Biases\n\n \n\n ### Other Known Limitations\n\n \n\n ## Additional Information\n\n ### Dataset Curators\n\n \n\n ### Licensing Information", "### Contributions\n\nThanks to @vinaykudari for adding this dataset." ]
[ "TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-n<1K #source_datasets-original #language-Telugu #license-unknown #region-us \n", "# Dataset Card for [telugu_books]\n\n ## Table of Contents\n - Dataset Description\n - Dataset Summary\n - Supported Tasks\n - Languages\n - Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n - Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n - Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n - Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n\n ## Dataset Description\n\n - Homepage:\n\n Telugu Books\n\n - Repository:\n - Paper:\n - Leaderboard:\n - Point of Contact:\n\n ### Dataset Summary\n\nThis dataset is created by scraping telugu novels from URL this dataset can be used for nlp tasks like topic modeling, word embeddings, transfer learning etc\n\n ### Supported Tasks and Leaderboards\n\n \n\n ### Languages\n\nTE - Telugu\n\n ## Dataset Structure\n\n ### Data Instances\n\n \n\n ### Data Fields\n\n - Text: Sentence from a novel\n\n ### Data Splits\n\n \n\n ## Dataset Creation\n\n ### Curation Rationale\n\n \n\n ### Source Data\n\n #### Initial Data Collection and Normalization\n\n \n\n #### Who are the source language producers?\n\n Anusha Motamarri\n\n ### Annotations\n\n #### Annotation process\n\n Anusha Motamarri\n\n #### Who are the annotators?\n\n Anusha Motamarri\n\n ### Personal and Sensitive Information\n\n \n\n ## Considerations for Using the Data\n\n ### Social Impact of Dataset\n\n \n\n ### Discussion of Biases\n\n \n\n ### Other Known Limitations\n\n \n\n ## Additional Information\n\n ### Dataset Curators\n\n \n\n ### Licensing Information", "### Contributions\n\nThanks to @vinaykudari for adding this dataset." ]
[ 112, 365, 18 ]
[ "passage: TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-n<1K #source_datasets-original #language-Telugu #license-unknown #region-us \n# Dataset Card for [telugu_books]\n\n ## Table of Contents\n - Dataset Description\n - Dataset Summary\n - Supported Tasks\n - Languages\n - Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n - Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n - Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n - Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n\n ## Dataset Description\n\n - Homepage:\n\n Telugu Books\n\n - Repository:\n - Paper:\n - Leaderboard:\n - Point of Contact:\n\n ### Dataset Summary\n\nThis dataset is created by scraping telugu novels from URL this dataset can be used for nlp tasks like topic modeling, word embeddings, transfer learning etc\n\n ### Supported Tasks and Leaderboards\n\n \n\n ### Languages\n\nTE - Telugu\n\n ## Dataset Structure\n\n ### Data Instances\n\n \n\n ### Data Fields\n\n - Text: Sentence from a novel\n\n ### Data Splits\n\n \n\n ## Dataset Creation\n\n ### Curation Rationale\n\n \n\n ### Source Data\n\n #### Initial Data Collection and Normalization\n\n \n\n #### Who are the source language producers?\n\n Anusha Motamarri\n\n ### Annotations\n\n #### Annotation process\n\n Anusha Motamarri\n\n #### Who are the annotators?\n\n Anusha Motamarri\n\n ### Personal and Sensitive Information\n\n \n\n ## Considerations for Using the Data\n\n ### Social Impact of Dataset\n\n \n\n ### Discussion of Biases\n\n \n\n ### Other Known Limitations\n\n \n\n ## Additional Information\n\n ### Dataset Curators\n\n \n\n ### Licensing Information### Contributions\n\nThanks to @vinaykudari for adding this dataset." ]
b6389659bc7503177e9addb898ef297807b64d48
# Dataset Card for [Dataset Name] ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.kaggle.com/sudalairajkumar/telugu-nlp?select=telugu_news - **Repository:** https://github.com/AnushaMotamarri/Telugu-Newspaper-Article-Dataset ### Dataset Summary This dataset contains Telugu language news articles along with respective topic labels (business, editorial, entertainment, nation, sport) extracted from the daily Andhra Jyoti. This dataset could be used to build Classification and Language Models. ### Supported Tasks and Leaderboards Multiclass classification, Topic Classification, Language Model ### Languages TE - Telugu, India ## Dataset Structure ### Data Instances Two CSV files (train, test) with five columns (sno, date, heading, body, topic). ### Data Fields - sno: id - date: publish date of the news article - heading: article heading/title - body: article body/content - topic: one of the following topics (business, editorial, entertainment, nation, sport) ### Data Splits Train and Test ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data - https://www.kaggle.com/sudalairajkumar/telugu-nlp?select=telugu_news - https://github.com/AnushaMotamarri/Telugu-Newspaper-Article-Dataset #### Initial Data Collection and Normalization The source data is scraped articles from archives of Telugu newspaper website Andhra Jyoti. A set of queries were created and the corresponding ground truth answers were retrieved by a combination of BM25 and tf-idf. #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators Sudalai Rajkumar, Anusha Motamarri ### Licensing Information [More Information Needed] ### Citation Information ``` @InProceedings{kaggle:dataset, title = {Telugu News - Natural Language Processing for Indian Languages}, authors={Sudalai Rajkumar, Anusha Motamarri}, year={2019} } ``` ### Contributions Thanks to [@oostopitre](https://github.com/oostopitre) for adding this dataset.
telugu_news
[ "task_categories:text-generation", "task_categories:fill-mask", "task_categories:text-classification", "task_ids:language-modeling", "task_ids:masked-language-modeling", "task_ids:multi-class-classification", "task_ids:topic-classification", "annotations_creators:machine-generated", "language_creators:other", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:te", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["machine-generated"], "language_creators": ["other"], "language": ["te"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-generation", "fill-mask", "text-classification"], "task_ids": ["language-modeling", "masked-language-modeling", "multi-class-classification", "topic-classification"], "pretty_name": "TeluguNews", "dataset_info": {"features": [{"name": "sno", "dtype": "int32"}, {"name": "date", "dtype": "string"}, {"name": "heading", "dtype": "string"}, {"name": "body", "dtype": "string"}, {"name": "topic", "dtype": {"class_label": {"names": {"0": "business", "1": "editorial", "2": "entertainment", "3": "nation", "4": "sports"}}}}], "splits": [{"name": "train", "num_bytes": 69400234, "num_examples": 17312}, {"name": "test", "num_bytes": 17265514, "num_examples": 4329}], "download_size": 0, "dataset_size": 86665748}}
2024-01-18T11:17:01+00:00
[]
[ "te" ]
TAGS #task_categories-text-generation #task_categories-fill-mask #task_categories-text-classification #task_ids-language-modeling #task_ids-masked-language-modeling #task_ids-multi-class-classification #task_ids-topic-classification #annotations_creators-machine-generated #language_creators-other #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Telugu #license-unknown #region-us
# Dataset Card for [Dataset Name] ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: URL ### Dataset Summary This dataset contains Telugu language news articles along with respective topic labels (business, editorial, entertainment, nation, sport) extracted from the daily Andhra Jyoti. This dataset could be used to build Classification and Language Models. ### Supported Tasks and Leaderboards Multiclass classification, Topic Classification, Language Model ### Languages TE - Telugu, India ## Dataset Structure ### Data Instances Two CSV files (train, test) with five columns (sno, date, heading, body, topic). ### Data Fields - sno: id - date: publish date of the news article - heading: article heading/title - body: article body/content - topic: one of the following topics (business, editorial, entertainment, nation, sport) ### Data Splits Train and Test ## Dataset Creation ### Curation Rationale ### Source Data - URL - URL #### Initial Data Collection and Normalization The source data is scraped articles from archives of Telugu newspaper website Andhra Jyoti. A set of queries were created and the corresponding ground truth answers were retrieved by a combination of BM25 and tf-idf. #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators Sudalai Rajkumar, Anusha Motamarri ### Licensing Information ### Contributions Thanks to @oostopitre for adding this dataset.
[ "# Dataset Card for [Dataset Name]", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL", "### Dataset Summary\n\nThis dataset contains Telugu language news articles along with respective topic \nlabels (business, editorial, entertainment, nation, sport) extracted from the daily Andhra Jyoti. \nThis dataset could be used to build Classification and Language Models.", "### Supported Tasks and Leaderboards\n\nMulticlass classification, Topic Classification, Language Model", "### Languages\n\nTE - Telugu, India", "## Dataset Structure", "### Data Instances\n\nTwo CSV files (train, test) with five columns (sno, date, heading, body, topic).", "### Data Fields\n\n- sno: id\n- date: publish date of the news article\n- heading: article heading/title\n- body: article body/content\n- topic: one of the following topics (business, editorial, entertainment, nation, sport)", "### Data Splits\n\nTrain and Test", "## Dataset Creation", "### Curation Rationale", "### Source Data\n\n- URL\n- URL", "#### Initial Data Collection and Normalization\n\nThe source data is scraped articles from archives of Telugu newspaper website Andhra Jyoti. \nA set of queries were created and the corresponding ground truth answers were retrieved by a combination of BM25 and tf-idf.", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nSudalai Rajkumar, Anusha Motamarri", "### Licensing Information", "### Contributions\n\nThanks to @oostopitre for adding this dataset." ]
[ "TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_categories-text-classification #task_ids-language-modeling #task_ids-masked-language-modeling #task_ids-multi-class-classification #task_ids-topic-classification #annotations_creators-machine-generated #language_creators-other #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Telugu #license-unknown #region-us \n", "# Dataset Card for [Dataset Name]", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL", "### Dataset Summary\n\nThis dataset contains Telugu language news articles along with respective topic \nlabels (business, editorial, entertainment, nation, sport) extracted from the daily Andhra Jyoti. \nThis dataset could be used to build Classification and Language Models.", "### Supported Tasks and Leaderboards\n\nMulticlass classification, Topic Classification, Language Model", "### Languages\n\nTE - Telugu, India", "## Dataset Structure", "### Data Instances\n\nTwo CSV files (train, test) with five columns (sno, date, heading, body, topic).", "### Data Fields\n\n- sno: id\n- date: publish date of the news article\n- heading: article heading/title\n- body: article body/content\n- topic: one of the following topics (business, editorial, entertainment, nation, sport)", "### Data Splits\n\nTrain and Test", "## Dataset Creation", "### Curation Rationale", "### Source Data\n\n- URL\n- URL", "#### Initial Data Collection and Normalization\n\nThe source data is scraped articles from archives of Telugu newspaper website Andhra Jyoti. \nA set of queries were created and the corresponding ground truth answers were retrieved by a combination of BM25 and tf-idf.", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nSudalai Rajkumar, Anusha Motamarri", "### Licensing Information", "### Contributions\n\nThanks to @oostopitre for adding this dataset." ]
[ 144, 10, 120, 14, 56, 21, 9, 6, 34, 54, 8, 5, 7, 8, 60, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 16, 6, 18 ]
[ "passage: TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_categories-text-classification #task_ids-language-modeling #task_ids-masked-language-modeling #task_ids-multi-class-classification #task_ids-topic-classification #annotations_creators-machine-generated #language_creators-other #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Telugu #license-unknown #region-us \n# Dataset Card for [Dataset Name]## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: URL### Dataset Summary\n\nThis dataset contains Telugu language news articles along with respective topic \nlabels (business, editorial, entertainment, nation, sport) extracted from the daily Andhra Jyoti. \nThis dataset could be used to build Classification and Language Models.### Supported Tasks and Leaderboards\n\nMulticlass classification, Topic Classification, Language Model### Languages\n\nTE - Telugu, India## Dataset Structure### Data Instances\n\nTwo CSV files (train, test) with five columns (sno, date, heading, body, topic).### Data Fields\n\n- sno: id\n- date: publish date of the news article\n- heading: article heading/title\n- body: article body/content\n- topic: one of the following topics (business, editorial, entertainment, nation, sport)### Data Splits\n\nTrain and Test## Dataset Creation### Curation Rationale### Source Data\n\n- URL\n- URL" ]
34f0027e529f6c28d0a0748097c34f5537a1c9ff
# Dataset Card for [tep_en_fa_para] ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:**[TEP: Tehran English-Persian parallel corpus](http://opus.nlpl.eu/TEP.php) - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary TEP: Tehran English-Persian parallel corpus. The first free Eng-Per corpus, provided by the Natural Language and Text Processing Laboratory, University of Tehran. ### Supported Tasks and Leaderboards The underlying task is machine translation for language pair English-Persian ### Languages English, Persian ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information M. T. Pilevar, H. Faili, and A. H. Pilevar, “TEP: Tehran English-Persian Parallel Corpus”, in proceedings of 12th International Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2011). ### Contributions Thanks to [@spatil6](https://github.com/spatil6) for adding this dataset.
tep_en_fa_para
[ "task_categories:translation", "annotations_creators:found", "language_creators:found", "multilinguality:translation", "size_categories:100K<n<1M", "source_datasets:original", "language:en", "language:fa", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["en", "fa"], "license": ["unknown"], "multilinguality": ["translation"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["translation"], "task_ids": [], "pretty_name": "TepEnFaPara", "dataset_info": {"features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "fa"]}}}], "config_name": "en-fa", "splits": [{"name": "train", "num_bytes": 58735557, "num_examples": 612087}], "download_size": 16353318, "dataset_size": 58735557}}
2024-01-18T11:17:02+00:00
[]
[ "en", "fa" ]
TAGS #task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-translation #size_categories-100K<n<1M #source_datasets-original #language-English #language-Persian #license-unknown #region-us
# Dataset Card for [tep_en_fa_para] ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage:TEP: Tehran English-Persian parallel corpus - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary TEP: Tehran English-Persian parallel corpus. The first free Eng-Per corpus, provided by the Natural Language and Text Processing Laboratory, University of Tehran. ### Supported Tasks and Leaderboards The underlying task is machine translation for language pair English-Persian ### Languages English, Persian ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information M. T. Pilevar, H. Faili, and A. H. Pilevar, “TEP: Tehran English-Persian Parallel Corpus”, in proceedings of 12th International Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2011). ### Contributions Thanks to @spatil6 for adding this dataset.
[ "# Dataset Card for [tep_en_fa_para]", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage:TEP: Tehran English-Persian parallel corpus\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nTEP: Tehran English-Persian parallel corpus. The first free Eng-Per corpus, provided by the Natural Language and Text Processing Laboratory, University of Tehran.", "### Supported Tasks and Leaderboards\n\nThe underlying task is machine translation for language pair English-Persian", "### Languages\n\nEnglish, Persian", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\n\n\n\nM. T. Pilevar, H. Faili, and A. H. Pilevar, “TEP: Tehran English-Persian Parallel Corpus”, in proceedings of 12th International Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2011).", "### Contributions\n\nThanks to @spatil6 for adding this dataset." ]
[ "TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-translation #size_categories-100K<n<1M #source_datasets-original #language-English #language-Persian #license-unknown #region-us \n", "# Dataset Card for [tep_en_fa_para]", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage:TEP: Tehran English-Persian parallel corpus\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nTEP: Tehran English-Persian parallel corpus. The first free Eng-Per corpus, provided by the Natural Language and Text Processing Laboratory, University of Tehran.", "### Supported Tasks and Leaderboards\n\nThe underlying task is machine translation for language pair English-Persian", "### Languages\n\nEnglish, Persian", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\n\n\n\nM. T. Pilevar, H. Faili, and A. H. Pilevar, “TEP: Tehran English-Persian Parallel Corpus”, in proceedings of 12th International Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2011).", "### Contributions\n\nThanks to @spatil6 for adding this dataset." ]
[ 76, 15, 120, 34, 42, 25, 8, 6, 6, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 67, 17 ]
[ "passage: TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-translation #size_categories-100K<n<1M #source_datasets-original #language-English #language-Persian #license-unknown #region-us \n# Dataset Card for [tep_en_fa_para]## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage:TEP: Tehran English-Persian parallel corpus\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:### Dataset Summary\n\nTEP: Tehran English-Persian parallel corpus. The first free Eng-Per corpus, provided by the Natural Language and Text Processing Laboratory, University of Tehran.### Supported Tasks and Leaderboards\n\nThe underlying task is machine translation for language pair English-Persian### Languages\n\nEnglish, Persian## Dataset Structure### Data Instances### Data Fields### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators" ]
4ab0e8edd74c33299bd99e3e2d47906d3a3bbcf8
# Dataset Card for text2log ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** - **Repository:** [GitHub](https://github.com/alevkov/text2log) - **Paper:** - **Leaderboard:** - **Point of Contact:** https://github.com/alevkov ### Dataset Summary The dataset contains 100,000 simple English sentences selected and filtered from `enTenTen15` and their translation into First Order Logic (FOL) using `ccg2lambda`. ### Supported Tasks and Leaderboards 'semantic-parsing': The data set is used to train models which can generate FOL statements from natural language text ### Languages en-US ## Dataset Structure ### Data Instances ``` { 'clean':'All things that are new are good.', 'trans':'all x1.(_thing(x1) -> (_new(x1) -> _good(x1)))' } ``` ### Data Fields - 'clean': a simple English sentence - 'trans': the corresponding translation into Lambda Dependency-based Compositional Semantics ### Data Splits No predefined train/test split is given. The authors used a 80/20 split ## Dataset Creation ### Curation Rationale The text2log data set is used to improve FOL statement generation from natural text ### Source Data #### Initial Data Collection and Normalization Short text samples selected from enTenTen15 #### Who are the source language producers? See https://www.sketchengine.eu/ententen-english-corpus/ ### Annotations #### Annotation process Machine generated using https://github.com/mynlp/ccg2lambda #### Who are the annotators? none ### Personal and Sensitive Information The dataset does not contain personal or sensitive information. ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information None given ### Citation Information ```bibtex @INPROCEEDINGS{9401852, author={Levkovskyi, Oleksii and Li, Wei}, booktitle={SoutheastCon 2021}, title={Generating Predicate Logic Expressions from Natural Language}, year={2021}, volume={}, number={}, pages={1-8}, doi={10.1109/SoutheastCon45413.2021.9401852} } ``` ### Contributions Thanks to [@apergo-ai](https://github.com/apergo-ai) for adding this dataset.
text2log
[ "task_categories:translation", "annotations_creators:machine-generated", "language_creators:machine-generated", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:en", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["machine-generated"], "language_creators": ["machine-generated"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["translation"], "task_ids": [], "pretty_name": "text2log", "dataset_info": {"features": [{"name": "sentence", "dtype": "string"}, {"name": "fol_translation", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 10358134, "num_examples": 101931}], "download_size": 9746473, "dataset_size": 10358134}}
2024-01-18T11:17:03+00:00
[]
[ "en" ]
TAGS #task_categories-translation #annotations_creators-machine-generated #language_creators-machine-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-unknown #region-us
# Dataset Card for text2log ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information ## Dataset Description - Homepage: - Repository: GitHub - Paper: - Leaderboard: - Point of Contact: URL ### Dataset Summary The dataset contains 100,000 simple English sentences selected and filtered from 'enTenTen15' and their translation into First Order Logic (FOL) using 'ccg2lambda'. ### Supported Tasks and Leaderboards 'semantic-parsing': The data set is used to train models which can generate FOL statements from natural language text ### Languages en-US ## Dataset Structure ### Data Instances ### Data Fields - 'clean': a simple English sentence - 'trans': the corresponding translation into Lambda Dependency-based Compositional Semantics ### Data Splits No predefined train/test split is given. The authors used a 80/20 split ## Dataset Creation ### Curation Rationale The text2log data set is used to improve FOL statement generation from natural text ### Source Data #### Initial Data Collection and Normalization Short text samples selected from enTenTen15 #### Who are the source language producers? See URL ### Annotations #### Annotation process Machine generated using URL #### Who are the annotators? none ### Personal and Sensitive Information The dataset does not contain personal or sensitive information. ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information None given ### Contributions Thanks to @apergo-ai for adding this dataset.
[ "# Dataset Card for text2log", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information", "## Dataset Description\n\n- Homepage:\n- Repository: GitHub\n- Paper:\n- Leaderboard:\n- Point of Contact: URL", "### Dataset Summary\n\nThe dataset contains 100,000 simple English sentences selected and filtered from 'enTenTen15' and their translation into First Order Logic (FOL) using 'ccg2lambda'.", "### Supported Tasks and Leaderboards\n\n'semantic-parsing': The data set is used to train models which can generate FOL statements from natural language text", "### Languages\n\nen-US", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- 'clean': a simple English sentence\n- 'trans': the corresponding translation into Lambda Dependency-based Compositional Semantics", "### Data Splits\n\nNo predefined train/test split is given. The authors used a 80/20 split", "## Dataset Creation", "### Curation Rationale\n\nThe text2log data set is used to improve FOL statement generation from natural text", "### Source Data", "#### Initial Data Collection and Normalization\n\nShort text samples selected from enTenTen15", "#### Who are the source language producers?\n\nSee URL", "### Annotations", "#### Annotation process\n\nMachine generated using URL", "#### Who are the annotators?\n\nnone", "### Personal and Sensitive Information\n\nThe dataset does not contain personal or sensitive information.", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nNone given", "### Contributions\n\nThanks to @apergo-ai for adding this dataset." ]
[ "TAGS\n#task_categories-translation #annotations_creators-machine-generated #language_creators-machine-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-unknown #region-us \n", "# Dataset Card for text2log", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information", "## Dataset Description\n\n- Homepage:\n- Repository: GitHub\n- Paper:\n- Leaderboard:\n- Point of Contact: URL", "### Dataset Summary\n\nThe dataset contains 100,000 simple English sentences selected and filtered from 'enTenTen15' and their translation into First Order Logic (FOL) using 'ccg2lambda'.", "### Supported Tasks and Leaderboards\n\n'semantic-parsing': The data set is used to train models which can generate FOL statements from natural language text", "### Languages\n\nen-US", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- 'clean': a simple English sentence\n- 'trans': the corresponding translation into Lambda Dependency-based Compositional Semantics", "### Data Splits\n\nNo predefined train/test split is given. The authors used a 80/20 split", "## Dataset Creation", "### Curation Rationale\n\nThe text2log data set is used to improve FOL statement generation from natural text", "### Source Data", "#### Initial Data Collection and Normalization\n\nShort text samples selected from enTenTen15", "#### Who are the source language producers?\n\nSee URL", "### Annotations", "#### Annotation process\n\nMachine generated using URL", "#### Who are the annotators?\n\nnone", "### Personal and Sensitive Information\n\nThe dataset does not contain personal or sensitive information.", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nNone given", "### Contributions\n\nThanks to @apergo-ai for adding this dataset." ]
[ 78, 8, 112, 28, 50, 37, 7, 6, 6, 38, 25, 5, 24, 4, 20, 12, 5, 10, 11, 19, 8, 7, 8, 7, 5, 6, 9, 19 ]
[ "passage: TAGS\n#task_categories-translation #annotations_creators-machine-generated #language_creators-machine-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-unknown #region-us \n# Dataset Card for text2log## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information## Dataset Description\n\n- Homepage:\n- Repository: GitHub\n- Paper:\n- Leaderboard:\n- Point of Contact: URL### Dataset Summary\n\nThe dataset contains 100,000 simple English sentences selected and filtered from 'enTenTen15' and their translation into First Order Logic (FOL) using 'ccg2lambda'.### Supported Tasks and Leaderboards\n\n'semantic-parsing': The data set is used to train models which can generate FOL statements from natural language text### Languages\n\nen-US## Dataset Structure### Data Instances### Data Fields\n\n- 'clean': a simple English sentence\n- 'trans': the corresponding translation into Lambda Dependency-based Compositional Semantics### Data Splits\n\nNo predefined train/test split is given. The authors used a 80/20 split## Dataset Creation### Curation Rationale\n\nThe text2log data set is used to improve FOL statement generation from natural text### Source Data#### Initial Data Collection and Normalization\n\nShort text samples selected from enTenTen15#### Who are the source language producers?\n\nSee URL### Annotations#### Annotation process\n\nMachine generated using URL#### Who are the annotators?\n\nnone### Personal and Sensitive Information\n\nThe dataset does not contain personal or sensitive information." ]
aa021e41d0ee6dbee2975fbed620ec8c586bdaf6
# Dataset Card for `thai_toxicity_tweet` ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://github.com/tmu-nlp/ThaiToxicityTweetCorpus/ - **Repository:** https://github.com/tmu-nlp/ThaiToxicityTweetCorpus/ - **Paper:** https://www.ta-cos.org/sites/ta-cos.org/files/1_W32.pdf - **Leaderboard:** - **Point of Contact:** https://www.ta-cos.org/sites/ta-cos.org/files/1_W32.pdf ### Dataset Summary Thai Toxicity Tweet Corpus contains 3,300 tweets (506 tweets with texts missing) annotated by humans with guidelines including a 44-word dictionary. The author obtained 2,027 and 1,273 toxic and non-toxic tweets, respectively; these were labeled by three annotators. The result of corpus analysis indicates that tweets that include toxic words are not always toxic. Further, it is more likely that a tweet is toxic, if it contains toxic words indicating their original meaning. Moreover, disagreements in annotation are primarily because of sarcasm, unclear existing target, and word sense ambiguity. Notes from data cleaner: The data is included into [huggingface/datasets](https://www.github.com/huggingface/datasets) in Dec 2020. By this time, 506 of the tweets are not available publicly anymore. We denote these by `TWEET_NOT_FOUND` in `tweet_text`. Processing can be found at [this PR](https://github.com/tmu-nlp/ThaiToxicityTweetCorpus/pull/1). ### Supported Tasks and Leaderboards text classification ### Languages Thai (`th`) ## Dataset Structure ### Data Instances ``` {'is_toxic': 0, 'nontoxic_votes': 3, 'toxic_votes': 0, 'tweet_id': '898576382384418817', 'tweet_text': 'วันๆ นี่คุยกะหมา แมว หมู ไก่ ม้า ควาย มากกว่าคุยกับคนไปละ'} {'is_toxic': 1, 'nontoxic_votes': 0, 'toxic_votes': 3, 'tweet_id': '898573084981985280', 'tweet_text': 'ควายแดงเมิงด่ารัฐบาลจนรองนายกป่วย พวกมึงกำลังทำลายชาติรู้มั้ย มั้ย มั้ย มั้ยยยยยยยยย news.voicetv.co.th/thailand/51672…'} ``` ### Data Fields "tweet_id": Id of tweet on Twitter "tweet_text": text of the tweet "toxic_votes": how many annotators say it is toxic, out of 3 annotators "nontoxic_votes": how many annotators say it is NOT toxic, out of 3 annotators "is_toxic": 1 if tweet is toxic else 0 (majority rules) ### Data Splits No explicit split is given. ## Dataset Creation ### Curation Rationale The dataset is created as part of [Sirihattasak et al (2019)](https://www.ta-cos.org/sites/ta-cos.org/files/1_W32.pdf). ### Source Data #### Initial Data Collection and Normalization The authors used the public Twitter Search API to collect 9,819 tweets from January–December 2017 based on our keyword dictionary. Then, they selected 75 tweets for each keyword. In total, they collected 3,300 tweets for annotation. To ensure quality of data, they set the following selection criteria. 1. All tweets are selected by humans to prevent word ambiguity. (The Twitter API selected the tweets based on characters in the keyword. For example, in the case of “บ้า(crazy),” the API will also select “บ้านนอก” (countryside)” which is not our target.) 2. The length of the tweet should be sufficiently long to discern the context of the tweet. Hence, they set five words as the minimum limit. 3. The tweets that contain only extremely toxic words, (for example: “damn, retard, bitch, f*ck, slut!!!”) are not considered. 4. In addition, they allowed tweets with English words if they were not critical elements in the labeling decision, for example, the word “f*ck.” As a result, our corpus contains English words, but they are less than 2% of the total. All hashtags, re-tweets, and links were removed from these tweets. However, they did not delete emoticons because these emotional icons can imply the real intent of the post owners. Furthermore, only in the case of annotation, some entries such as the names of famous people were replaced with a tag <ไม่ขอเปิดเผยชื่อ>, for anonymity to prevent individual bias. #### Who are the source language producers? Twitter users in Thailand ### Annotations #### Annotation process We manually annotated our dataset with two labels: Toxic and Non-Toxic. We define a message as toxic if it indicates any harmful, damage, or negative intent based on our definition of toxicity. Furthermore, all the tweets were annotated by three annotators to identify toxicity; the conditions used for this identification are presented in the following list. - A toxic message is a message that should be deleted or not be allowed in public. - A message’s target or consequence must exist. It can either be an individual or a generalized group based on a commonality such as religion or ethnicity, or an entire community. - Self-complain is not considered toxic, because it is not harmful to anyone. However, if self-complain is intended to indicate something bad, it will be considered as toxic. - Both direct and indirect messages including those with sarcasm are taken into consideration. We strictly instructed all the annotators about these concepts and asked them to perform a small test to ensure they understood these conditions. The annotation process was divided into two rounds. We asked the candidates to annotate their answers in the first round to learn our annotation standard. Then, we asked them to annotate a different dataset and selected the ones who obtained a full-score for the second round as an annotator. From among these annotators, 20% of the annotators failed the first round and were not involved in the final annotation. #### Who are the annotators? Three annotators hired by [Sirihattasak et al (2019)](https://www.ta-cos.org/sites/ta-cos.org/files/1_W32.pdf) ### Personal and Sensitive Information Despite all tweets being public, due to the nature of toxic tweets, there might be personal attacks and toxic language used. ## Considerations for Using the Data ### Social Impact of Dataset - toxic social media message classification dataset ### Discussion of Biases - Users are masked before annotation by the annotators to prevent biases based on tweet authors ### Other Known Limitations - The data is included into [huggingface/datasets](https://www.github.com/huggingface/datasets) in Dec 2020. By this time, 506 of the tweets are not available publicly anymore. We denote these by `TWEET_NOT_FOUND` in `tweet_text`. ## Additional Information ### Dataset Curators [Sirihattasak et al (2019)](https://www.ta-cos.org/sites/ta-cos.org/files/1_W32.pdf) ### Licensing Information CC-BY-NC 3.0 ### Citation Information Please cite the following if you make use of the dataset: ``` @article{sirihattasak2019annotation, title={Annotation and Classification of Toxicity for Thai Twitter}, author={Sirihattasak, Sugan and Komachi, Mamoru and Ishikawa, Hiroshi}, year={2019} } ``` ### Contributions Thanks to [@cstorm125](https://github.com/cstorm125) for adding this dataset.
thai_toxicity_tweet
[ "task_categories:text-classification", "task_ids:sentiment-classification", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:th", "license:cc-by-nc-3.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found"], "language": ["th"], "license": ["cc-by-nc-3.0"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["sentiment-classification"], "pretty_name": "ThaiToxicityTweet", "dataset_info": {"features": [{"name": "tweet_id", "dtype": "string"}, {"name": "tweet_text", "dtype": "string"}, {"name": "toxic_votes", "dtype": "int32"}, {"name": "nontoxic_votes", "dtype": "int32"}, {"name": "is_toxic", "dtype": {"class_label": {"names": {"0": "neg", "1": "pos"}}}}], "config_name": "thai_toxicity_tweet", "splits": [{"name": "train", "num_bytes": 637387, "num_examples": 3300}], "download_size": 194740, "dataset_size": 637387}}
2024-01-18T11:17:04+00:00
[]
[ "th" ]
TAGS #task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Thai #license-cc-by-nc-3.0 #region-us
# Dataset Card for 'thai_toxicity_tweet' ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: URL - Paper: URL - Leaderboard: - Point of Contact: URL ### Dataset Summary Thai Toxicity Tweet Corpus contains 3,300 tweets (506 tweets with texts missing) annotated by humans with guidelines including a 44-word dictionary. The author obtained 2,027 and 1,273 toxic and non-toxic tweets, respectively; these were labeled by three annotators. The result of corpus analysis indicates that tweets that include toxic words are not always toxic. Further, it is more likely that a tweet is toxic, if it contains toxic words indicating their original meaning. Moreover, disagreements in annotation are primarily because of sarcasm, unclear existing target, and word sense ambiguity. Notes from data cleaner: The data is included into huggingface/datasets in Dec 2020. By this time, 506 of the tweets are not available publicly anymore. We denote these by 'TWEET_NOT_FOUND' in 'tweet_text'. Processing can be found at this PR. ### Supported Tasks and Leaderboards text classification ### Languages Thai ('th') ## Dataset Structure ### Data Instances ### Data Fields "tweet_id": Id of tweet on Twitter "tweet_text": text of the tweet "toxic_votes": how many annotators say it is toxic, out of 3 annotators "nontoxic_votes": how many annotators say it is NOT toxic, out of 3 annotators "is_toxic": 1 if tweet is toxic else 0 (majority rules) ### Data Splits No explicit split is given. ## Dataset Creation ### Curation Rationale The dataset is created as part of Sirihattasak et al (2019). ### Source Data #### Initial Data Collection and Normalization The authors used the public Twitter Search API to collect 9,819 tweets from January–December 2017 based on our keyword dictionary. Then, they selected 75 tweets for each keyword. In total, they collected 3,300 tweets for annotation. To ensure quality of data, they set the following selection criteria. 1. All tweets are selected by humans to prevent word ambiguity. (The Twitter API selected the tweets based on characters in the keyword. For example, in the case of “บ้า(crazy),” the API will also select “บ้านนอก” (countryside)” which is not our target.) 2. The length of the tweet should be sufficiently long to discern the context of the tweet. Hence, they set five words as the minimum limit. 3. The tweets that contain only extremely toxic words, (for example: “damn, retard, bitch, f*ck, slut!!!”) are not considered. 4. In addition, they allowed tweets with English words if they were not critical elements in the labeling decision, for example, the word “f*ck.” As a result, our corpus contains English words, but they are less than 2% of the total. All hashtags, re-tweets, and links were removed from these tweets. However, they did not delete emoticons because these emotional icons can imply the real intent of the post owners. Furthermore, only in the case of annotation, some entries such as the names of famous people were replaced with a tag <ไม่ขอเปิดเผยชื่อ>, for anonymity to prevent individual bias. #### Who are the source language producers? Twitter users in Thailand ### Annotations #### Annotation process We manually annotated our dataset with two labels: Toxic and Non-Toxic. We define a message as toxic if it indicates any harmful, damage, or negative intent based on our definition of toxicity. Furthermore, all the tweets were annotated by three annotators to identify toxicity; the conditions used for this identification are presented in the following list. - A toxic message is a message that should be deleted or not be allowed in public. - A message’s target or consequence must exist. It can either be an individual or a generalized group based on a commonality such as religion or ethnicity, or an entire community. - Self-complain is not considered toxic, because it is not harmful to anyone. However, if self-complain is intended to indicate something bad, it will be considered as toxic. - Both direct and indirect messages including those with sarcasm are taken into consideration. We strictly instructed all the annotators about these concepts and asked them to perform a small test to ensure they understood these conditions. The annotation process was divided into two rounds. We asked the candidates to annotate their answers in the first round to learn our annotation standard. Then, we asked them to annotate a different dataset and selected the ones who obtained a full-score for the second round as an annotator. From among these annotators, 20% of the annotators failed the first round and were not involved in the final annotation. #### Who are the annotators? Three annotators hired by Sirihattasak et al (2019) ### Personal and Sensitive Information Despite all tweets being public, due to the nature of toxic tweets, there might be personal attacks and toxic language used. ## Considerations for Using the Data ### Social Impact of Dataset - toxic social media message classification dataset ### Discussion of Biases - Users are masked before annotation by the annotators to prevent biases based on tweet authors ### Other Known Limitations - The data is included into huggingface/datasets in Dec 2020. By this time, 506 of the tweets are not available publicly anymore. We denote these by 'TWEET_NOT_FOUND' in 'tweet_text'. ## Additional Information ### Dataset Curators Sirihattasak et al (2019) ### Licensing Information CC-BY-NC 3.0 Please cite the following if you make use of the dataset: ### Contributions Thanks to @cstorm125 for adding this dataset.
[ "# Dataset Card for 'thai_toxicity_tweet'", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: URL\n- Leaderboard: \n- Point of Contact: URL", "### Dataset Summary\n\nThai Toxicity Tweet Corpus contains 3,300 tweets (506 tweets with texts missing) annotated by humans with guidelines including a 44-word dictionary.\nThe author obtained 2,027 and 1,273 toxic and non-toxic tweets, respectively; these were labeled by three annotators. The result of corpus\nanalysis indicates that tweets that include toxic words are not always toxic. Further, it is more likely that a tweet is toxic, if it contains\ntoxic words indicating their original meaning. Moreover, disagreements in annotation are primarily because of sarcasm, unclear existing\ntarget, and word sense ambiguity.\n\nNotes from data cleaner: The data is included into huggingface/datasets in Dec 2020. By this time, 506 of the tweets are not available publicly anymore. We denote these by 'TWEET_NOT_FOUND' in 'tweet_text'.\nProcessing can be found at this PR.", "### Supported Tasks and Leaderboards\n\ntext classification", "### Languages\n\nThai ('th')", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n\"tweet_id\": Id of tweet on Twitter\n\"tweet_text\": text of the tweet\n\"toxic_votes\": how many annotators say it is toxic, out of 3 annotators\n\"nontoxic_votes\": how many annotators say it is NOT toxic, out of 3 annotators\n\"is_toxic\": 1 if tweet is toxic else 0 (majority rules)", "### Data Splits\n\nNo explicit split is given.", "## Dataset Creation", "### Curation Rationale\n\nThe dataset is created as part of Sirihattasak et al (2019).", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe authors used the public Twitter Search API to collect 9,819 tweets from January–December 2017 based on our keyword dictionary. Then, they selected 75 tweets for each keyword. In total, they collected 3,300 tweets for annotation. To ensure quality of data, they set the following selection criteria.\n\n1. All tweets are selected by humans to prevent word ambiguity. (The Twitter API selected the tweets based on characters in the keyword. For example, in the case of “บ้า(crazy),” the API will also select “บ้านนอก” (countryside)” which is not our target.) \n2. The length of the tweet should be sufficiently long to discern the context of the tweet. Hence, they set five words as the minimum limit.\n3. The tweets that contain only extremely toxic words, (for example: “damn, retard, bitch, f*ck, slut!!!”) are not considered.\n4. In addition, they allowed tweets with English words if they were not critical elements in the labeling decision, for example, the word “f*ck.” As a result, our corpus contains English words, but they are less than 2% of the total.\n\nAll hashtags, re-tweets, and links were removed from these tweets. However, they did not delete emoticons because these emotional icons can imply the real intent of the post owners. Furthermore, only in the case of annotation, some entries such as the names of famous people were replaced with a tag <ไม่ขอเปิดเผยชื่อ>, for anonymity to prevent individual bias.", "#### Who are the source language producers?\n\nTwitter users in Thailand", "### Annotations", "#### Annotation process\n\nWe manually annotated our dataset with two labels: Toxic and Non-Toxic. We define a message as toxic if it indicates any harmful, damage, or negative intent based on our definition of toxicity. Furthermore, all the tweets were annotated by three annotators to identify toxicity; the conditions used for this identification are presented in the following list.\n\n- A toxic message is a message that should be deleted or not be allowed in public.\n- A message’s target or consequence must exist. It can either be an individual or a generalized group based on a commonality such as religion or ethnicity, or an entire community.\n- Self-complain is not considered toxic, because it is not harmful to anyone. However, if self-complain is intended to indicate something bad, it will be considered as toxic.\n- Both direct and indirect messages including those with sarcasm are taken into consideration.\n\nWe strictly instructed all the annotators about these concepts and asked them to perform a small test to ensure they understood these conditions. The annotation process was divided into two rounds. We asked the candidates to annotate their answers in the first round to learn our annotation standard. Then, we asked them to annotate a different dataset and selected the ones who obtained a full-score for the second round as an annotator. From among these annotators, 20% of the annotators failed the first round and were not involved in the final annotation.", "#### Who are the annotators?\n\nThree annotators hired by Sirihattasak et al (2019)", "### Personal and Sensitive Information\n\nDespite all tweets being public, due to the nature of toxic tweets, there might be personal attacks and toxic language used.", "## Considerations for Using the Data", "### Social Impact of Dataset\n\n- toxic social media message classification dataset", "### Discussion of Biases\n\n- Users are masked before annotation by the annotators to prevent biases based on tweet authors", "### Other Known Limitations\n\n- The data is included into huggingface/datasets in Dec 2020. By this time, 506 of the tweets are not available publicly anymore. We denote these by 'TWEET_NOT_FOUND' in 'tweet_text'.", "## Additional Information", "### Dataset Curators\n\nSirihattasak et al (2019)", "### Licensing Information\n\nCC-BY-NC 3.0\n\n\n\nPlease cite the following if you make use of the dataset:", "### Contributions\n\nThanks to @cstorm125 for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Thai #license-cc-by-nc-3.0 #region-us \n", "# Dataset Card for 'thai_toxicity_tweet'", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: URL\n- Leaderboard: \n- Point of Contact: URL", "### Dataset Summary\n\nThai Toxicity Tweet Corpus contains 3,300 tweets (506 tweets with texts missing) annotated by humans with guidelines including a 44-word dictionary.\nThe author obtained 2,027 and 1,273 toxic and non-toxic tweets, respectively; these were labeled by three annotators. The result of corpus\nanalysis indicates that tweets that include toxic words are not always toxic. Further, it is more likely that a tweet is toxic, if it contains\ntoxic words indicating their original meaning. Moreover, disagreements in annotation are primarily because of sarcasm, unclear existing\ntarget, and word sense ambiguity.\n\nNotes from data cleaner: The data is included into huggingface/datasets in Dec 2020. By this time, 506 of the tweets are not available publicly anymore. We denote these by 'TWEET_NOT_FOUND' in 'tweet_text'.\nProcessing can be found at this PR.", "### Supported Tasks and Leaderboards\n\ntext classification", "### Languages\n\nThai ('th')", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n\"tweet_id\": Id of tweet on Twitter\n\"tweet_text\": text of the tweet\n\"toxic_votes\": how many annotators say it is toxic, out of 3 annotators\n\"nontoxic_votes\": how many annotators say it is NOT toxic, out of 3 annotators\n\"is_toxic\": 1 if tweet is toxic else 0 (majority rules)", "### Data Splits\n\nNo explicit split is given.", "## Dataset Creation", "### Curation Rationale\n\nThe dataset is created as part of Sirihattasak et al (2019).", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe authors used the public Twitter Search API to collect 9,819 tweets from January–December 2017 based on our keyword dictionary. Then, they selected 75 tweets for each keyword. In total, they collected 3,300 tweets for annotation. To ensure quality of data, they set the following selection criteria.\n\n1. All tweets are selected by humans to prevent word ambiguity. (The Twitter API selected the tweets based on characters in the keyword. For example, in the case of “บ้า(crazy),” the API will also select “บ้านนอก” (countryside)” which is not our target.) \n2. The length of the tweet should be sufficiently long to discern the context of the tweet. Hence, they set five words as the minimum limit.\n3. The tweets that contain only extremely toxic words, (for example: “damn, retard, bitch, f*ck, slut!!!”) are not considered.\n4. In addition, they allowed tweets with English words if they were not critical elements in the labeling decision, for example, the word “f*ck.” As a result, our corpus contains English words, but they are less than 2% of the total.\n\nAll hashtags, re-tweets, and links were removed from these tweets. However, they did not delete emoticons because these emotional icons can imply the real intent of the post owners. Furthermore, only in the case of annotation, some entries such as the names of famous people were replaced with a tag <ไม่ขอเปิดเผยชื่อ>, for anonymity to prevent individual bias.", "#### Who are the source language producers?\n\nTwitter users in Thailand", "### Annotations", "#### Annotation process\n\nWe manually annotated our dataset with two labels: Toxic and Non-Toxic. We define a message as toxic if it indicates any harmful, damage, or negative intent based on our definition of toxicity. Furthermore, all the tweets were annotated by three annotators to identify toxicity; the conditions used for this identification are presented in the following list.\n\n- A toxic message is a message that should be deleted or not be allowed in public.\n- A message’s target or consequence must exist. It can either be an individual or a generalized group based on a commonality such as religion or ethnicity, or an entire community.\n- Self-complain is not considered toxic, because it is not harmful to anyone. However, if self-complain is intended to indicate something bad, it will be considered as toxic.\n- Both direct and indirect messages including those with sarcasm are taken into consideration.\n\nWe strictly instructed all the annotators about these concepts and asked them to perform a small test to ensure they understood these conditions. The annotation process was divided into two rounds. We asked the candidates to annotate their answers in the first round to learn our annotation standard. Then, we asked them to annotate a different dataset and selected the ones who obtained a full-score for the second round as an annotator. From among these annotators, 20% of the annotators failed the first round and were not involved in the final annotation.", "#### Who are the annotators?\n\nThree annotators hired by Sirihattasak et al (2019)", "### Personal and Sensitive Information\n\nDespite all tweets being public, due to the nature of toxic tweets, there might be personal attacks and toxic language used.", "## Considerations for Using the Data", "### Social Impact of Dataset\n\n- toxic social media message classification dataset", "### Discussion of Biases\n\n- Users are masked before annotation by the annotators to prevent biases based on tweet authors", "### Other Known Limitations\n\n- The data is included into huggingface/datasets in Dec 2020. By this time, 506 of the tweets are not available publicly anymore. We denote these by 'TWEET_NOT_FOUND' in 'tweet_text'.", "## Additional Information", "### Dataset Curators\n\nSirihattasak et al (2019)", "### Licensing Information\n\nCC-BY-NC 3.0\n\n\n\nPlease cite the following if you make use of the dataset:", "### Contributions\n\nThanks to @cstorm125 for adding this dataset." ]
[ 93, 13, 120, 28, 222, 13, 10, 6, 6, 96, 11, 5, 23, 4, 355, 14, 5, 342, 24, 38, 8, 17, 32, 62, 5, 14, 25, 17 ]
[ "passage: TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Thai #license-cc-by-nc-3.0 #region-us \n# Dataset Card for 'thai_toxicity_tweet'## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: URL\n- Leaderboard: \n- Point of Contact: URL### Dataset Summary\n\nThai Toxicity Tweet Corpus contains 3,300 tweets (506 tweets with texts missing) annotated by humans with guidelines including a 44-word dictionary.\nThe author obtained 2,027 and 1,273 toxic and non-toxic tweets, respectively; these were labeled by three annotators. The result of corpus\nanalysis indicates that tweets that include toxic words are not always toxic. Further, it is more likely that a tweet is toxic, if it contains\ntoxic words indicating their original meaning. Moreover, disagreements in annotation are primarily because of sarcasm, unclear existing\ntarget, and word sense ambiguity.\n\nNotes from data cleaner: The data is included into huggingface/datasets in Dec 2020. By this time, 506 of the tweets are not available publicly anymore. We denote these by 'TWEET_NOT_FOUND' in 'tweet_text'.\nProcessing can be found at this PR.### Supported Tasks and Leaderboards\n\ntext classification### Languages\n\nThai ('th')## Dataset Structure", "passage: ### Data Instances### Data Fields\n\n\"tweet_id\": Id of tweet on Twitter\n\"tweet_text\": text of the tweet\n\"toxic_votes\": how many annotators say it is toxic, out of 3 annotators\n\"nontoxic_votes\": how many annotators say it is NOT toxic, out of 3 annotators\n\"is_toxic\": 1 if tweet is toxic else 0 (majority rules)### Data Splits\n\nNo explicit split is given.## Dataset Creation### Curation Rationale\n\nThe dataset is created as part of Sirihattasak et al (2019).### Source Data#### Initial Data Collection and Normalization\n\nThe authors used the public Twitter Search API to collect 9,819 tweets from January–December 2017 based on our keyword dictionary. Then, they selected 75 tweets for each keyword. In total, they collected 3,300 tweets for annotation. To ensure quality of data, they set the following selection criteria.\n\n1. All tweets are selected by humans to prevent word ambiguity. (The Twitter API selected the tweets based on characters in the keyword. For example, in the case of “บ้า(crazy),” the API will also select “บ้านนอก” (countryside)” which is not our target.) \n2. The length of the tweet should be sufficiently long to discern the context of the tweet. Hence, they set five words as the minimum limit.\n3. The tweets that contain only extremely toxic words, (for example: “damn, retard, bitch, f*ck, slut!!!”) are not considered.\n4. In addition, they allowed tweets with English words if they were not critical elements in the labeling decision, for example, the word “f*ck.” As a result, our corpus contains English words, but they are less than 2% of the total.\n\nAll hashtags, re-tweets, and links were removed from these tweets. However, they did not delete emoticons because these emotional icons can imply the real intent of the post owners. Furthermore, only in the case of annotation, some entries such as the names of famous people were replaced with a tag <ไม่ขอเปิดเผยชื่อ>, for anonymity to prevent individual bias.", "passage: #### Who are the source language producers?\n\nTwitter users in Thailand### Annotations#### Annotation process\n\nWe manually annotated our dataset with two labels: Toxic and Non-Toxic. We define a message as toxic if it indicates any harmful, damage, or negative intent based on our definition of toxicity. Furthermore, all the tweets were annotated by three annotators to identify toxicity; the conditions used for this identification are presented in the following list.\n\n- A toxic message is a message that should be deleted or not be allowed in public.\n- A message’s target or consequence must exist. It can either be an individual or a generalized group based on a commonality such as religion or ethnicity, or an entire community.\n- Self-complain is not considered toxic, because it is not harmful to anyone. However, if self-complain is intended to indicate something bad, it will be considered as toxic.\n- Both direct and indirect messages including those with sarcasm are taken into consideration.\n\nWe strictly instructed all the annotators about these concepts and asked them to perform a small test to ensure they understood these conditions. The annotation process was divided into two rounds. We asked the candidates to annotate their answers in the first round to learn our annotation standard. Then, we asked them to annotate a different dataset and selected the ones who obtained a full-score for the second round as an annotator. From among these annotators, 20% of the annotators failed the first round and were not involved in the final annotation.#### Who are the annotators?\n\nThree annotators hired by Sirihattasak et al (2019)### Personal and Sensitive Information\n\nDespite all tweets being public, due to the nature of toxic tweets, there might be personal attacks and toxic language used.## Considerations for Using the Data### Social Impact of Dataset\n\n- toxic social media message classification dataset### Discussion of Biases\n\n- Users are masked before annotation by the annotators to prevent biases based on tweet authors" ]
048e01a12f5bde5988f5f9b34be694c9e451ffa0
# Dataset Card for `thainer` ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://github.com/wannaphong/thai-ner - **Repository:** https://github.com/wannaphong/thai-ner - **Paper:** - **Leaderboard:** - **Point of Contact:** https://github.com/wannaphong/ ### Dataset Summary ThaiNER (v1.3) is a 6,456-sentence named entity recognition dataset created from expanding the 2,258-sentence [unnamed dataset](http://pioneer.chula.ac.th/~awirote/Data-Nutcha.zip) by [Tirasaroj and Aroonmanakun (2012)](http://pioneer.chula.ac.th/~awirote/publications/). It is used to train NER taggers in [PyThaiNLP](https://github.com/PyThaiNLP/pythainlp). The NER tags are annotated by [Tirasaroj and Aroonmanakun (2012)]((http://pioneer.chula.ac.th/~awirote/publications/)) for 2,258 sentences and the rest by [@wannaphong](https://github.com/wannaphong/). The POS tags are done by [PyThaiNLP](https://github.com/PyThaiNLP/pythainlp)'s `perceptron` engine trained on `orchid_ud`. [@wannaphong](https://github.com/wannaphong/) is now the only maintainer of this dataset. ### Supported Tasks and Leaderboards - named entity recognition - pos tagging ### Languages Thai ## Dataset Structure ### Data Instances ``` {'id': 100, 'ner_tags': [27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27], 'pos_tags': [6, 12, 13, 1, 6, 5, 11, 7, 11, 6, 5, 13, 6, 6, 6, 11, 6, 6, 11, 6, 6, 11, 6, 6, 13, 6, 11, 11, 6, 11, 6, 11, 6, 11, 6, 11, 11, 6, 6, 11, 12, 6, 13, 5, 11, 7, 11, 6, 3, 11, 12, 3, 13, 6, 1, 6, 12, 13, 1, 6, 6, 5, 11, 3, 11, 5, 4, 6, 13, 6, 13, 6, 10, 3, 13, 13, 12, 13, 12, 0, 1, 10, 11, 6, 6, 11, 6, 11, 6, 12, 13, 5, 12, 3, 13, 13, 1, 6, 1, 6, 13], 'tokens': ['เชื้อโรค', 'ที่', 'ปรากฏ', 'ใน', 'สัตว์', 'ทั้ง', ' ', '4', ' ', 'ชนิด', 'นี้', 'เป็น', 'เชื้อ', 'โรคไข้หวัด', 'นก', ' ', 'เอช', 'พี', ' ', 'เอ', 'เวียน', ' ', 'อิน', 'ฟลู', 'เอน', 'ซา', ' ', '(', 'Hight', ' ', 'Polygenic', ' ', 'Avain', ' ', 'Influenza', ')', ' ', 'ชนิด', 'รุนแรง', ' ', 'ซึ่ง', 'การ', 'ตั้งชื่อ', 'ทั้ง', ' ', '4', ' ', 'ขึ้น', 'มา', ' ', 'เพื่อที่จะ', 'สามารถ', 'ระบุ', 'เชื้อ', 'ของ', 'ไวรัส', 'ที่', 'ทำอันตราย', 'ตาม', 'สิ่งมีชีวิต', 'ประเภท', 'ต่างๆ', ' ', 'ได้', ' ', 'อีก', 'ทั้ง', 'การ', 'ระบุ', 'สถานที่', 'คือ', 'ประเทศ', 'ไทย', 'จะ', 'ทำให้', 'รู้', 'ว่า', 'พบ', 'ที่', 'แรก', 'ใน', 'ไทย', ' ', 'ส่วน', 'วัน', ' ', 'เดือน', ' ', 'ปี', 'ที่', 'พบ', 'นั้น', 'ก็', 'จะ', 'ทำให้', 'ทราบ', 'ถึง', 'ครั้งแรก', 'ของ', 'การ', 'ค้นพบ']} {'id': 107, 'ner_tags': [27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27, 27], 'pos_tags': [0, 1, 6, 5, 11, 12, 3, 3, 13, 6, 13, 12, 0, 2, 12, 11, 6, 5, 13, 6, 5, 1, 6, 6, 1, 10, 11, 4, 13, 6, 11, 12, 6, 6, 10, 11, 13, 6, 1, 6, 4, 6, 1, 6, 6, 11, 4, 6, 1, 5, 6, 12, 2, 13, 6, 6, 5, 1, 11, 12, 13, 1, 6, 6, 11, 13, 11, 6, 6, 6, 11, 11, 6, 11, 11, 4, 10, 11, 11, 6, 11], 'tokens': ['ล่าสุด', 'ใน', 'เรื่อง', 'นี้', ' ', 'ทั้งนี้', 'คง', 'ต้อง', 'มี', 'การ', 'ตรวจสอบ', 'ให้', 'ชัดเจน', 'อีกครั้ง', 'ว่า', ' ', 'ไวรัส', 'นี้', 'เป็น', 'ชนิด', 'เดียว', 'กับ', 'ไข้หวัด', 'นก', 'ใน', 'ไทย', ' ', 'หรือ', 'เป็น', 'การกลายพันธุ์', ' ', 'โดยที่', 'คณะ', 'สัตวแพทย์', 'มหาวิทยาลัยเกษตรศาสตร์', ' ', 'จัด', 'ระดมสมอง', 'จาก', 'คณบดี', 'และ', 'ผู้เชี่ยวชาญ', 'จาก', 'คณะ', 'สัตวแพทย์', ' ', 'และ', 'ปศุสัตว์', 'ของ', 'หลาย', 'มหาวิทยาลัย', 'เพื่อ', 'ร่วมกัน', 'หา', 'ข้อมูล', 'เรื่อง', 'นี้', 'ด้วย', ' ', 'โดย', 'ประสาน', 'กับ', 'เจ้าหน้าที่', 'ระหว่างประเทศ', ' ', 'คือ', ' ', 'องค์การ', 'สุขภาพ', 'สัตว์โลก', ' ', '(', 'OIE', ')', ' ', 'และ', 'องค์การอนามัยโลก', ' ', '(', 'WHO', ')']} ``` ### Data Fields - `id`: sentence id - `tokens`: word tokens by [PyThaiNLP](https://github.com/PyThaiNLP/pythainlp)'s dictionary-based tokenizer `newmm` - `pos_tags`: POS tags tagged by [PyThaiNLP](https://github.com/PyThaiNLP/pythainlp)'s `perceptron` engine trained on `orchid_ud` - `ner_tags`: NER tags tagged by humans ### Data Splits No explicit split is given ## Dataset Creation ### Curation Rationale ThaiNER (v1.3) is a 6,456-sentence named entity recognition dataset created from expanding the 2,258-sentence [unnamed dataset](http://pioneer.chula.ac.th/~awirote/Data-Nutcha.zip) by [Tirasaroj and Aroonmanakun (2012)](http://pioneer.chula.ac.th/~awirote/publications/). It is used to train NER taggers in [PyThaiNLP](https://github.com/PyThaiNLP/pythainlp). ### Source Data #### Initial Data Collection and Normalization The earlier part of the dataset is all news articles, whereas the part added by [@wannaphong](https://github.com/wannaphong/) includes news articles, public announcements and [@wannaphong](https://github.com/wannaphong/)'s own chat messages with personal and sensitive information removed. #### Who are the source language producers? News articles and public announcements are created by their respective authors. Chat messages are created by [@wannaphong](https://github.com/wannaphong/). ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [Tirasaroj and Aroonmanakun (2012)](http://pioneer.chula.ac.th/~awirote/publications/) for the earlier 2,258 sentences and [@wannaphong](https://github.com/wannaphong/) for the rest ### Personal and Sensitive Information News articles and public announcements are not expected to include personal and sensitive information. [@wannaphong](https://github.com/wannaphong/) has removed such information from his own chat messages. ## Considerations for Using the Data ### Social Impact of Dataset - named entity recognition in Thai ### Discussion of Biases Since almost all of collection and annotation is done by [@wannaphong](https://github.com/wannaphong/), his biases are expected to be reflected in the dataset. ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [Tirasaroj and Aroonmanakun (2012)](http://pioneer.chula.ac.th/~awirote/publications/) for the earlier 2,258 sentences and [@wannaphong](https://github.com/wannaphong/) for the rest ### Licensing Information CC-BY 3.0 ### Citation Information ``` @misc{Wannaphong Phatthiyaphaibun_2019, title={wannaphongcom/thai-ner: ThaiNER 1.3}, url={https://zenodo.org/record/3550546}, DOI={10.5281/ZENODO.3550546}, abstractNote={Thai Named Entity Recognition}, publisher={Zenodo}, author={Wannaphong Phatthiyaphaibun}, year={2019}, month={Nov} } ``` Work extended from: [Tirasaroj, N. and Aroonmanakun, W. 2012. Thai NER using CRF model based on surface features. In Proceedings of SNLP-AOS 2011, 9-10 February, 2012, Bangkok, pages 176-180.](http://pioneer.chula.ac.th/~awirote/publications/) ### Contributions Thanks to [@cstorm125](https://github.com/cstorm125) for adding this dataset.
thainer
[ "task_categories:token-classification", "task_ids:named-entity-recognition", "task_ids:part-of-speech", "annotations_creators:expert-generated", "annotations_creators:machine-generated", "language_creators:found", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:extended|other-tirasaroj-aroonmanakun", "language:th", "license:cc-by-3.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated", "machine-generated"], "language_creators": ["found", "expert-generated"], "language": ["th"], "license": ["cc-by-3.0"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["extended|other-tirasaroj-aroonmanakun"], "task_categories": ["token-classification"], "task_ids": ["named-entity-recognition", "part-of-speech"], "pretty_name": "thainer", "dataset_info": {"features": [{"name": "id", "dtype": "int32"}, {"name": "tokens", "sequence": "string"}, {"name": "pos_tags", "sequence": {"class_label": {"names": {"0": "ADJ", "1": "ADP", "2": "ADV", "3": "AUX", "4": "CCONJ", "5": "DET", "6": "NOUN", "7": "NUM", "8": "PART", "9": "PRON", "10": "PROPN", "11": "PUNCT", "12": "SCONJ", "13": "VERB"}}}}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "B-DATE", "1": "B-EMAIL", "2": "B-LAW", "3": "B-LEN", "4": "B-LOCATION", "5": "B-MONEY", "6": "B-ORGANIZATION", "7": "B-PERCENT", "8": "B-PERSON", "9": "B-PHONE", "10": "B-TIME", "11": "B-URL", "12": "B-ZIP", "13": "B-\u0e44\u0e21\u0e48\u0e22\u0e37\u0e19\u0e22\u0e31\u0e19", "14": "I-DATE", "15": "I-EMAIL", "16": "I-LAW", "17": "I-LEN", "18": "I-LOCATION", "19": "I-MONEY", "20": "I-ORGANIZATION", "21": "I-PERCENT", "22": "I-PERSON", "23": "I-PHONE", "24": "I-TIME", "25": "I-URL", "26": "I-\u0e44\u0e21\u0e48\u0e22\u0e37\u0e19\u0e22\u0e31\u0e19", "27": "O"}}}}], "config_name": "thainer", "splits": [{"name": "train", "num_bytes": 8117902, "num_examples": 6348}], "download_size": 5456461, "dataset_size": 8117902}}
2024-01-18T11:17:05+00:00
[]
[ "th" ]
TAGS #task_categories-token-classification #task_ids-named-entity-recognition #task_ids-part-of-speech #annotations_creators-expert-generated #annotations_creators-machine-generated #language_creators-found #language_creators-expert-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-extended|other-tirasaroj-aroonmanakun #language-Thai #license-cc-by-3.0 #region-us
# Dataset Card for 'thainer' ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: URL - Paper: - Leaderboard: - Point of Contact: URL ### Dataset Summary ThaiNER (v1.3) is a 6,456-sentence named entity recognition dataset created from expanding the 2,258-sentence unnamed dataset by Tirasaroj and Aroonmanakun (2012). It is used to train NER taggers in PyThaiNLP. The NER tags are annotated by Tirasaroj and Aroonmanakun (2012)) for 2,258 sentences and the rest by @wannaphong. The POS tags are done by PyThaiNLP's 'perceptron' engine trained on 'orchid_ud'. @wannaphong is now the only maintainer of this dataset. ### Supported Tasks and Leaderboards - named entity recognition - pos tagging ### Languages Thai ## Dataset Structure ### Data Instances ### Data Fields - 'id': sentence id - 'tokens': word tokens by PyThaiNLP's dictionary-based tokenizer 'newmm' - 'pos_tags': POS tags tagged by PyThaiNLP's 'perceptron' engine trained on 'orchid_ud' - 'ner_tags': NER tags tagged by humans ### Data Splits No explicit split is given ## Dataset Creation ### Curation Rationale ThaiNER (v1.3) is a 6,456-sentence named entity recognition dataset created from expanding the 2,258-sentence unnamed dataset by Tirasaroj and Aroonmanakun (2012). It is used to train NER taggers in PyThaiNLP. ### Source Data #### Initial Data Collection and Normalization The earlier part of the dataset is all news articles, whereas the part added by @wannaphong includes news articles, public announcements and @wannaphong's own chat messages with personal and sensitive information removed. #### Who are the source language producers? News articles and public announcements are created by their respective authors. Chat messages are created by @wannaphong. ### Annotations #### Annotation process #### Who are the annotators? Tirasaroj and Aroonmanakun (2012) for the earlier 2,258 sentences and @wannaphong for the rest ### Personal and Sensitive Information News articles and public announcements are not expected to include personal and sensitive information. @wannaphong has removed such information from his own chat messages. ## Considerations for Using the Data ### Social Impact of Dataset - named entity recognition in Thai ### Discussion of Biases Since almost all of collection and annotation is done by @wannaphong, his biases are expected to be reflected in the dataset. ### Other Known Limitations ## Additional Information ### Dataset Curators Tirasaroj and Aroonmanakun (2012) for the earlier 2,258 sentences and @wannaphong for the rest ### Licensing Information CC-BY 3.0 Work extended from: Tirasaroj, N. and Aroonmanakun, W. 2012. Thai NER using CRF model based on surface features. In Proceedings of SNLP-AOS 2011, 9-10 February, 2012, Bangkok, pages 176-180. ### Contributions Thanks to @cstorm125 for adding this dataset.
[ "# Dataset Card for 'thainer'", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper:\n- Leaderboard:\n- Point of Contact: URL", "### Dataset Summary\n\nThaiNER (v1.3) is a 6,456-sentence named entity recognition dataset created from expanding the 2,258-sentence unnamed dataset by Tirasaroj and Aroonmanakun (2012). It is used to train NER taggers in PyThaiNLP. The NER tags are annotated by Tirasaroj and Aroonmanakun (2012)) for 2,258 sentences and the rest by @wannaphong. The POS tags are done by PyThaiNLP's 'perceptron' engine trained on 'orchid_ud'. @wannaphong is now the only maintainer of this dataset.", "### Supported Tasks and Leaderboards\n\n- named entity recognition\n- pos tagging", "### Languages\n\nThai", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- 'id': sentence id\n- 'tokens': word tokens by PyThaiNLP's dictionary-based tokenizer 'newmm'\n- 'pos_tags': POS tags tagged by PyThaiNLP's 'perceptron' engine trained on 'orchid_ud'\n- 'ner_tags': NER tags tagged by humans", "### Data Splits\n\nNo explicit split is given", "## Dataset Creation", "### Curation Rationale\n\nThaiNER (v1.3) is a 6,456-sentence named entity recognition dataset created from expanding the 2,258-sentence unnamed dataset by Tirasaroj and Aroonmanakun (2012). It is used to train NER taggers in PyThaiNLP.", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe earlier part of the dataset is all news articles, whereas the part added by @wannaphong includes news articles, public announcements and @wannaphong's own chat messages with personal and sensitive information removed.", "#### Who are the source language producers?\n\nNews articles and public announcements are created by their respective authors. Chat messages are created by @wannaphong.", "### Annotations", "#### Annotation process", "#### Who are the annotators?\n\nTirasaroj and Aroonmanakun (2012) for the earlier 2,258 sentences and @wannaphong for the rest", "### Personal and Sensitive Information\n\nNews articles and public announcements are not expected to include personal and sensitive information. @wannaphong has removed such information from his own chat messages.", "## Considerations for Using the Data", "### Social Impact of Dataset\n\n- named entity recognition in Thai", "### Discussion of Biases\n\nSince almost all of collection and annotation is done by @wannaphong, his biases are expected to be reflected in the dataset.", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nTirasaroj and Aroonmanakun (2012) for the earlier 2,258 sentences and @wannaphong for the rest", "### Licensing Information\n\nCC-BY 3.0\n\n\n\n\n\nWork extended from:\nTirasaroj, N. and Aroonmanakun, W. 2012. Thai NER using CRF model based on surface features. In Proceedings of SNLP-AOS 2011, 9-10 February, 2012, Bangkok, pages 176-180.", "### Contributions\n\nThanks to @cstorm125 for adding this dataset." ]
[ "TAGS\n#task_categories-token-classification #task_ids-named-entity-recognition #task_ids-part-of-speech #annotations_creators-expert-generated #annotations_creators-machine-generated #language_creators-found #language_creators-expert-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-extended|other-tirasaroj-aroonmanakun #language-Thai #license-cc-by-3.0 #region-us \n", "# Dataset Card for 'thainer'", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper:\n- Leaderboard:\n- Point of Contact: URL", "### Dataset Summary\n\nThaiNER (v1.3) is a 6,456-sentence named entity recognition dataset created from expanding the 2,258-sentence unnamed dataset by Tirasaroj and Aroonmanakun (2012). It is used to train NER taggers in PyThaiNLP. The NER tags are annotated by Tirasaroj and Aroonmanakun (2012)) for 2,258 sentences and the rest by @wannaphong. The POS tags are done by PyThaiNLP's 'perceptron' engine trained on 'orchid_ud'. @wannaphong is now the only maintainer of this dataset.", "### Supported Tasks and Leaderboards\n\n- named entity recognition\n- pos tagging", "### Languages\n\nThai", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- 'id': sentence id\n- 'tokens': word tokens by PyThaiNLP's dictionary-based tokenizer 'newmm'\n- 'pos_tags': POS tags tagged by PyThaiNLP's 'perceptron' engine trained on 'orchid_ud'\n- 'ner_tags': NER tags tagged by humans", "### Data Splits\n\nNo explicit split is given", "## Dataset Creation", "### Curation Rationale\n\nThaiNER (v1.3) is a 6,456-sentence named entity recognition dataset created from expanding the 2,258-sentence unnamed dataset by Tirasaroj and Aroonmanakun (2012). It is used to train NER taggers in PyThaiNLP.", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe earlier part of the dataset is all news articles, whereas the part added by @wannaphong includes news articles, public announcements and @wannaphong's own chat messages with personal and sensitive information removed.", "#### Who are the source language producers?\n\nNews articles and public announcements are created by their respective authors. Chat messages are created by @wannaphong.", "### Annotations", "#### Annotation process", "#### Who are the annotators?\n\nTirasaroj and Aroonmanakun (2012) for the earlier 2,258 sentences and @wannaphong for the rest", "### Personal and Sensitive Information\n\nNews articles and public announcements are not expected to include personal and sensitive information. @wannaphong has removed such information from his own chat messages.", "## Considerations for Using the Data", "### Social Impact of Dataset\n\n- named entity recognition in Thai", "### Discussion of Biases\n\nSince almost all of collection and annotation is done by @wannaphong, his biases are expected to be reflected in the dataset.", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nTirasaroj and Aroonmanakun (2012) for the earlier 2,258 sentences and @wannaphong for the rest", "### Licensing Information\n\nCC-BY 3.0\n\n\n\n\n\nWork extended from:\nTirasaroj, N. and Aroonmanakun, W. 2012. Thai NER using CRF model based on surface features. In Proceedings of SNLP-AOS 2011, 9-10 February, 2012, Bangkok, pages 176-180.", "### Contributions\n\nThanks to @cstorm125 for adding this dataset." ]
[ 145, 9, 120, 27, 146, 20, 5, 6, 6, 88, 10, 5, 69, 4, 57, 35, 5, 5, 33, 39, 8, 15, 39, 7, 5, 30, 64, 17 ]
[ "passage: TAGS\n#task_categories-token-classification #task_ids-named-entity-recognition #task_ids-part-of-speech #annotations_creators-expert-generated #annotations_creators-machine-generated #language_creators-found #language_creators-expert-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-extended|other-tirasaroj-aroonmanakun #language-Thai #license-cc-by-3.0 #region-us \n# Dataset Card for 'thainer'## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper:\n- Leaderboard:\n- Point of Contact: URL### Dataset Summary\n\nThaiNER (v1.3) is a 6,456-sentence named entity recognition dataset created from expanding the 2,258-sentence unnamed dataset by Tirasaroj and Aroonmanakun (2012). It is used to train NER taggers in PyThaiNLP. The NER tags are annotated by Tirasaroj and Aroonmanakun (2012)) for 2,258 sentences and the rest by @wannaphong. The POS tags are done by PyThaiNLP's 'perceptron' engine trained on 'orchid_ud'. @wannaphong is now the only maintainer of this dataset.### Supported Tasks and Leaderboards\n\n- named entity recognition\n- pos tagging### Languages\n\nThai## Dataset Structure### Data Instances" ]
bd4954649e6fc5691427fd45fc6fff00547d7383
# Dataset Card for `thaiqa-squad` ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** http://github.com/pythainlp/thaiqa_squad (original `thaiqa` at https://aiforthai.in.th/) - **Repository:** http://github.com/pythainlp/thaiqa_squad - **Paper:** - **Leaderboard:** - **Point of Contact:**http://github.com/pythainlp/ (original `thaiqa` at https://aiforthai.in.th/) ### Dataset Summary `thaiqa_squad` is an open-domain, extractive question answering dataset (4,000 questions in `train` and 74 questions in `dev`) in [SQuAD](https://rajpurkar.github.io/SQuAD-explorer/) format, originally created by [NECTEC](https://www.nectec.or.th/en/) from Wikipedia articles and adapted to [SQuAD](https://rajpurkar.github.io/SQuAD-explorer/) format by [PyThaiNLP](https://github.com/PyThaiNLP/). ### Supported Tasks and Leaderboards extractive question answering ### Languages Thai ## Dataset Structure ### Data Instances ``` {'answers': {'answer': ['ฮิกกิ้นส์'], 'answer_begin_position': [528], 'answer_end_position': [537]}, 'article_id': 115035, 'context': '<doc id="115035" url="https://th.wikipedia.org/wiki?curid=115035" title="เบนจี้">เบนจี้ เบนจี้ () เป็นชื่อตัวละครหมาพันทางแสนรู้ ที่ปรากฏอยู่ในภาพยนตร์หลายเรื่องที่เขียนบท และกำกับโดย โจ แคมป์ ในช่วงทศวรรษ 1970 ถึง 1980 ภาพยนตร์เรื่องแรกในชุด ใช้ชื่อเรื่องว่า เบนจี้ เช่นเดียวกับตัวละคร ถ่ายทำที่เมืองดัลลัส รัฐเทกซัส ฉายครั้งแรกในปี พ.ศ. 2517 ภาพยนตร์ได้รับการเสนอชื่อเข้าชิงรางวัลออสการ์ และได้รางวัลลูกโลกทองคำ สาขาเพลงประกอบยอดเยี่ยม จากเพลง Benji\'s Theme (I Feel Love) ร้องโดย ชาร์ลี ริช หมาที่แสดงเป็นเบนจี้ตัวแรก ชื่อว่า ฮิกกิ้นส์ (พ.ศ. 2502 - พ.ศ. 2518) มีอายุถึง 15 ปีแล้วในขณะแสดง หลังจากภาพยนตร์ออกฉายได้ไม่นาน มันก็ตายในปี พ.ศ. 2518เบนจี้ในภาพยนตร์เบนจี้ในภาพยนตร์. - พ.ศ. 2517, Benji (ภาพยนตร์) - พ.ศ. 2520, For the Love of Benji (ภาพยนตร์) - พ.ศ. 2521, Benji\'s Very Own Christmas Story (ภาพยนตร์โทรทัศน์) - พ.ศ. 2523, Oh Heavenly Dog (ภาพยนตร์) - พ.ศ. 2523, Benji at Work (ภาพยนตร์โทรทัศน์) - พ.ศ. 2524, Benji Takes a Dive at Marineland (ภาพยนตร์โทรทัศน์) - พ.ศ. 2526, Benji, Zax & the Alien Prince (ภาพยนตร์ซีรีส์) - พ.ศ. 2530, Benji the Hunted (ภาพยนตร์) - พ.ศ. 2547, Benji: Off the Leash! (ภาพยนตร์) - พ.ศ. 2550, Benji: The Barkening (ภาพยนตร์)</doc>\n', 'question': 'สุนัขตัวแรกรับบทเป็นเบนจี้ในภาพยนตร์เรื่อง Benji ที่ออกฉายในปี พ.ศ. 2517 มีชื่อว่าอะไร', 'question_id': 1} {'answers': {'answer': ['ชาร์ลี ริช'], 'answer_begin_position': [482], 'answer_end_position': [492]}, 'article_id': 115035, 'context': '<doc id="115035" url="https://th.wikipedia.org/wiki?curid=115035" title="เบนจี้">เบนจี้ เบนจี้ () เป็นชื่อตัวละครหมาพันทางแสนรู้ ที่ปรากฏอยู่ในภาพยนตร์หลายเรื่องที่เขียนบท และกำกับโดย โจ แคมป์ ในช่วงทศวรรษ 1970 ถึง 1980 ภาพยนตร์เรื่องแรกในชุด ใช้ชื่อเรื่องว่า เบนจี้ เช่นเดียวกับตัวละคร ถ่ายทำที่เมืองดัลลัส รัฐเทกซัส ฉายครั้งแรกในปี พ.ศ. 2517 ภาพยนตร์ได้รับการเสนอชื่อเข้าชิงรางวัลออสการ์ และได้รางวัลลูกโลกทองคำ สาขาเพลงประกอบยอดเยี่ยม จากเพลง Benji\'s Theme (I Feel Love) ร้องโดย ชาร์ลี ริช หมาที่แสดงเป็นเบนจี้ตัวแรก ชื่อว่า ฮิกกิ้นส์ (พ.ศ. 2502 - พ.ศ. 2518) มีอายุถึง 15 ปีแล้วในขณะแสดง หลังจากภาพยนตร์ออกฉายได้ไม่นาน มันก็ตายในปี พ.ศ. 2518เบนจี้ในภาพยนตร์เบนจี้ในภาพยนตร์. - พ.ศ. 2517, Benji (ภาพยนตร์) - พ.ศ. 2520, For the Love of Benji (ภาพยนตร์) - พ.ศ. 2521, Benji\'s Very Own Christmas Story (ภาพยนตร์โทรทัศน์) - พ.ศ. 2523, Oh Heavenly Dog (ภาพยนตร์) - พ.ศ. 2523, Benji at Work (ภาพยนตร์โทรทัศน์) - พ.ศ. 2524, Benji Takes a Dive at Marineland (ภาพยนตร์โทรทัศน์) - พ.ศ. 2526, Benji, Zax & the Alien Prince (ภาพยนตร์ซีรีส์) - พ.ศ. 2530, Benji the Hunted (ภาพยนตร์) - พ.ศ. 2547, Benji: Off the Leash! (ภาพยนตร์) - พ.ศ. 2550, Benji: The Barkening (ภาพยนตร์)</doc>\n', 'question': "เพลง Benji's Theme ใช้ประกอบภาพยนตร์เรื่อง Benji ในปีพ.ศ. 2517 ขับร้องโดยใคร", 'question_id': 2035} ``` ### Data Fields ``` { "question_id": question id "article_id": article id "context": article texts "question": question "answers": { "answer": answer text "answer_begin_position": answer beginning position "answer_end_position": answer exclusive upper bound position } ), } ``` ### Data Splits | | train | valid | |-------------------------|-------------|-------------| | # questions | 4000 | 74 | | # avg words in context | 1186.740750 | 1016.459459 | | # avg words in question | 14.325500 | 12.743243 | | # avg words in answer | 3.279750 | 4.608108 | ## Dataset Creation ### Curation Rationale [PyThaiNLP](https://github.com/PyThaiNLP/) created `thaiqa_squad` as a [SQuAD](https://rajpurkar.github.io/SQuAD-explorer/) version of [thaiqa](http://copycatch.in.th/thai-qa-task.html). [thaiqa](https://aiforthai.in.th/corpus.php) is part of [The 2nd Question answering program from Thai Wikipedia](http://copycatch.in.th/thai-qa-task.html) of [National Software Contest 2020](http://nsc.siit.tu.ac.th/GENA2/login.php). ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? Wikipedia authors for contexts and [NECTEC](https://www.nectec.or.th/en/) for questions and answer annotations ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [NECTEC](https://www.nectec.or.th/en/) ### Personal and Sensitive Information All contents are from Wikipedia. No personal and sensitive information is expected to be included. ## Considerations for Using the Data ### Social Impact of Dataset - open-domain, extractive question answering in Thai ### Discussion of Biases [More Information Needed] ### Other Known Limitations Dataset provided for research purposes only. Please check dataset license for additional information. The contexts include `<doc>` tags at start and at the end ## Additional Information ### Dataset Curators [NECTEC](https://www.nectec.or.th/en/) for original [thaiqa](https://aiforthai.in.th/corpus.php). SQuAD formattting by [PyThaiNLP](https://github.com/PyThaiNLP/). ### Licensing Information CC-BY-NC-SA 3.0 ### Citation Information No clear citation guidelines from source: https://aiforthai.in.th/corpus.php SQuAD version: https://github.com/PyThaiNLP/thaiqa_squad ### Contributions Thanks to [@cstorm125](https://github.com/cstorm125) for adding this dataset.
thaiqa_squad
[ "task_categories:question-answering", "task_ids:extractive-qa", "task_ids:open-domain-qa", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:extended|other-thaiqa", "language:th", "license:cc-by-nc-sa-3.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found"], "language": ["th"], "license": ["cc-by-nc-sa-3.0"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["extended|other-thaiqa"], "task_categories": ["question-answering"], "task_ids": ["extractive-qa", "open-domain-qa"], "pretty_name": "thaiqa-squad", "dataset_info": {"features": [{"name": "question_id", "dtype": "int32"}, {"name": "article_id", "dtype": "int32"}, {"name": "context", "dtype": "string"}, {"name": "question", "dtype": "string"}, {"name": "answers", "sequence": [{"name": "answer", "dtype": "string"}, {"name": "answer_begin_position", "dtype": "int32"}, {"name": "answer_end_position", "dtype": "int32"}]}], "config_name": "thaiqa_squad", "splits": [{"name": "train", "num_bytes": 47905050, "num_examples": 4000}, {"name": "validation", "num_bytes": 744813, "num_examples": 74}], "download_size": 10003354, "dataset_size": 48649863}}
2024-01-18T11:17:06+00:00
[]
[ "th" ]
TAGS #task_categories-question-answering #task_ids-extractive-qa #task_ids-open-domain-qa #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-extended|other-thaiqa #language-Thai #license-cc-by-nc-sa-3.0 #region-us
Dataset Card for 'thaiqa-squad' =============================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL (original 'thaiqa' at URL * Repository: URL * Paper: * Leaderboard: * Point of Contact:URL (original 'thaiqa' at URL ### Dataset Summary 'thaiqa\_squad' is an open-domain, extractive question answering dataset (4,000 questions in 'train' and 74 questions in 'dev') in SQuAD format, originally created by NECTEC from Wikipedia articles and adapted to SQuAD format by PyThaiNLP. ### Supported Tasks and Leaderboards extractive question answering ### Languages Thai Dataset Structure ----------------- ### Data Instances ### Data Fields ### Data Splits train: # questions, valid: 4000 train: # avg words in context, valid: 1186.740750 train: # avg words in question, valid: 14.325500 train: # avg words in answer, valid: 3.279750 Dataset Creation ---------------- ### Curation Rationale PyThaiNLP created 'thaiqa\_squad' as a SQuAD version of thaiqa. thaiqa is part of The 2nd Question answering program from Thai Wikipedia of National Software Contest 2020. ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? Wikipedia authors for contexts and NECTEC for questions and answer annotations ### Annotations #### Annotation process #### Who are the annotators? NECTEC ### Personal and Sensitive Information All contents are from Wikipedia. No personal and sensitive information is expected to be included. Considerations for Using the Data --------------------------------- ### Social Impact of Dataset * open-domain, extractive question answering in Thai ### Discussion of Biases ### Other Known Limitations Dataset provided for research purposes only. Please check dataset license for additional information. The contexts include '' tags at start and at the end Additional Information ---------------------- ### Dataset Curators NECTEC for original thaiqa. SQuAD formattting by PyThaiNLP. ### Licensing Information CC-BY-NC-SA 3.0 No clear citation guidelines from source: URL SQuAD version: URL ### Contributions Thanks to @cstorm125 for adding this dataset.
[ "### Dataset Summary\n\n\n'thaiqa\\_squad' is an open-domain, extractive question answering dataset (4,000 questions in 'train' and 74 questions in 'dev') in SQuAD format, originally created by NECTEC from Wikipedia articles and adapted to SQuAD format by PyThaiNLP.", "### Supported Tasks and Leaderboards\n\n\nextractive question answering", "### Languages\n\n\nThai\n\n\nDataset Structure\n-----------------", "### Data Instances", "### Data Fields", "### Data Splits\n\n\ntrain: # questions, valid: 4000\ntrain: # avg words in context, valid: 1186.740750\ntrain: # avg words in question, valid: 14.325500\ntrain: # avg words in answer, valid: 3.279750\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nPyThaiNLP created 'thaiqa\\_squad' as a SQuAD version of thaiqa. thaiqa is part of The 2nd Question answering program from Thai Wikipedia of National Software Contest 2020.", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?\n\n\nWikipedia authors for contexts and NECTEC for questions and answer annotations", "### Annotations", "#### Annotation process", "#### Who are the annotators?\n\n\nNECTEC", "### Personal and Sensitive Information\n\n\nAll contents are from Wikipedia. No personal and sensitive information is expected to be included.\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset\n\n\n* open-domain, extractive question answering in Thai", "### Discussion of Biases", "### Other Known Limitations\n\n\nDataset provided for research purposes only. Please check dataset license for additional information.\n\n\nThe contexts include '' tags at start and at the end\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nNECTEC for original thaiqa. SQuAD formattting by PyThaiNLP.", "### Licensing Information\n\n\nCC-BY-NC-SA 3.0\n\n\nNo clear citation guidelines from source: URL\n\n\nSQuAD version: URL", "### Contributions\n\n\nThanks to @cstorm125 for adding this dataset." ]
[ "TAGS\n#task_categories-question-answering #task_ids-extractive-qa #task_ids-open-domain-qa #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-extended|other-thaiqa #language-Thai #license-cc-by-nc-sa-3.0 #region-us \n", "### Dataset Summary\n\n\n'thaiqa\\_squad' is an open-domain, extractive question answering dataset (4,000 questions in 'train' and 74 questions in 'dev') in SQuAD format, originally created by NECTEC from Wikipedia articles and adapted to SQuAD format by PyThaiNLP.", "### Supported Tasks and Leaderboards\n\n\nextractive question answering", "### Languages\n\n\nThai\n\n\nDataset Structure\n-----------------", "### Data Instances", "### Data Fields", "### Data Splits\n\n\ntrain: # questions, valid: 4000\ntrain: # avg words in context, valid: 1186.740750\ntrain: # avg words in question, valid: 14.325500\ntrain: # avg words in answer, valid: 3.279750\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nPyThaiNLP created 'thaiqa\\_squad' as a SQuAD version of thaiqa. thaiqa is part of The 2nd Question answering program from Thai Wikipedia of National Software Contest 2020.", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?\n\n\nWikipedia authors for contexts and NECTEC for questions and answer annotations", "### Annotations", "#### Annotation process", "#### Who are the annotators?\n\n\nNECTEC", "### Personal and Sensitive Information\n\n\nAll contents are from Wikipedia. No personal and sensitive information is expected to be included.\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset\n\n\n* open-domain, extractive question answering in Thai", "### Discussion of Biases", "### Other Known Limitations\n\n\nDataset provided for research purposes only. Please check dataset license for additional information.\n\n\nThe contexts include '' tags at start and at the end\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nNECTEC for original thaiqa. SQuAD formattting by PyThaiNLP.", "### Licensing Information\n\n\nCC-BY-NC-SA 3.0\n\n\nNo clear citation guidelines from source: URL\n\n\nSQuAD version: URL", "### Contributions\n\n\nThanks to @cstorm125 for adding this dataset." ]
[ 115, 75, 15, 12, 6, 5, 63, 51, 4, 10, 27, 5, 5, 12, 36, 20, 8, 44, 26, 29, 17 ]
[ "passage: TAGS\n#task_categories-question-answering #task_ids-extractive-qa #task_ids-open-domain-qa #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-extended|other-thaiqa #language-Thai #license-cc-by-nc-sa-3.0 #region-us \n### Dataset Summary\n\n\n'thaiqa\\_squad' is an open-domain, extractive question answering dataset (4,000 questions in 'train' and 74 questions in 'dev') in SQuAD format, originally created by NECTEC from Wikipedia articles and adapted to SQuAD format by PyThaiNLP.### Supported Tasks and Leaderboards\n\n\nextractive question answering### Languages\n\n\nThai\n\n\nDataset Structure\n-----------------### Data Instances### Data Fields### Data Splits\n\n\ntrain: # questions, valid: 4000\ntrain: # avg words in context, valid: 1186.740750\ntrain: # avg words in question, valid: 14.325500\ntrain: # avg words in answer, valid: 3.279750\n\n\nDataset Creation\n----------------### Curation Rationale\n\n\nPyThaiNLP created 'thaiqa\\_squad' as a SQuAD version of thaiqa. thaiqa is part of The 2nd Question answering program from Thai Wikipedia of National Software Contest 2020.### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?\n\n\nWikipedia authors for contexts and NECTEC for questions and answer annotations### Annotations#### Annotation process#### Who are the annotators?\n\n\nNECTEC### Personal and Sensitive Information\n\n\nAll contents are from Wikipedia. No personal and sensitive information is expected to be included.\n\n\nConsiderations for Using the Data\n---------------------------------### Social Impact of Dataset\n\n\n* open-domain, extractive question answering in Thai### Discussion of Biases" ]
616357f471a24974284b7066056d5b17ac9bc7d5
# Dataset Card for ThaiSum ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://github.com/nakhunchumpolsathien/ThaiSum - **Repository:** https://github.com/nakhunchumpolsathien/ThaiSum - **Paper:** - **Leaderboard:** - **Point of Contact:** https://github.com/nakhunchumpolsathien ### Dataset Summary ThaiSum is a large-scale corpus for Thai text summarization obtained from several online news websites namely Thairath, ThaiPBS, Prachathai, and The Standard. This dataset consists of over 350,000 article and summary pairs written by journalists. ### Supported Tasks and Leaderboards summarization, language modeling ### Languages Thai ## Dataset Structure ### Data Instances ``` {'body': 'กีเก ซานเชซ ฟลอเรส\xa0 กุนซือเลือดกระทิงของทีมวัตฟอร์ด\xa0 เมินประเด็นจุดโทษปัญหาในเกมพรีเมียร์ลีก อังกฤษ นัดที่แตนอาละวาดเปิดบ้านพ่าย คริสตัล พาเลซ 0-1ชี้ทีมของเขาเล่นไม่ดีพอเอง,สำนักข่าวต่างประเทศรายงานวันที่ 27 ก.ย. ว่า กีเก ซานเชซ ฟลอเรส\xa0 ผู้จัดการทีมชาวสเปน ของ แตนอาละวาด วัตฟอร์ด\xa0 ยอมรับทีมของเขาเล่นได้ไม่ดีพอเอง ในเกมพรีเมียร์ลีก อังกฤษ นัดเปิดบ้านพ่าย อินทรีผงาด คริสตัล พาเลซ 0-1 เมื่อคืนวันอาทิตย์ที่ผ่านมา,เกมนี้จุดเปลี่ยนมาอยู่ที่การได้จุดโทษในช่วงครึ่งหลังของ คริสตัล พาเลซ ซึ่งไม่ค่อยชัดเจนเท่าไหร่ว่า อัลลัน นียอม นั้นไปทำฟาล์วใส่ วิลฟรีด ซาฮา ในเขตโทษหรือไม่ แต่ผู้ตัดสินก็ชี้เป็นจุดโทษ ซึ่ง โยอัน กาบาย สังหารไม่พลาด และเป็นประตูชัยช่วยให้ คริสตัล พาเลซ เอาชนะ วัตฟอร์ด ไป 1-0 และเป็นการพ่ายแพ้ในบ้านนัดแรกของวัตฟอร์ดในฤดูกาลนี้อีกด้วย,ฟลอเรส กล่าวว่า มันเป็นเรื่องยากในการหยุดเกมรุกของคริสตัล พาเลซ ซึ่งมันอึดอัดจริงๆสำหรับเรา เราเล่นกันได้ไม่ดีนักในตอนที่ได้ครองบอล เราต้องเล่นทางริมเส้นให้มากกว่านี้ เราไม่สามารถหยุดเกมสวนกลับของพวกเขาได้ และแนวรับของเราก็ยืนไม่เป็นระเบียบสักเท่าไหร่ในช่วงครึ่งแรก ส่วนเรื่องจุดโทษการตัดสินใจขั้นสุดท้ายมันอยู่ที่ผู้ตัดสิน ซึ่งมันเป็นการตัดสินใจที่สำคัญ ผมเองก็ไม่รู้ว่าเขาตัดสินถูกหรือเปล่า บางทีมันอาจเป็นจุดที่ตัดสินเกมนี้เลย แต่เราไม่ได้แพ้เกมนี้เพราะจุดโทษ เราแพ้ในวันนี้เพราะเราเล่นไม่ดีและคริสตัล พาเลซ เล่นดีกว่าเรา เราไม่ได้มีฟอร์มการเล่นที่ดีในเกมนี้เลย', 'summary': 'กีเก ซานเชซ ฟลอเรส กุนซือเลือดกระทิงของทีมวัตฟอร์ด เมินประเด็นจุดโทษปัญหาในเกมพรีเมียร์ลีก อังกฤษ นัดที่แตนอาละวาดเปิดบ้านพ่าย คริสตัล พาเลซ 0-1ชี้ทีมของเขาเล่นไม่ดีพอเอง', 'tags': 'พรีเมียร์ลีก,วัตฟอร์ด,คริสตัล พาเลซ,กีเก ซานเชซ ฟลอเรส,ข่าวกีฬา,ข่าว,ไทยรัฐออนไลน์', 'title': 'ฟลอเรส รับ วัตฟอร์ดห่วยเองเกมพ่ายพาเลซคาบ้าน', 'type': '', 'url': 'https://www.thairath.co.th/content/528322'} ``` ### Data Fields - `title`: title of article - `body`: body of article - `summary`: summary of article - `type`: type of article, if any - `tags`: tags of article, separated by `,` - `url`: URL of article ### Data Splits train/valid/test: 358868 / 11000 / 11000 ## Dataset Creation ### Curation Rationale Sequence-to-sequence (Seq2Seq) models have shown great achievement in text summarization. However, Seq2Seq model often requires large-scale training data to achieve effective results. Although many impressive advancements in text summarization field have been made, most of summarization studies focus on resource-rich languages. The progress of Thai text summarization is still far behind. The dearth of large-scale dataset keeps Thai text summarization in its infancy. As far as our knowledge goes, there is not a large-scale dataset for Thai text summarization available anywhere. Thus, we present ThaiSum, a large-scale corpus for Thai text summarization obtained from several online news websites namely Thairath, ThaiPBS, Prachathai, and The Standard. ### Source Data #### Initial Data Collection and Normalization We used a python library named Scrapy to crawl articles from several news websites namely Thairath, Prachatai, ThaiPBS and, The Standard. We first collected news URLs provided in their sitemaps. During web-crawling, we used HTML markup and metadata available in HTML pages to identify article text, summary, headline, tags and label. Collected articles were published online from 2014 to August 2020. <br> <br> We further performed data cleansing process to minimize noisy data. We filtered out articles that their article text or summary is missing. Articles that contains article text with less than 150 words or summary with less than 15 words were removed. We also discarded articles that contain at least one of these following tags: ‘ดวง’ (horoscope), ‘นิยาย’ (novel), ‘อินสตราแกรมดารา’ (celebrity Instagram), ‘คลิปสุดฮา’(funny video) and ‘สรุปข่าว’ (highlight news). Some summaries were completely irrelevant to their original article texts. To eliminate those irrelevant summaries, we calculated abstractedness score between summary and its article text. Abstractedness score is written formally as: <br> <center><a href="https://www.codecogs.com/eqnedit.php?latex=\begin{equation}&space;\frac{|S-A|}{r}&space;\times&space;100&space;\end{equation}" target="_blank"><img src="https://latex.codecogs.com/gif.latex?\begin{equation}&space;\frac{|S-A|}{r}&space;\times&space;100&space;\end{equation}" title="\begin{equation} \frac{|S-A|}{r} \times 100 \end{equation}" /></a></center><br> <br>Where 𝑆 denotes set of article tokens. 𝐴 denotes set of summary tokens. 𝑟 denotes a total number of summary tokens. We omitted articles that have abstractedness score at 1-grams higher than 60%. <br><br> It is important to point out that we used [PyThaiNLP](https://github.com/PyThaiNLP/pythainlp), version 2.2.4, tokenizing engine = newmm, to process Thai texts in this study. It is challenging to tokenize running Thai text into words or sentences because there are not clear word/sentence delimiters in Thai language. Therefore, using different tokenization engines may result in different segment of words/sentences. After data-cleansing process, ThaiSum dataset contains over 358,000 articles. The size of this dataset is comparable to a well-known English document summarization dataset, CNN/Dily mail dataset. Moreover, we analyse the characteristics of this dataset by measuring the abstractedness level, compassion rate, and content diversity. For more details, see [thaisum_exploration.ipynb](https://github.com/nakhunchumpolsathien/ThaiSum/blob/master/thaisum_exploration.ipynb). #### Dataset Statistics ThaiSum dataset consists of 358,868 articles. Average lengths of article texts and summaries are approximately 530 and 37 words respectively. As mentioned earlier, we also collected headlines, tags and labels provided in each article. Tags are similar to keywords of the article. An article normally contains several tags but a few labels. Tags can be name of places or persons that article is about while labels indicate news category (politic, entertainment, etc.). Ultimatly, ThaiSum contains 538,059 unique tags and 59 unique labels. Note that not every article contains tags or labels. |Dataset Size| 358,868 | articles | |:---|---:|---:| |Avg. Article Length| 529.5 | words| |Avg. Summary Length | 37.3 | words| |Avg. Headline Length | 12.6 | words| |Unique Vocabulary Size | 407,355 | words| |Occurring > 10 times | 81,761 | words| |Unique News Tag Size | 538,059 | tags| |Unique News Label Size | 59 | labels| #### Who are the source language producers? Journalists of respective articles ### Annotations #### Annotation process `summary`, `type` and `tags` are created by journalists who wrote the articles and/or their publishers. #### Who are the annotators? `summary`, `type` and `tags` are created by journalists who wrote the articles and/or their publishers. ### Personal and Sensitive Information All data are public news articles. No personal and sensitive information is expected to be included. ## Considerations for Using the Data ### Social Impact of Dataset - News summarization in Thai - Language modeling for Thai news ### Discussion of Biases - [ThaiPBS](https://www.thaipbs.or.th/home) [receives funding from Thai government](https://www.bangkokbiznews.com/blog/detail/648740). - [Thairath](https://www.thairath.co.th/) is known as [the most popular newspaper in Thailand](https://mgronline.com/onlinesection/detail/9620000058532); no clear political leaning. - [The Standard](https://thestandard.co/) is a left-leaning online magazine. - [Prachathai](https://prachatai.com/) is a left-leaning, human-right-focused news site. ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [@nakhunchumpolsathien](https://github.com/nakhunchumpolsathien/) [@caramelWaffle](https://github.com/caramelWaffle) ### Licensing Information MIT License ### Citation Information ``` @mastersthesis{chumpolsathien_2020, title={Using Knowledge Distillation from Keyword Extraction to Improve the Informativeness of Neural Cross-lingual Summarization}, author={Chumpolsathien, Nakhun}, year={2020}, school={Beijing Institute of Technology} ``` ### Contributions Thanks to [@cstorm125](https://github.com/cstorm125) for adding this dataset.
thaisum
[ "task_categories:summarization", "task_categories:text-generation", "task_categories:fill-mask", "task_ids:language-modeling", "task_ids:masked-language-modeling", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:th", "license:mit", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["found"], "language": ["th"], "license": ["mit"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["summarization", "text-generation", "fill-mask"], "task_ids": ["language-modeling", "masked-language-modeling"], "pretty_name": "ThaiSum", "dataset_info": {"features": [{"name": "title", "dtype": "string"}, {"name": "body", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "tags", "dtype": "string"}, {"name": "url", "dtype": "string"}], "config_name": "thaisum", "splits": [{"name": "train", "num_bytes": 2945472406, "num_examples": 358868}, {"name": "validation", "num_bytes": 118437310, "num_examples": 11000}, {"name": "test", "num_bytes": 119496704, "num_examples": 11000}], "download_size": 647582078, "dataset_size": 3183406420}}
2024-01-18T11:17:07+00:00
[]
[ "th" ]
TAGS #task_categories-summarization #task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Thai #license-mit #region-us
Dataset Card for ThaiSum ======================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: URL * Paper: * Leaderboard: * Point of Contact: URL ### Dataset Summary ThaiSum is a large-scale corpus for Thai text summarization obtained from several online news websites namely Thairath, ThaiPBS, Prachathai, and The Standard. This dataset consists of over 350,000 article and summary pairs written by journalists. ### Supported Tasks and Leaderboards summarization, language modeling ### Languages Thai Dataset Structure ----------------- ### Data Instances ### Data Fields * 'title': title of article * 'body': body of article * 'summary': summary of article * 'type': type of article, if any * 'tags': tags of article, separated by ',' * 'url': URL of article ### Data Splits train/valid/test: 358868 / 11000 / 11000 Dataset Creation ---------------- ### Curation Rationale Sequence-to-sequence (Seq2Seq) models have shown great achievement in text summarization. However, Seq2Seq model often requires large-scale training data to achieve effective results. Although many impressive advancements in text summarization field have been made, most of summarization studies focus on resource-rich languages. The progress of Thai text summarization is still far behind. The dearth of large-scale dataset keeps Thai text summarization in its infancy. As far as our knowledge goes, there is not a large-scale dataset for Thai text summarization available anywhere. Thus, we present ThaiSum, a large-scale corpus for Thai text summarization obtained from several online news websites namely Thairath, ThaiPBS, Prachathai, and The Standard. ### Source Data #### Initial Data Collection and Normalization We used a python library named Scrapy to crawl articles from several news websites namely Thairath, Prachatai, ThaiPBS and, The Standard. We first collected news URLs provided in their sitemaps. During web-crawling, we used HTML markup and metadata available in HTML pages to identify article text, summary, headline, tags and label. Collected articles were published online from 2014 to August 2020. We further performed data cleansing process to minimize noisy data. We filtered out articles that their article text or summary is missing. Articles that contains article text with less than 150 words or summary with less than 15 words were removed. We also discarded articles that contain at least one of these following tags: ‘ดวง’ (horoscope), ‘นิยาย’ (novel), ‘อินสตราแกรมดารา’ (celebrity Instagram), ‘คลิปสุดฮา’(funny video) and ‘สรุปข่าว’ (highlight news). Some summaries were completely irrelevant to their original article texts. To eliminate those irrelevant summaries, we calculated abstractedness score between summary and its article text. Abstractedness score is written formally as: [![](URL title=)](URL target=) Where 𝑆 denotes set of article tokens. 𝐴 denotes set of summary tokens. 𝑟 denotes a total number of summary tokens. We omitted articles that have abstractedness score at 1-grams higher than 60%. It is important to point out that we used PyThaiNLP, version 2.2.4, tokenizing engine = newmm, to process Thai texts in this study. It is challenging to tokenize running Thai text into words or sentences because there are not clear word/sentence delimiters in Thai language. Therefore, using different tokenization engines may result in different segment of words/sentences. After data-cleansing process, ThaiSum dataset contains over 358,000 articles. The size of this dataset is comparable to a well-known English document summarization dataset, CNN/Dily mail dataset. Moreover, we analyse the characteristics of this dataset by measuring the abstractedness level, compassion rate, and content diversity. For more details, see thaisum\_exploration.ipynb. #### Dataset Statistics ThaiSum dataset consists of 358,868 articles. Average lengths of article texts and summaries are approximately 530 and 37 words respectively. As mentioned earlier, we also collected headlines, tags and labels provided in each article. Tags are similar to keywords of the article. An article normally contains several tags but a few labels. Tags can be name of places or persons that article is about while labels indicate news category (politic, entertainment, etc.). Ultimatly, ThaiSum contains 538,059 unique tags and 59 unique labels. Note that not every article contains tags or labels. #### Who are the source language producers? Journalists of respective articles ### Annotations #### Annotation process 'summary', 'type' and 'tags' are created by journalists who wrote the articles and/or their publishers. #### Who are the annotators? 'summary', 'type' and 'tags' are created by journalists who wrote the articles and/or their publishers. ### Personal and Sensitive Information All data are public news articles. No personal and sensitive information is expected to be included. Considerations for Using the Data --------------------------------- ### Social Impact of Dataset * News summarization in Thai * Language modeling for Thai news ### Discussion of Biases * ThaiPBS receives funding from Thai government. * Thairath is known as the most popular newspaper in Thailand; no clear political leaning. * The Standard is a left-leaning online magazine. * Prachathai is a left-leaning, human-right-focused news site. ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators @nakhunchumpolsathien @caramelWaffle ### Licensing Information MIT License ### Contributions Thanks to @cstorm125 for adding this dataset.
[ "### Dataset Summary\n\n\nThaiSum is a large-scale corpus for Thai text summarization obtained from several online news websites namely Thairath, ThaiPBS, Prachathai, and The Standard. This dataset consists of over 350,000 article and summary pairs written by journalists.", "### Supported Tasks and Leaderboards\n\n\nsummarization, language modeling", "### Languages\n\n\nThai\n\n\nDataset Structure\n-----------------", "### Data Instances", "### Data Fields\n\n\n* 'title': title of article\n* 'body': body of article\n* 'summary': summary of article\n* 'type': type of article, if any\n* 'tags': tags of article, separated by ','\n* 'url': URL of article", "### Data Splits\n\n\ntrain/valid/test: 358868 / 11000 / 11000\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nSequence-to-sequence (Seq2Seq) models have shown great achievement in text summarization. However, Seq2Seq model often requires large-scale training data to achieve effective results. Although many impressive advancements in text summarization field have been made, most of summarization studies focus on resource-rich languages. The progress of Thai text summarization is still far behind. The dearth of large-scale dataset keeps Thai text summarization in its infancy. As far as our knowledge goes, there is not a large-scale dataset for Thai text summarization available anywhere. Thus, we present ThaiSum, a large-scale corpus for Thai text summarization obtained from several online news websites namely Thairath, ThaiPBS, Prachathai, and The Standard.", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nWe used a python library named Scrapy to crawl articles from several news websites namely Thairath, Prachatai, ThaiPBS and, The Standard. We first collected news URLs provided in their sitemaps. During web-crawling, we used HTML markup and metadata available in HTML pages to identify article text, summary, headline, tags and label. Collected articles were published online from 2014 to August 2020. \n \n\nWe further performed data cleansing process to minimize noisy data. We filtered out articles that their article text or summary is missing. Articles that contains article text with less than 150 words or summary with less than 15 words were removed. We also discarded articles that contain at least one of these following tags: ‘ดวง’ (horoscope), ‘นิยาย’ (novel), ‘อินสตราแกรมดารา’ (celebrity Instagram), ‘คลิปสุดฮา’(funny video) and ‘สรุปข่าว’ (highlight news). Some summaries were completely irrelevant to their original article texts. To eliminate those irrelevant summaries, we calculated abstractedness score between summary and its article text. Abstractedness score is written formally as: \n\n\n\n[![](URL title=)](URL target=) \n\n \nWhere 𝑆 denotes set of article tokens. 𝐴 denotes set of summary tokens. 𝑟 denotes a total number of summary tokens. We omitted articles that have abstractedness score at 1-grams higher than 60%.\n \n \n\nIt is important to point out that we used PyThaiNLP, version 2.2.4, tokenizing engine = newmm, to process Thai texts in this study. It is challenging to tokenize running Thai text into words or sentences because there are not clear word/sentence delimiters in Thai language. Therefore, using different tokenization engines may result in different segment of words/sentences.\n\n\nAfter data-cleansing process, ThaiSum dataset contains over 358,000 articles. The size of this dataset is comparable to a well-known English document summarization dataset, CNN/Dily mail dataset. Moreover, we analyse the characteristics of this dataset by measuring the abstractedness level, compassion rate, and content diversity. For more details, see thaisum\\_exploration.ipynb.", "#### Dataset Statistics\n\n\nThaiSum dataset consists of 358,868 articles. Average lengths of article texts and summaries are approximately 530 and 37 words respectively. As mentioned earlier, we also collected headlines, tags and labels provided in each article. Tags are similar to keywords of the article. An article normally contains several tags but a few labels. Tags can be name of places or persons that article is about while labels indicate news category (politic, entertainment, etc.). Ultimatly, ThaiSum contains 538,059 unique tags and 59 unique labels. Note that not every article contains tags or labels.", "#### Who are the source language producers?\n\n\nJournalists of respective articles", "### Annotations", "#### Annotation process\n\n\n'summary', 'type' and 'tags' are created by journalists who wrote the articles and/or their publishers.", "#### Who are the annotators?\n\n\n'summary', 'type' and 'tags' are created by journalists who wrote the articles and/or their publishers.", "### Personal and Sensitive Information\n\n\nAll data are public news articles. No personal and sensitive information is expected to be included.\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset\n\n\n* News summarization in Thai\n* Language modeling for Thai news", "### Discussion of Biases\n\n\n* ThaiPBS receives funding from Thai government.\n* Thairath is known as the most popular newspaper in Thailand; no clear political leaning.\n* The Standard is a left-leaning online magazine.\n* Prachathai is a left-leaning, human-right-focused news site.", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\n@nakhunchumpolsathien\n@caramelWaffle", "### Licensing Information\n\n\nMIT License", "### Contributions\n\n\nThanks to @cstorm125 for adding this dataset." ]
[ "TAGS\n#task_categories-summarization #task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Thai #license-mit #region-us \n", "### Dataset Summary\n\n\nThaiSum is a large-scale corpus for Thai text summarization obtained from several online news websites namely Thairath, ThaiPBS, Prachathai, and The Standard. This dataset consists of over 350,000 article and summary pairs written by journalists.", "### Supported Tasks and Leaderboards\n\n\nsummarization, language modeling", "### Languages\n\n\nThai\n\n\nDataset Structure\n-----------------", "### Data Instances", "### Data Fields\n\n\n* 'title': title of article\n* 'body': body of article\n* 'summary': summary of article\n* 'type': type of article, if any\n* 'tags': tags of article, separated by ','\n* 'url': URL of article", "### Data Splits\n\n\ntrain/valid/test: 358868 / 11000 / 11000\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nSequence-to-sequence (Seq2Seq) models have shown great achievement in text summarization. However, Seq2Seq model often requires large-scale training data to achieve effective results. Although many impressive advancements in text summarization field have been made, most of summarization studies focus on resource-rich languages. The progress of Thai text summarization is still far behind. The dearth of large-scale dataset keeps Thai text summarization in its infancy. As far as our knowledge goes, there is not a large-scale dataset for Thai text summarization available anywhere. Thus, we present ThaiSum, a large-scale corpus for Thai text summarization obtained from several online news websites namely Thairath, ThaiPBS, Prachathai, and The Standard.", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nWe used a python library named Scrapy to crawl articles from several news websites namely Thairath, Prachatai, ThaiPBS and, The Standard. We first collected news URLs provided in their sitemaps. During web-crawling, we used HTML markup and metadata available in HTML pages to identify article text, summary, headline, tags and label. Collected articles were published online from 2014 to August 2020. \n \n\nWe further performed data cleansing process to minimize noisy data. We filtered out articles that their article text or summary is missing. Articles that contains article text with less than 150 words or summary with less than 15 words were removed. We also discarded articles that contain at least one of these following tags: ‘ดวง’ (horoscope), ‘นิยาย’ (novel), ‘อินสตราแกรมดารา’ (celebrity Instagram), ‘คลิปสุดฮา’(funny video) and ‘สรุปข่าว’ (highlight news). Some summaries were completely irrelevant to their original article texts. To eliminate those irrelevant summaries, we calculated abstractedness score between summary and its article text. Abstractedness score is written formally as: \n\n\n\n[![](URL title=)](URL target=) \n\n \nWhere 𝑆 denotes set of article tokens. 𝐴 denotes set of summary tokens. 𝑟 denotes a total number of summary tokens. We omitted articles that have abstractedness score at 1-grams higher than 60%.\n \n \n\nIt is important to point out that we used PyThaiNLP, version 2.2.4, tokenizing engine = newmm, to process Thai texts in this study. It is challenging to tokenize running Thai text into words or sentences because there are not clear word/sentence delimiters in Thai language. Therefore, using different tokenization engines may result in different segment of words/sentences.\n\n\nAfter data-cleansing process, ThaiSum dataset contains over 358,000 articles. The size of this dataset is comparable to a well-known English document summarization dataset, CNN/Dily mail dataset. Moreover, we analyse the characteristics of this dataset by measuring the abstractedness level, compassion rate, and content diversity. For more details, see thaisum\\_exploration.ipynb.", "#### Dataset Statistics\n\n\nThaiSum dataset consists of 358,868 articles. Average lengths of article texts and summaries are approximately 530 and 37 words respectively. As mentioned earlier, we also collected headlines, tags and labels provided in each article. Tags are similar to keywords of the article. An article normally contains several tags but a few labels. Tags can be name of places or persons that article is about while labels indicate news category (politic, entertainment, etc.). Ultimatly, ThaiSum contains 538,059 unique tags and 59 unique labels. Note that not every article contains tags or labels.", "#### Who are the source language producers?\n\n\nJournalists of respective articles", "### Annotations", "#### Annotation process\n\n\n'summary', 'type' and 'tags' are created by journalists who wrote the articles and/or their publishers.", "#### Who are the annotators?\n\n\n'summary', 'type' and 'tags' are created by journalists who wrote the articles and/or their publishers.", "### Personal and Sensitive Information\n\n\nAll data are public news articles. No personal and sensitive information is expected to be included.\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset\n\n\n* News summarization in Thai\n* Language modeling for Thai news", "### Discussion of Biases\n\n\n* ThaiPBS receives funding from Thai government.\n* Thairath is known as the most popular newspaper in Thailand; no clear political leaning.\n* The Standard is a left-leaning online magazine.\n* Prachathai is a left-leaning, human-right-focused news site.", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\n@nakhunchumpolsathien\n@caramelWaffle", "### Licensing Information\n\n\nMIT License", "### Contributions\n\n\nThanks to @cstorm125 for adding this dataset." ]
[ 120, 65, 17, 12, 6, 64, 27, 192, 4, 518, 143, 15, 5, 33, 37, 36, 21, 71, 14, 20, 8, 17 ]
[ "passage: TAGS\n#task_categories-summarization #task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Thai #license-mit #region-us \n### Dataset Summary\n\n\nThaiSum is a large-scale corpus for Thai text summarization obtained from several online news websites namely Thairath, ThaiPBS, Prachathai, and The Standard. This dataset consists of over 350,000 article and summary pairs written by journalists.### Supported Tasks and Leaderboards\n\n\nsummarization, language modeling### Languages\n\n\nThai\n\n\nDataset Structure\n-----------------### Data Instances### Data Fields\n\n\n* 'title': title of article\n* 'body': body of article\n* 'summary': summary of article\n* 'type': type of article, if any\n* 'tags': tags of article, separated by ','\n* 'url': URL of article### Data Splits\n\n\ntrain/valid/test: 358868 / 11000 / 11000\n\n\nDataset Creation\n----------------### Curation Rationale\n\n\nSequence-to-sequence (Seq2Seq) models have shown great achievement in text summarization. However, Seq2Seq model often requires large-scale training data to achieve effective results. Although many impressive advancements in text summarization field have been made, most of summarization studies focus on resource-rich languages. The progress of Thai text summarization is still far behind. The dearth of large-scale dataset keeps Thai text summarization in its infancy. As far as our knowledge goes, there is not a large-scale dataset for Thai text summarization available anywhere. Thus, we present ThaiSum, a large-scale corpus for Thai text summarization obtained from several online news websites namely Thairath, ThaiPBS, Prachathai, and The Standard.### Source Data" ]
148e1d5e8349977c76f673190424a2faf6980a1d
# Dataset Card for The Pile ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) This model card is a work in progress. Please also see [our datasheet](https://arxiv.org/abs/2201.07311) for more detailed info. ## Dataset Description - **Homepage:** https://pile.eleuther.ai/ - **Repository:** https://github.com/EleutherAI/the-pile - **Paper:** [The Pile: An 800GB Dataset of Diverse Text for Language Modeling](https://arxiv.org/abs/2101.00027) - **Leaderboard:** - **Point of Contact:** [EleutherAI](mailto:contact@eleuther.ai) - **Datasheet:** [Datasheet for the Pile](https://arxiv.org/abs/2201.07311) ### Dataset Summary The Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality datasets combined together. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages This dataset is in English (`EN`). ## Dataset Structure ### Data Instances #### all ``` { 'meta': {'pile_set_name': 'Pile-CC'}, 'text': 'It is done, and submitted. You can play “Survival of the Tastiest” on Android, and on the web. Playing on...' } ``` <details> <summary>Expand to see individual components</summary> #### enron_emails ``` { 'text': 'Name\t\t\tNew Title\t\t\t\tEffective Date\t\t\tMid Year promotion Yes/No\n\nFloyd, Jodie\t\tSr Cust Svc Rep (no change)\t\t7/16/01\t\t\t\tNo\n\nBuehler, Craig\t\tSr Mkt/Sup Analyst (no change)\t\t7/16/01\t\t\t\tNo\n\nWagoner, Mike\t\tTeam Advisor - Gas Control\t\t7/1/01\t\t\t\tNo\n\nClapper, Karen\t\tSr Cust Svc Rep\t\t\t8/1/01\t\t\t\tYes\n\nGreaney, Chris\t\tSr Cust Svc Rep\t\t\t8/1/01\t\t\t\tYes\n\nWilkens, Jerry\t\tSr Cust Svc Rep\t\t\t8/1/01\t\t\t\tYes\n\nMinton, Kevin\t\tPipeline Controller\t\t\t8/1/01\t\t\t\tYes\n\nCox, Don\t\tPipeline Controller\t\t\t8/1/01\t\t\t\tYes\n\nHanagriff, Richard\tSr Accounting Control Spec\t\t8/1/01\t\t\t\tYes\n\n\nThanks,\nMS' 'meta': "{}", } ``` #### europarl ``` { 'text': 'Uvádění biocidních přípravků na trh - Nový návrh revize týkající se biocidních přípravků (rozprava) \nPředsedající\nDalším bodem je společná rozprava o následujících tématech:\nzpráva paní Sârbuové za Výbor pro životní prostředí, veřejné zdraví a bezpečnost potravin o návrhu...' 'meta': "{'language': 'cs'}", } ``` #### free_law ``` { 'meta': "{'case_jurisdiction': 'scotus.tar.gz', 'case_ID': '110921.json','date_created': '2010-04-28T17:12:49Z'}", 'text': '\n461 U.S. 238 (1983)\nOLIM ET AL.\nv.\nWAKINEKONA\nNo. 81-1581.\nSupreme Court of United States.\nArgued...' } ``` #### hacker_news ``` { 'text': "\nChina Deserves Donald Trump - rm2889\nhttps://www.nytimes.com/2019/05/21/opinion/china-trump-trade.html\n======\nNotPaidToPost\n> so he’d be wise to curb his nationalistic “no-one-tells-China-what-to-do”\n> bluster\n\nThis comment highlights both ignorance of Chinese history and continuing\nAmerican arrogance.\n\nChina has been painfully dictated what to do during the last 200 years. This\nhas had a profound effect on the country and has led to the collapse of\nimperial rule and the drive to 'rejuvenate'...", 'meta': "{'id': '19979654'}", } ``` #### nih_exporter ``` { 'text': "The National Domestic Violence Hotline (NDVH) and the National Dating Abuse Helpline (NDAH), which are supported by the Division of Family Violence Prevention and Services within the Family and Youth Services Bureau, serve as critical partners in the intervention, prevention, and resource assistance efforts of the network of family violence, domestic violence, and dating violence service providers. They provide crisis intervention and support services; information about resources on domestic...", 'meta': " {'APPLICATION_ID': 100065}", } ``` #### pubmed ``` { 'meta': {'pmid': 11409574, 'language': 'eng'}, 'text': 'Epidemiology of hypoxaemia in children with acute lower respiratory infection.\nTo determine the prevalence of hypoxaemia in children aged under 5 years suffering acute lower respiratory infections (ALRI), the risk factors for hypoxaemia in children under 5 years of age with ALRI, and the association of hypoxaemia with an increased risk of dying in children of the same age. Systematic review of the published literature. Out-patient clinics, emergency departments and hospitalisation wards in 23 health centres from 10 countries. Cohort studies reporting the frequency of hypoxaemia in children under 5 years of age with ALRI, and the association between hypoxaemia and the risk of dying. Prevalence of hypoxaemia measured in children with ARI and relative risks for the association between the severity of illness and the frequency of hypoxaemia, and between hypoxaemia and the risk of dying. Seventeen published studies were found that included 4,021 children under 5 with acute respiratory infections (ARI) and reported the prevalence of hypoxaemia. Out-patient children and those with a clinical diagnosis of upper ARI had a low risk of hypoxaemia (pooled estimate of 6% to 9%). The prevalence increased to 31% and to 43% in patients in emergency departments and in cases with clinical pneumonia, respectively, and it was even higher among hospitalised children (47%) and in those with radiographically confirmed pneumonia (72%). The cumulated data also suggest that hypoxaemia is more frequent in children living at high altitude. Three papers reported an association between hypoxaemia and death, with relative risks varying between 1.4 and 4.6. Papers describing predictors of hypoxaemia have focused on clinical signs for detecting hypoxaemia rather than on identifying risk factors for developing this complication. Hypoxaemia is a common and potentially lethal complication of ALRI in children under 5, particularly among those with severe disease and those living at high altitude. Given the observed high prevalence of hypoxaemia and its likely association with increased mortality, efforts should be made to improve the detection of hypoxaemia and to provide oxygen earlier to more children with severe ALRI.' } ``` #### pubmed_central ``` { 'meta': "{id': 'PMC5595690'}", 'text': 'Introduction {#acel12642-sec-0001}\n============\n\nAlzheimer\\\'s disease (AD), the most common cause of...' } ``` #### ubuntu_irc ``` { 'text': "#ubuntu 2004-07-05\n* Window 3\n* \tServer: [0] <None>\n* \tScreen: 0x817e90c\n* \tGeometry Info: [0 11 0 11 11 11] \n* \tCO, LI are [94 49] \n* \tCurrent channel: #ubuntu\n* \tQuery User: <None> \n*\tPrompt: <None>\n* \tSecond status line is OFF\n* \tSplit line is ON triple is OFF\n* \tLogging is ON\n* \tLogfile is irclogs/ubuntu.log\n* \tNotification is OFF\n* \tHold mode is OFF\n* \tWindow level is NONE\n* \tLastlog level is ALL\n* \tNotify level is ALL\n<mdz> lifeless: using tla effectively for all packages in Warty requ...", 'meta': "{'channel': 'ubuntu', 'month': 7}" } ``` #### uspto ``` { 'text': "1. Field of the Invention\nIn an extensive plant breeding program, Grant Merrill, originator and now deceased, originated a large number of new and distinct varieties of fruit trees, and which included the herein-claimed variety of peach tree. Such plant breeding program was undertaken in originator's experimental orchard located near Exeter, Tulare County, Calif.\n2. Prior Varieties\nAmong the existent varieties of peach trees which were known to originator, particular reference is made to Gemfree (U.S. Plant Pat. No. 1,409) and June Lady (U.S. Plant Pat. No. 3,022) hereinafter mentioned for the purpose of comparison.", 'meta': "{'bibliographic_information': {'Patent Number': 'PP0049700', 'Series Code': '6', 'Application Number': '2845415', 'Application Type': '6', 'Art unit': '337', 'Application Filing Date': '19810720', 'Title of Invention': 'Peach tree (A3-10)', 'Issue Date': '19830104', 'Number of Claims': '1', 'Exemplary Claim Number(s)': '1', 'Primary Examiner': 'Bagwill; Robert E.', 'Number of Drawing Sheets': '1', 'Number of figures': '1'}, 'source_file': 'https://bulkdata.uspto.gov/data/patent/grant/redbook/fulltext/1983/pftaps19830104_wk01.zip', 'abstract': 'A peach tree which is large, vigorous, and spreading; foliated with large, lanceolate leaves having a finely serrate margin, a petiole of medium length and thickness, and medium size, reniform glands; blooms from medium size, conic, plump, pubescent buds; the flowers, medium in blooming period compared with other varieties, being of medium size, and pink; and is a regular and very productive bearer of medium but variable size, round truncate, clingstone fruit having yellow skin substantially overspread with red, yellow flesh mottled with red adjacent the skin, and an amber stone.', 'classifications': [{'OCL': ['Plt', '43'], 'EDF': ['3'], 'ICL': ['A01H', '503'], 'FSC': ['Plt'], 'FSS': ['43']}], 'inventors': [{'inventor name': 'Merrill, deceased; Grant', 'Street': '325 Breese Ave.', 'City': 'late of Red Bluff', 'State': 'CA'}, {'inventor name': 'Merrill, executrix; by Lucile B.', 'Street': '325 Breese Ave.', 'City': 'Red Bluff', 'State': 'CA', 'Zip code': '96080'}]}" } ``` #### github ``` { 'text': "/* filesystem.c\n * Filesystem utility routines\n *\n * Wireshark - Network traffic analyzer\n * By Gerald Combs <gerald@wireshark.org>\n * Copyright 1998 Gerald Combs\n *\n * SPDX-License-Identifier: GPL-2.0-or-later\n */\n\n#include <config.h>\n\n#include <stdio.h>\n#include <stdlib.h>\n#include <string.h>\n#include <errno.h>\n\n#include <glib.h>...", 'meta': "{'repo_name': 'wireshark/wireshark', 'stars': '2789', 'repo_language': 'C', 'file_name': 'packet-mpeg-audio-template.c', 'mime_type': 'text/x-c'}" } ``` </details> ### Data Fields #### all - `text` (str): Text. - `meta` (dict): Metadata of the data instance with keys: - pile_set_name: Name of the subset. <details> <summary>Expand to see individual components</summary> #### enron_emails - `text` (str): Text. - `meta` (str): Metadata of the data instance. #### europarl - `text` (str): Text. - `meta` (str): Metadata of the data instance with: language. #### free_law - `text` (str): Text. - `meta` (str): Metadata of the data instance with: case_ID, case_jurisdiction, date_created. #### hacker_news - `text` (str): Text. - `meta` (str): Metadata of the data instance with: id. #### nih_exporter - `text` (str): Text. - `meta` (str): Metadata of the data instance with: APPLICATION_ID. #### pubmed - `text` (str): Text. - `meta` (str): Metadata of the data instance with: pmid, language. #### pubmed_central - `text` (str): Text. - `meta` (str): Metadata of the data instance with: ID of the data instance. #### ubuntu_irc - `text` (str): Text. - `meta` (str): Metadata of the data instance with: channel, month. #### uspto - `text` (str): Text. - `meta` (str): Metadata of the data instance with: bibliographic_information, source_file, abstract, classifications, inventors. #### github - `text` (str): Text. - `meta` (str): Metadata of the data instance with: repo_name, stars, repo_language, file_name, mime_type. ### Data Splits The "all" configuration is composed of 3 splits: train, validation and test. </details> ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators This dataset was primarily curated by Leo Gao and Stella Biderman, with assistance from other authors of the Pile paper. ### Licensing Information Please refer to the specific license depending on the subset you use: - PubMed Central: [MIT License](https://github.com/EleutherAI/pile-pubmedcentral/blob/master/LICENSE) ### Citation Information ``` @article{gao2020pile, title={The {P}ile: An 800{GB} dataset of diverse text for language modeling}, author={Gao, Leo and Biderman, Stella and Black, Sid and Golding, Laurence and Hoppe, Travis and Foster, Charles and Phang, Jason and He, Horace and Thite, Anish and Nabeshima, Noa and others}, journal={arXiv preprint arXiv:2101.00027}, year={2020} } @article{biderman2022datasheet, title={Datasheet for the pile}, author={Biderman, Stella and Bicheno, Kieran and Gao, Leo}, journal={arXiv preprint arXiv:2201.07311}, year={2022} } ``` ### Contributions Thanks to [@github-username](https://github.com/<github-username>) for adding this dataset.
EleutherAI/pile
[ "task_categories:text-generation", "task_categories:fill-mask", "task_ids:language-modeling", "task_ids:masked-language-modeling", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:monolingual", "size_categories:100B<n<1T", "source_datasets:original", "language:en", "license:other", "arxiv:2201.07311", "arxiv:2101.00027", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["found"], "language": ["en"], "license": "other", "multilinguality": ["monolingual"], "size_categories": ["100B<n<1T"], "source_datasets": ["original"], "task_categories": ["text-generation", "fill-mask"], "task_ids": ["language-modeling", "masked-language-modeling"], "paperswithcode_id": "the-pile", "pretty_name": "the Pile"}
2023-05-03T14:58:14+00:00
[ "2201.07311", "2101.00027" ]
[ "en" ]
TAGS #task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-100B<n<1T #source_datasets-original #language-English #license-other #arxiv-2201.07311 #arxiv-2101.00027 #region-us
# Dataset Card for The Pile ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions This model card is a work in progress. Please also see our datasheet for more detailed info. ## Dataset Description - Homepage: URL - Repository: URL - Paper: The Pile: An 800GB Dataset of Diverse Text for Language Modeling - Leaderboard: - Point of Contact: EleutherAI - Datasheet: Datasheet for the Pile ### Dataset Summary The Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality datasets combined together. ### Supported Tasks and Leaderboards ### Languages This dataset is in English ('EN'). ## Dataset Structure ### Data Instances #### all <details> <summary>Expand to see individual components</summary> #### enron_emails #### europarl #### free_law #### hacker_news #### nih_exporter #### pubmed #### pubmed_central #### ubuntu_irc #### uspto #### github </details> ### Data Fields #### all - 'text' (str): Text. - 'meta' (dict): Metadata of the data instance with keys: - pile_set_name: Name of the subset. <details> <summary>Expand to see individual components</summary> #### enron_emails - 'text' (str): Text. - 'meta' (str): Metadata of the data instance. #### europarl - 'text' (str): Text. - 'meta' (str): Metadata of the data instance with: language. #### free_law - 'text' (str): Text. - 'meta' (str): Metadata of the data instance with: case_ID, case_jurisdiction, date_created. #### hacker_news - 'text' (str): Text. - 'meta' (str): Metadata of the data instance with: id. #### nih_exporter - 'text' (str): Text. - 'meta' (str): Metadata of the data instance with: APPLICATION_ID. #### pubmed - 'text' (str): Text. - 'meta' (str): Metadata of the data instance with: pmid, language. #### pubmed_central - 'text' (str): Text. - 'meta' (str): Metadata of the data instance with: ID of the data instance. #### ubuntu_irc - 'text' (str): Text. - 'meta' (str): Metadata of the data instance with: channel, month. #### uspto - 'text' (str): Text. - 'meta' (str): Metadata of the data instance with: bibliographic_information, source_file, abstract, classifications, inventors. #### github - 'text' (str): Text. - 'meta' (str): Metadata of the data instance with: repo_name, stars, repo_language, file_name, mime_type. ### Data Splits The "all" configuration is composed of 3 splits: train, validation and test. </details> ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators This dataset was primarily curated by Leo Gao and Stella Biderman, with assistance from other authors of the Pile paper. ### Licensing Information Please refer to the specific license depending on the subset you use: - PubMed Central: MIT License ### Contributions Thanks to @github-username for adding this dataset.
[ "# Dataset Card for The Pile", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions\n\nThis model card is a work in progress. Please also see our datasheet for more detailed info.", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: The Pile: An 800GB Dataset of Diverse Text for Language Modeling\n- Leaderboard:\n- Point of Contact: EleutherAI\n- Datasheet: Datasheet for the Pile", "### Dataset Summary\n\nThe Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality\ndatasets combined together.", "### Supported Tasks and Leaderboards", "### Languages\n\nThis dataset is in English ('EN').", "## Dataset Structure", "### Data Instances", "#### all\n\n<details>\n <summary>Expand to see individual components</summary>", "#### enron_emails", "#### europarl", "#### free_law", "#### hacker_news", "#### nih_exporter", "#### pubmed", "#### pubmed_central", "#### ubuntu_irc", "#### uspto", "#### github\n\n\n</details>", "### Data Fields", "#### all\n\n- 'text' (str): Text.\n- 'meta' (dict): Metadata of the data instance with keys:\n - pile_set_name: Name of the subset.\n\n<details>\n <summary>Expand to see individual components</summary>", "#### enron_emails\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance.", "#### europarl\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: language.", "#### free_law\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: case_ID, case_jurisdiction, date_created.", "#### hacker_news\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: id.", "#### nih_exporter\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: APPLICATION_ID.", "#### pubmed\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: pmid, language.", "#### pubmed_central\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: ID of the data instance.", "#### ubuntu_irc\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: channel, month.", "#### uspto\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: bibliographic_information, source_file, abstract, classifications, \n inventors.", "#### github\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: repo_name, stars, repo_language, file_name, mime_type.", "### Data Splits\n\nThe \"all\" configuration is composed of 3 splits: train, validation and test.\n\n</details>", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThis dataset was primarily curated by Leo Gao and Stella Biderman, with assistance from other authors of the Pile paper.", "### Licensing Information\n\nPlease refer to the specific license depending on the subset you use:\n- PubMed Central: MIT License", "### Contributions\n\nThanks to @github-username for adding this dataset." ]
[ "TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-100B<n<1T #source_datasets-original #language-English #license-other #arxiv-2201.07311 #arxiv-2101.00027 #region-us \n", "# Dataset Card for The Pile", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions\n\nThis model card is a work in progress. Please also see our datasheet for more detailed info.", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: The Pile: An 800GB Dataset of Diverse Text for Language Modeling\n- Leaderboard:\n- Point of Contact: EleutherAI\n- Datasheet: Datasheet for the Pile", "### Dataset Summary\n\nThe Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality\ndatasets combined together.", "### Supported Tasks and Leaderboards", "### Languages\n\nThis dataset is in English ('EN').", "## Dataset Structure", "### Data Instances", "#### all\n\n<details>\n <summary>Expand to see individual components</summary>", "#### enron_emails", "#### europarl", "#### free_law", "#### hacker_news", "#### nih_exporter", "#### pubmed", "#### pubmed_central", "#### ubuntu_irc", "#### uspto", "#### github\n\n\n</details>", "### Data Fields", "#### all\n\n- 'text' (str): Text.\n- 'meta' (dict): Metadata of the data instance with keys:\n - pile_set_name: Name of the subset.\n\n<details>\n <summary>Expand to see individual components</summary>", "#### enron_emails\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance.", "#### europarl\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: language.", "#### free_law\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: case_ID, case_jurisdiction, date_created.", "#### hacker_news\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: id.", "#### nih_exporter\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: APPLICATION_ID.", "#### pubmed\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: pmid, language.", "#### pubmed_central\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: ID of the data instance.", "#### ubuntu_irc\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: channel, month.", "#### uspto\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: bibliographic_information, source_file, abstract, classifications, \n inventors.", "#### github\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: repo_name, stars, repo_language, file_name, mime_type.", "### Data Splits\n\nThe \"all\" configuration is composed of 3 splits: train, validation and test.\n\n</details>", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThis dataset was primarily curated by Leo Gao and Stella Biderman, with assistance from other authors of the Pile paper.", "### Licensing Information\n\nPlease refer to the specific license depending on the subset you use:\n- PubMed Central: MIT License", "### Contributions\n\nThanks to @github-username for adding this dataset." ]
[ 126, 8, 146, 59, 41, 10, 15, 6, 6, 22, 7, 4, 5, 5, 6, 4, 6, 7, 4, 9, 5, 61, 30, 30, 45, 31, 36, 33, 36, 35, 45, 48, 30, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 35, 28, 19 ]
[ "passage: TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-100B<n<1T #source_datasets-original #language-English #license-other #arxiv-2201.07311 #arxiv-2101.00027 #region-us \n# Dataset Card for The Pile## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions\n\nThis model card is a work in progress. Please also see our datasheet for more detailed info.## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: The Pile: An 800GB Dataset of Diverse Text for Language Modeling\n- Leaderboard:\n- Point of Contact: EleutherAI\n- Datasheet: Datasheet for the Pile### Dataset Summary\n\nThe Pile is a 825 GiB diverse, open source language modelling data set that consists of 22 smaller, high-quality\ndatasets combined together.### Supported Tasks and Leaderboards### Languages\n\nThis dataset is in English ('EN').## Dataset Structure### Data Instances#### all\n\n<details>\n <summary>Expand to see individual components</summary>#### enron_emails#### europarl#### free_law#### hacker_news#### nih_exporter#### pubmed#### pubmed_central#### ubuntu_irc#### uspto#### github\n\n\n</details>### Data Fields", "passage: #### all\n\n- 'text' (str): Text.\n- 'meta' (dict): Metadata of the data instance with keys:\n - pile_set_name: Name of the subset.\n\n<details>\n <summary>Expand to see individual components</summary>#### enron_emails\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance.#### europarl\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: language.#### free_law\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: case_ID, case_jurisdiction, date_created.#### hacker_news\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: id.#### nih_exporter\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: APPLICATION_ID.#### pubmed\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: pmid, language.#### pubmed_central\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: ID of the data instance.#### ubuntu_irc\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: channel, month.#### uspto\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: bibliographic_information, source_file, abstract, classifications, \n inventors.#### github\n\n- 'text' (str): Text.\n- 'meta' (str): Metadata of the data instance with: repo_name, stars, repo_language, file_name, mime_type.### Data Splits\n\nThe \"all\" configuration is composed of 3 splits: train, validation and test.\n\n</details>## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information" ]
381bff9b9c647d5c0fa794e90f4f3170bf4a121d
# Dataset Card for the_pile_books3 ## Table of Contents - [Dataset Card for the_pile_books3](#dataset-card-for-the_pile_books3) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [|split|num examples|](#splitnum-examples) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Initial Data Collection and Normalization](#initial-data-collection-and-normalization) - [Who are the source language producers?](#who-are-the-source-language-producers) - [Annotations](#annotations) - [Annotation process](#annotation-process) - [Who are the annotators?](#who-are-the-annotators) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [GitHub](https://github.com/soskek/bookcorpus/issues/27#issuecomment-716104208) - **Repository:** [Needs More Information] - **Paper:** [arXiv](https://arxiv.org/abs/2101.00027) - **Leaderboard:** [Needs More Information] - **Point of Contact:** [Needs More Information] ### Dataset Summary <div class="course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400"> <p><b>Defunct:</b> Dataset "the_pile_books3" is defunct and no longer accessible due to reported copyright infringement.</p> </div> This dataset is Shawn Presser's work and is part of EleutherAi/The Pile dataset. This dataset contains all of bibliotik in plain .txt form, aka 197,000 books processed in exactly the same way as did for bookcorpusopen (a.k.a. books1). seems to be similar to OpenAI's mysterious "books2" dataset referenced in their papers. Unfortunately OpenAI will not give details, so we know very little about any differences. People suspect it's "all of libgen", but it's purely conjecture. |download_size|36.8 Gib| |dataset_size|100.9 Gib| ### Supported Tasks and Leaderboards This dataset is used for Language Modeling. ### Languages The dataset is in English. ## Dataset Structure ### Data Instances ``` {'title': '07 LEGO Ninjago - The Search For Zane (Scholastic) - Kate Howard (retail)' 'text': '\n\nTITLE PAGE\n\nFROM THE JOURNAL OF SENSEI GARMADON\n\nCHAPTER 1\n\nCHAPTER 2\n\nCHAPTER 3\n\nCHAPTER 4\n\nCHAPTER 5\n\nCHAPTER 6\n\nCHAPTER 7\n\nCHAPTER 8\n\nCHAPTER 9\n\nCOPYRIGHT\n\nThroughout Ninjago", five ninja are well-known for their speed, strength, and  of course  the elemental powers that help them protect our world from evil. But there are others who possess some of the same powers as the ninja. Others who may not always use their powers for good.\n\nBefore now, the ninja believed they were special. They di.......'} ``` ### Data Fields - `title`: title of the book - `text`: text content of the book ### Data Splits |split|num examples| -------------------------------- |train|196640| ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information MIT ### Citation Information ``` @article{pile, title={The {P}ile: An 800GB Dataset of Diverse Text for Language Modeling}, author={Gao, Leo and Biderman, Stella and Black, Sid and Golding, Laurence and Hoppe, Travis and Foster, Charles and Phang, Jason and He, Horace and Thite, Anish and Nabeshima, Noa and Presser, Shawn and Leahy, Connor}, journal={arXiv preprint arXiv:2101.00027}, year={2020} } ``` ### Contributions Thanks to [@shawwn](https://github.com/shawwn) for creating this dataset. Thanks to [@richarddwang](https://github.com/richarddwang) for adding this dataset.
the_pile_books3
[ "task_categories:text-generation", "task_categories:fill-mask", "task_ids:language-modeling", "task_ids:masked-language-modeling", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:en", "license:mit", "arxiv:2101.00027", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["found"], "language": ["en"], "license": ["mit"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["text-generation", "fill-mask"], "task_ids": ["language-modeling", "masked-language-modeling"], "pretty_name": "Books3", "viewer": false, "dataset_info": {"features": [{"name": "title", "dtype": "string"}, {"name": "text", "dtype": "string"}], "config_name": "plain_text", "splits": [{"name": "train", "num_bytes": 108392037000, "num_examples": 196639}], "download_size": 39516981435, "dataset_size": 108392037000}}
2024-01-18T11:17:08+00:00
[ "2101.00027" ]
[ "en" ]
TAGS #task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-mit #arxiv-2101.00027 #region-us
# Dataset Card for the_pile_books3 ## Table of Contents - Dataset Card for the_pile_books3 - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - |split|num examples| - Dataset Creation - Curation Rationale - Source Data - Initial Data Collection and Normalization - Who are the source language producers? - Annotations - Annotation process - Who are the annotators? - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: GitHub - Repository: - Paper: arXiv - Leaderboard: - Point of Contact: ### Dataset Summary <div class="course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400"> <p><b>Defunct:</b> Dataset "the_pile_books3" is defunct and no longer accessible due to reported copyright infringement.</p> </div> This dataset is Shawn Presser's work and is part of EleutherAi/The Pile dataset. This dataset contains all of bibliotik in plain .txt form, aka 197,000 books processed in exactly the same way as did for bookcorpusopen (a.k.a. books1). seems to be similar to OpenAI's mysterious "books2" dataset referenced in their papers. Unfortunately OpenAI will not give details, so we know very little about any differences. People suspect it's "all of libgen", but it's purely conjecture. |download_size|36.8 Gib| |dataset_size|100.9 Gib| ### Supported Tasks and Leaderboards This dataset is used for Language Modeling. ### Languages The dataset is in English. ## Dataset Structure ### Data Instances ### Data Fields - 'title': title of the book - 'text': text content of the book ### Data Splits |split|num examples| -------------------------------- |train|196640| ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information MIT ### Contributions Thanks to @shawwn for creating this dataset. Thanks to @richarddwang for adding this dataset.
[ "# Dataset Card for the_pile_books3", "## Table of Contents\n- Dataset Card for the_pile_books3\n - Table of Contents\n - Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n - Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n - |split|num examples|\n - Dataset Creation\n - Curation Rationale\n - Source Data\n - Initial Data Collection and Normalization\n - Who are the source language producers?\n - Annotations\n - Annotation process\n - Who are the annotators?\n - Personal and Sensitive Information\n - Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n - Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: GitHub\n- Repository: \n- Paper: arXiv\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\n<div class=\"course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400\">\n <p><b>Defunct:</b> Dataset \"the_pile_books3\" is defunct and no longer accessible due to reported copyright infringement.</p>\n</div>\n\n\nThis dataset is Shawn Presser's work and is part of EleutherAi/The Pile dataset. \n\nThis dataset contains all of bibliotik in plain .txt form, aka 197,000 books processed in exactly the same way as did for bookcorpusopen (a.k.a. books1). seems to be similar to OpenAI's mysterious \"books2\" dataset referenced in their papers. Unfortunately OpenAI will not give details, so we know very little about any differences. People suspect it's \"all of libgen\", but it's purely conjecture.\n\n|download_size|36.8 Gib|\n|dataset_size|100.9 Gib|", "### Supported Tasks and Leaderboards\n\nThis dataset is used for Language Modeling.", "### Languages\n\nThe dataset is in English.", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- 'title': title of the book\n- 'text': text content of the book", "### Data Splits\n\n|split|num examples|\n--------------------------------\n|train|196640|", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nMIT", "### Contributions\n\nThanks to @shawwn for creating this dataset.\nThanks to @richarddwang for adding this dataset." ]
[ "TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-mit #arxiv-2101.00027 #region-us \n", "# Dataset Card for the_pile_books3", "## Table of Contents\n- Dataset Card for the_pile_books3\n - Table of Contents\n - Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n - Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n - |split|num examples|\n - Dataset Creation\n - Curation Rationale\n - Source Data\n - Initial Data Collection and Normalization\n - Who are the source language producers?\n - Annotations\n - Annotation process\n - Who are the annotators?\n - Personal and Sensitive Information\n - Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n - Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: GitHub\n- Repository: \n- Paper: arXiv\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\n<div class=\"course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400\">\n <p><b>Defunct:</b> Dataset \"the_pile_books3\" is defunct and no longer accessible due to reported copyright infringement.</p>\n</div>\n\n\nThis dataset is Shawn Presser's work and is part of EleutherAi/The Pile dataset. \n\nThis dataset contains all of bibliotik in plain .txt form, aka 197,000 books processed in exactly the same way as did for bookcorpusopen (a.k.a. books1). seems to be similar to OpenAI's mysterious \"books2\" dataset referenced in their papers. Unfortunately OpenAI will not give details, so we know very little about any differences. People suspect it's \"all of libgen\", but it's purely conjecture.\n\n|download_size|36.8 Gib|\n|dataset_size|100.9 Gib|", "### Supported Tasks and Leaderboards\n\nThis dataset is used for Language Modeling.", "### Languages\n\nThe dataset is in English.", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- 'title': title of the book\n- 'text': text content of the book", "### Data Splits\n\n|split|num examples|\n--------------------------------\n|train|196640|", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nMIT", "### Contributions\n\nThanks to @shawwn for creating this dataset.\nThanks to @richarddwang for adding this dataset." ]
[ 117, 12, 177, 30, 304, 20, 11, 6, 6, 24, 25, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 7, 30 ]
[ "passage: TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-mit #arxiv-2101.00027 #region-us \n# Dataset Card for the_pile_books3## Table of Contents\n- Dataset Card for the_pile_books3\n - Table of Contents\n - Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n - Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n - |split|num examples|\n - Dataset Creation\n - Curation Rationale\n - Source Data\n - Initial Data Collection and Normalization\n - Who are the source language producers?\n - Annotations\n - Annotation process\n - Who are the annotators?\n - Personal and Sensitive Information\n - Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n - Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: GitHub\n- Repository: \n- Paper: arXiv\n- Leaderboard: \n- Point of Contact:" ]
cc3c2717731b0f847dd4047b1a67325f89210f12
# Dataset Card for the_pile_openwebtext2 ## Table of Contents - [Dataset Card for the_pile_openwebtext2](#dataset-card-for-the_pile_openwebtext2) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [|split|num examples|](#splitnum-examples) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Initial Data Collection and Normalization](#initial-data-collection-and-normalization) - [Who are the source language producers?](#who-are-the-source-language-producers) - [Annotations](#annotations) - [Annotation process](#annotation-process) - [Who are the annotators?](#who-are-the-annotators) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://openwebtext2.readthedocs.io/en/latest/ - **Repository:** [GitHub](https://github.com/EleutherAI/openwebtext2) - **Paper:** https://arxiv.org/abs/2101.00027 - **Leaderboard:** [Needs More Information] - **Point of Contact:** [Needs More Information] ### Dataset Summary <div class="course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400"> <p><b>Defunct:</b> Dataset "the_pile_openwebtext2" is defunct and no longer accessible due to unavailability of the source data.</p> </div> OpenWebText2 is part of EleutherAi/The Pile dataset and is an enhanced version of the original OpenWebTextCorpus covering all Reddit submissions from 2005 up until April 2020, with further months becoming available after the corresponding PushShift dump files are released. |download_size|27.3 Gib| |dataset_size|63.8 Gib| ### Supported Tasks and Leaderboards This dataset is used for Language Modeling. ### Languages This dataset is in English. ## Dataset Structure ### Data Instances ``` This example was too long and was cropped: {'title': Xiaomi Mi Note 10 Gearbest Coupon Promo Code [6+128GB] [France Warehouse], 'text': '27% off Xiaomi Mi Note 10 (CC9 Pro) 108MP Penta Camera Mobile Phone Global Version Online Smartphone – Black Gearbest Coupon Promo Code\n\nGearbest Coupon Price :$439.99\n\nRegular Price : $603.19 Your Save : $163.20 Coupon Limit: 100 times Warehouse: France Expires : September 30, 2020 Coupon Valid for...', 'reddit_scores': [6],} ``` ### Data Fields - `title`: title of the web page - `text`: text content of the web page - `reddit_scores`: scores of the reddit submissions that mention this web page, as a list of integers ### Data Splits |split|num examples| -------------------------------- |train|17103059| ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information [Needs More Information] ### Citation Information ``` @article{pile, title={The {P}ile: An 800GB Dataset of Diverse Text for Language Modeling}, author={Gao, Leo and Biderman, Stella and Black, Sid and Golding, Laurence and Hoppe, Travis and Foster, Charles and Phang, Jason and He, Horace and Thite, Anish and Nabeshima, Noa and Presser, Shawn and Leahy, Connor}, journal={arXiv preprint arXiv:2101.00027}, year={2020} } ``` ### Contributions [researcher2](https://github.com/researcher2) Wrote much of this code, with inspiration and some straight copying of the scraping code found [here](https://github.com/yet-another-account/openwebtext/).<br/> [sdtblck](https://github.com/sdtblck/) kindly put together the Colab notebook, and performed a chunk of the scraping. <br/> [leogao2](https://github.com/leogao2/) provided overall design guidance, lm_dataformat, and performed another chunk of scraping. <br /> [Colaboratory](https://colab.research.google.com/) VMs helped with about 10% of our overall scraping. <br /> [The Eye](http://the-eye.eu/) host the processed datasets.<br /> [Read The Docs](https://readthedocs.org/) host our documentation.<br /> [@richarddwang](https://github.com/richarddwang) added this dataset to HF/datasets.
the_pile_openwebtext2
[ "task_categories:text-generation", "task_categories:fill-mask", "task_categories:text-classification", "task_ids:language-modeling", "task_ids:masked-language-modeling", "task_ids:text-scoring", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:monolingual", "size_categories:10M<n<100M", "source_datasets:original", "language:en", "license:mit", "arxiv:2101.00027", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["found"], "language": ["en"], "license": ["mit"], "multilinguality": ["monolingual"], "size_categories": ["10M<n<100M"], "source_datasets": ["original"], "task_categories": ["text-generation", "fill-mask", "text-classification"], "task_ids": ["language-modeling", "masked-language-modeling", "text-scoring"], "pretty_name": "OpenWebText2", "dataset_info": {"features": [{"name": "title", "dtype": "string"}, {"name": "text", "dtype": "string"}], "config_name": "plain_text", "splits": [{"name": "train", "num_bytes": 68571017395, "num_examples": 17103059}], "download_size": 29344276480, "dataset_size": 68571017395}, "viewer": false}
2023-11-27T14:54:23+00:00
[ "2101.00027" ]
[ "en" ]
TAGS #task_categories-text-generation #task_categories-fill-mask #task_categories-text-classification #task_ids-language-modeling #task_ids-masked-language-modeling #task_ids-text-scoring #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-10M<n<100M #source_datasets-original #language-English #license-mit #arxiv-2101.00027 #region-us
# Dataset Card for the_pile_openwebtext2 ## Table of Contents - Dataset Card for the_pile_openwebtext2 - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - |split|num examples| - Dataset Creation - Curation Rationale - Source Data - Initial Data Collection and Normalization - Who are the source language producers? - Annotations - Annotation process - Who are the annotators? - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: GitHub - Paper: URL - Leaderboard: - Point of Contact: ### Dataset Summary <div class="course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400"> <p><b>Defunct:</b> Dataset "the_pile_openwebtext2" is defunct and no longer accessible due to unavailability of the source data.</p> </div> OpenWebText2 is part of EleutherAi/The Pile dataset and is an enhanced version of the original OpenWebTextCorpus covering all Reddit submissions from 2005 up until April 2020, with further months becoming available after the corresponding PushShift dump files are released. |download_size|27.3 Gib| |dataset_size|63.8 Gib| ### Supported Tasks and Leaderboards This dataset is used for Language Modeling. ### Languages This dataset is in English. ## Dataset Structure ### Data Instances ### Data Fields - 'title': title of the web page - 'text': text content of the web page - 'reddit_scores': scores of the reddit submissions that mention this web page, as a list of integers ### Data Splits |split|num examples| -------------------------------- |train|17103059| ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions researcher2 Wrote much of this code, with inspiration and some straight copying of the scraping code found here.<br/> sdtblck kindly put together the Colab notebook, and performed a chunk of the scraping. <br/> leogao2 provided overall design guidance, lm_dataformat, and performed another chunk of scraping. <br /> Colaboratory VMs helped with about 10% of our overall scraping. <br /> The Eye host the processed datasets.<br /> Read The Docs host our documentation.<br /> @richarddwang added this dataset to HF/datasets.
[ "# Dataset Card for the_pile_openwebtext2", "## Table of Contents\n- Dataset Card for the_pile_openwebtext2\n - Table of Contents\n - Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n - Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n - |split|num examples|\n - Dataset Creation\n - Curation Rationale\n - Source Data\n - Initial Data Collection and Normalization\n - Who are the source language producers?\n - Annotations\n - Annotation process\n - Who are the annotators?\n - Personal and Sensitive Information\n - Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n - Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: GitHub\n- Paper: URL\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\n<div class=\"course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400\">\n <p><b>Defunct:</b> Dataset \"the_pile_openwebtext2\" is defunct and no longer accessible due to unavailability of the source data.</p>\n</div>\n\nOpenWebText2 is part of EleutherAi/The Pile dataset and is an enhanced version of the original OpenWebTextCorpus covering all Reddit submissions from 2005 up until April 2020, with further months becoming available after the corresponding PushShift dump files are released.\n\n|download_size|27.3 Gib|\n|dataset_size|63.8 Gib|", "### Supported Tasks and Leaderboards\n\nThis dataset is used for Language Modeling.", "### Languages\n\nThis dataset is in English.", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- 'title': title of the web page\n- 'text': text content of the web page\n- 'reddit_scores': scores of the reddit submissions that mention this web page, as a list of integers", "### Data Splits\n\n|split|num examples|\n--------------------------------\n|train|17103059|", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nresearcher2 Wrote much of this code, with inspiration and some straight copying of the scraping code found here.<br/>\nsdtblck kindly put together the Colab notebook, and performed a chunk of the scraping. <br/>\nleogao2 provided overall design guidance, lm_dataformat, and performed another chunk of scraping. <br />\nColaboratory VMs helped with about 10% of our overall scraping. <br />\nThe Eye host the processed datasets.<br />\nRead The Docs host our documentation.<br />\n\n@richarddwang added this dataset to HF/datasets." ]
[ "TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_categories-text-classification #task_ids-language-modeling #task_ids-masked-language-modeling #task_ids-text-scoring #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-10M<n<100M #source_datasets-original #language-English #license-mit #arxiv-2101.00027 #region-us \n", "# Dataset Card for the_pile_openwebtext2", "## Table of Contents\n- Dataset Card for the_pile_openwebtext2\n - Table of Contents\n - Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n - Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n - |split|num examples|\n - Dataset Creation\n - Curation Rationale\n - Source Data\n - Initial Data Collection and Normalization\n - Who are the source language producers?\n - Annotations\n - Annotation process\n - Who are the annotators?\n - Personal and Sensitive Information\n - Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n - Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: GitHub\n- Paper: URL\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\n<div class=\"course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400\">\n <p><b>Defunct:</b> Dataset \"the_pile_openwebtext2\" is defunct and no longer accessible due to unavailability of the source data.</p>\n</div>\n\nOpenWebText2 is part of EleutherAi/The Pile dataset and is an enhanced version of the original OpenWebTextCorpus covering all Reddit submissions from 2005 up until April 2020, with further months becoming available after the corresponding PushShift dump files are released.\n\n|download_size|27.3 Gib|\n|dataset_size|63.8 Gib|", "### Supported Tasks and Leaderboards\n\nThis dataset is used for Language Modeling.", "### Languages\n\nThis dataset is in English.", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- 'title': title of the web page\n- 'text': text content of the web page\n- 'reddit_scores': scores of the reddit submissions that mention this web page, as a list of integers", "### Data Splits\n\n|split|num examples|\n--------------------------------\n|train|17103059|", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nresearcher2 Wrote much of this code, with inspiration and some straight copying of the scraping code found here.<br/>\nsdtblck kindly put together the Colab notebook, and performed a chunk of the scraping. <br/>\nleogao2 provided overall design guidance, lm_dataformat, and performed another chunk of scraping. <br />\nColaboratory VMs helped with about 10% of our overall scraping. <br />\nThe Eye host the processed datasets.<br />\nRead The Docs host our documentation.<br />\n\n@richarddwang added this dataset to HF/datasets." ]
[ 138, 14, 179, 29, 237, 20, 11, 6, 6, 54, 26, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 147 ]
[ "passage: TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_categories-text-classification #task_ids-language-modeling #task_ids-masked-language-modeling #task_ids-text-scoring #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-10M<n<100M #source_datasets-original #language-English #license-mit #arxiv-2101.00027 #region-us \n# Dataset Card for the_pile_openwebtext2## Table of Contents\n- Dataset Card for the_pile_openwebtext2\n - Table of Contents\n - Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n - Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n - |split|num examples|\n - Dataset Creation\n - Curation Rationale\n - Source Data\n - Initial Data Collection and Normalization\n - Who are the source language producers?\n - Annotations\n - Annotation process\n - Who are the annotators?\n - Personal and Sensitive Information\n - Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n - Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: GitHub\n- Paper: URL\n- Leaderboard: \n- Point of Contact:" ]
f171faab8c7d2192a01295fa386c00156cf9fe46
# Dataset Card for Stack Exchange ## Table of Contents - [Dataset Card for Stack Exchange](#dataset-card-for-the_pile_stack_exchange) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [|split|num examples|](#splitnum-examples) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Initial Data Collection and Normalization](#initial-data-collection-and-normalization) - [Who are the source language producers?](#who-are-the-source-language-producers) - [Annotations](#annotations) - [Annotation process](#annotation-process) - [Who are the annotators?](#who-are-the-annotators) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [GitHub](https://github.com/EleutherAI/stackexchange-dataset) - **Repository:** [Needs More Information] - **Paper:** [arXiv](https://arxiv.org/abs/2101.00027) - **Leaderboard:** [Needs More Information] - **Point of Contact:** [Needs More Information] ### Dataset Summary <div class="course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400"> <p><b>Defunct:</b> Dataset "the_pile_stack_exchange" is defunct and no longer accessible due to unavailability of the source data.</p> </div> This dataset is part of EleutherAI/The Pile dataset and is a dataset for Language Models from processing stackexchange data dump, which is an anonymized dump of all user-contributed content on the Stack Exchange network. |download_size|34.28 Gib| |dataset_size|10.3 Gib| ### Supported Tasks and Leaderboards The dataset is used for Language Modeling. ### Languages The dataset is in English. ## Dataset Structure ### Data Instances ``` {'domain': 'chemistry', 'text':"\nQ: \n \nReviving old questions or asking a new one? \n \nI'm relatively new to the Chemistry SE community, and sometimes when I go to ask a question, I notice that the same (or similar) question has \nalready been asked. However, the previous question doesn't have a good answer (or is unanswered). In this case, is it better to ask the questi\non again in a new post (which might be marked as duplicate) or comment on the old post (which might be several years old)? In other words, wha\nt are the customs of this site in regards to reviving old questions/discussions?\n\nA:\n\nAs Martin commented, it really depends on the type of question. In any case, you always have the following possibilities:\n\nAsk a new question\nEdit the question to bump it to the first page\nAdd a bounty\nBring it to the attention of people in chat\n\nConsider the following cases:\n\nI have exactly the same question as asked and unanswered before!\n\nIf you ask a new question which turns out to be the same question, it may be closed as a dupe (depending on whether users remember the old que\nstion). Not the ideal option.\nIf you can find something substantial to edit and bump the question, do so. Maybe add a comment that you would really love an answer.\nIf you can spare some rep for a bounty (50 is usually enough), do so.\nYou can always bring it to the attention of people in chat.\n",} ``` ### Data Fields - `domain`: Stack Exchange domain of the sample - `text`: Text content containing both the question and the answer ### Data Splits |split|num examples| -------------------------------- |train|5096117| ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information [Needs More Information] ### Citation Information ``` @article{pile, title={The {P}ile: An 800GB Dataset of Diverse Text for Language Modeling}, author={Gao, Leo and Biderman, Stella and Black, Sid and Golding, Laurence and Hoppe, Travis and Foster, Charles and Phang, Jason and He, Horace and Thite, Anish and Nabeshima, Noa and Presser, Shawn and Leahy, Connor}, journal={arXiv preprint arXiv:2101.00027}, year={2020} } ``` ### Contributions Thanks to [sdtblck](https://github.com/sdtblck) for creating the dataset. Thanks to [richarddwang](https://github.com/richarddwang) for adding the dataset.
the_pile_stack_exchange
[ "task_categories:text-generation", "task_categories:fill-mask", "task_ids:language-modeling", "task_ids:masked-language-modeling", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:monolingual", "size_categories:1M<n<10M", "source_datasets:original", "language:en", "license:cc-by-sa-4.0", "arxiv:2101.00027", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["found"], "language": ["en"], "license": ["cc-by-sa-4.0"], "multilinguality": ["monolingual"], "size_categories": ["1M<n<10M"], "source_datasets": ["original"], "task_categories": ["text-generation", "fill-mask"], "task_ids": ["language-modeling", "masked-language-modeling"], "pretty_name": "Stack Exchange", "dataset_info": {"features": [{"name": "domain", "dtype": "string"}, {"name": "text", "dtype": "string"}], "config_name": "plain_text", "splits": [{"name": "train", "num_bytes": 11075434609, "num_examples": 5096117}], "download_size": 36802959360, "dataset_size": 11075434609}, "viewer": false}
2023-11-27T15:00:44+00:00
[ "2101.00027" ]
[ "en" ]
TAGS #task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-1M<n<10M #source_datasets-original #language-English #license-cc-by-sa-4.0 #arxiv-2101.00027 #region-us
# Dataset Card for Stack Exchange ## Table of Contents - Dataset Card for Stack Exchange - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - |split|num examples| - Dataset Creation - Curation Rationale - Source Data - Initial Data Collection and Normalization - Who are the source language producers? - Annotations - Annotation process - Who are the annotators? - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: GitHub - Repository: - Paper: arXiv - Leaderboard: - Point of Contact: ### Dataset Summary <div class="course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400"> <p><b>Defunct:</b> Dataset "the_pile_stack_exchange" is defunct and no longer accessible due to unavailability of the source data.</p> </div> This dataset is part of EleutherAI/The Pile dataset and is a dataset for Language Models from processing stackexchange data dump, which is an anonymized dump of all user-contributed content on the Stack Exchange network. |download_size|34.28 Gib| |dataset_size|10.3 Gib| ### Supported Tasks and Leaderboards The dataset is used for Language Modeling. ### Languages The dataset is in English. ## Dataset Structure ### Data Instances ### Data Fields - 'domain': Stack Exchange domain of the sample - 'text': Text content containing both the question and the answer ### Data Splits |split|num examples| -------------------------------- |train|5096117| ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to sdtblck for creating the dataset. Thanks to richarddwang for adding the dataset.
[ "# Dataset Card for Stack Exchange", "## Table of Contents\n- Dataset Card for Stack Exchange\n - Table of Contents\n - Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n - Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n - |split|num examples|\n - Dataset Creation\n - Curation Rationale\n - Source Data\n - Initial Data Collection and Normalization\n - Who are the source language producers?\n - Annotations\n - Annotation process\n - Who are the annotators?\n - Personal and Sensitive Information\n - Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n - Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: GitHub\n- Repository: \n- Paper: arXiv\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\n<div class=\"course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400\">\n <p><b>Defunct:</b> Dataset \"the_pile_stack_exchange\" is defunct and no longer accessible due to unavailability of the source data.</p>\n</div>\n\nThis dataset is part of EleutherAI/The Pile dataset and is a dataset for Language Models from processing stackexchange data dump, which is an anonymized dump of all user-contributed content on the Stack Exchange network.\n\n|download_size|34.28 Gib|\n|dataset_size|10.3 Gib|", "### Supported Tasks and Leaderboards\n\nThe dataset is used for Language Modeling.", "### Languages\n\nThe dataset is in English.", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- 'domain': Stack Exchange domain of the sample\n- 'text': Text content containing both the question and the answer", "### Data Splits\n\n|split|num examples|\n--------------------------------\n|train|5096117|", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\nThanks to sdtblck for creating the dataset.\nThanks to richarddwang for adding the dataset." ]
[ "TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-1M<n<10M #source_datasets-original #language-English #license-cc-by-sa-4.0 #arxiv-2101.00027 #region-us \n", "# Dataset Card for Stack Exchange", "## Table of Contents\n- Dataset Card for Stack Exchange\n - Table of Contents\n - Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n - Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n - |split|num examples|\n - Dataset Creation\n - Curation Rationale\n - Source Data\n - Initial Data Collection and Normalization\n - Who are the source language producers?\n - Annotations\n - Annotation process\n - Who are the annotators?\n - Personal and Sensitive Information\n - Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n - Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: GitHub\n- Repository: \n- Paper: arXiv\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\n<div class=\"course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400\">\n <p><b>Defunct:</b> Dataset \"the_pile_stack_exchange\" is defunct and no longer accessible due to unavailability of the source data.</p>\n</div>\n\nThis dataset is part of EleutherAI/The Pile dataset and is a dataset for Language Models from processing stackexchange data dump, which is an anonymized dump of all user-contributed content on the Stack Exchange network.\n\n|download_size|34.28 Gib|\n|dataset_size|10.3 Gib|", "### Supported Tasks and Leaderboards\n\nThe dataset is used for Language Modeling.", "### Languages\n\nThe dataset is in English.", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- 'domain': Stack Exchange domain of the sample\n- 'text': Text content containing both the question and the answer", "### Data Splits\n\n|split|num examples|\n--------------------------------\n|train|5096117|", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\nThanks to sdtblck for creating the dataset.\nThanks to richarddwang for adding the dataset." ]
[ 123, 8, 173, 30, 235, 20, 11, 6, 6, 33, 26, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 29 ]
[ "passage: TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-1M<n<10M #source_datasets-original #language-English #license-cc-by-sa-4.0 #arxiv-2101.00027 #region-us \n# Dataset Card for Stack Exchange## Table of Contents\n- Dataset Card for Stack Exchange\n - Table of Contents\n - Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n - Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n - |split|num examples|\n - Dataset Creation\n - Curation Rationale\n - Source Data\n - Initial Data Collection and Normalization\n - Who are the source language producers?\n - Annotations\n - Annotation process\n - Who are the annotators?\n - Personal and Sensitive Information\n - Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n - Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: GitHub\n- Repository: \n- Paper: arXiv\n- Leaderboard: \n- Point of Contact:" ]
c8d4e75c6a01b7a1248b36b11bf70913afc6178a
# Dataset Card for Tilde Multilingual Open Data for European Languages ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** http://opus.nlpl.eu/TildeMODEL.php - **Repository:** None - **Paper:** https://www.aclweb.org/anthology/W17-0235.pdf - **Leaderboard:** [More Information Needed] - **Point of Contact:** [More Information Needed] ### Dataset Summary To load a language pair which isn't part of the config, all you need to do is specify the language code as pairs. You can find the valid pairs in Homepage section of Dataset Description: http://opus.nlpl.eu/TildeMODEL.php E.g. `dataset = load_dataset("tilde_model", lang1="en", lang2="lv")` ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances Here are some examples of questions and facts: ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@abhishekkrthakur](https://github.com/abhishekkrthakur) for adding this dataset.
tilde_model
[ "task_categories:translation", "annotations_creators:found", "language_creators:found", "multilinguality:multilingual", "size_categories:n<1K", "source_datasets:original", "language:bg", "language:cs", "language:da", "language:de", "language:el", "language:en", "language:es", "language:et", "language:fi", "language:fr", "language:hr", "language:hu", "language:is", "language:it", "language:lt", "language:lv", "language:mt", "language:nl", "language:no", "language:pl", "language:pt", "language:ro", "language:ru", "language:sk", "language:sl", "language:sq", "language:sr", "language:sv", "language:tr", "language:uk", "license:cc-by-sa-4.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["bg", "cs", "da", "de", "el", "en", "es", "et", "fi", "fr", "hr", "hu", "is", "it", "lt", "lv", "mt", "nl", "no", "pl", "pt", "ro", "ru", "sk", "sl", "sq", "sr", "sv", "tr", "uk"], "license": ["cc-by-sa-4.0"], "multilinguality": ["multilingual"], "size_categories": ["n<1K"], "source_datasets": ["original"], "task_categories": ["translation"], "task_ids": [], "paperswithcode_id": "tilde-model-corpus", "pretty_name": "Tilde Multilingual Open Data for European Languages", "dataset_info": [{"config_name": "bg-el", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["bg", "el"]}}}], "splits": [{"name": "train", "num_bytes": 258081, "num_examples": 455}], "download_size": 64430, "dataset_size": 258081}, {"config_name": "cs-en", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["cs", "en"]}}}], "splits": [{"name": "train", "num_bytes": 709168, "num_examples": 3100}], "download_size": 201503, "dataset_size": 709168}, {"config_name": "de-hr", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["de", "hr"]}}}], "splits": [{"name": "train", "num_bytes": 180148538, "num_examples": 683194}], "download_size": 49585877, "dataset_size": 180148538}, {"config_name": "en-no", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "no"]}}}], "splits": [{"name": "train", "num_bytes": 73797124, "num_examples": 348141}], "download_size": 17852861, "dataset_size": 73797124}, {"config_name": "es-pt", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["es", "pt"]}}}], "splits": [{"name": "train", "num_bytes": 3808423, "num_examples": 13464}], "download_size": 1160892, "dataset_size": 3808423}]}
2024-01-18T11:17:10+00:00
[]
[ "bg", "cs", "da", "de", "el", "en", "es", "et", "fi", "fr", "hr", "hu", "is", "it", "lt", "lv", "mt", "nl", "no", "pl", "pt", "ro", "ru", "sk", "sl", "sq", "sr", "sv", "tr", "uk" ]
TAGS #task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-n<1K #source_datasets-original #language-Bulgarian #language-Czech #language-Danish #language-German #language-Modern Greek (1453-) #language-English #language-Spanish #language-Estonian #language-Finnish #language-French #language-Croatian #language-Hungarian #language-Icelandic #language-Italian #language-Lithuanian #language-Latvian #language-Maltese #language-Dutch #language-Norwegian #language-Polish #language-Portuguese #language-Romanian #language-Russian #language-Slovak #language-Slovenian #language-Albanian #language-Serbian #language-Swedish #language-Turkish #language-Ukrainian #license-cc-by-sa-4.0 #region-us
# Dataset Card for Tilde Multilingual Open Data for European Languages ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: None - Paper: URL - Leaderboard: - Point of Contact: ### Dataset Summary To load a language pair which isn't part of the config, all you need to do is specify the language code as pairs. You can find the valid pairs in Homepage section of Dataset Description: URL E.g. 'dataset = load_dataset("tilde_model", lang1="en", lang2="lv")' ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances Here are some examples of questions and facts: ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @abhishekkrthakur for adding this dataset.
[ "# Dataset Card for Tilde Multilingual Open Data for European Languages", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: URL\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\nTo load a language pair which isn't part of the config, all you need to do is specify the language code as pairs.\nYou can find the valid pairs in Homepage section of Dataset Description: URL\nE.g.\n\n'dataset = load_dataset(\"tilde_model\", lang1=\"en\", lang2=\"lv\")'", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances\n\nHere are some examples of questions and facts:", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ "TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-n<1K #source_datasets-original #language-Bulgarian #language-Czech #language-Danish #language-German #language-Modern Greek (1453-) #language-English #language-Spanish #language-Estonian #language-Finnish #language-French #language-Croatian #language-Hungarian #language-Icelandic #language-Italian #language-Lithuanian #language-Latvian #language-Maltese #language-Dutch #language-Norwegian #language-Polish #language-Portuguese #language-Romanian #language-Russian #language-Slovak #language-Slovenian #language-Albanian #language-Serbian #language-Swedish #language-Turkish #language-Ukrainian #license-cc-by-sa-4.0 #region-us \n", "# Dataset Card for Tilde Multilingual Open Data for European Languages", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: URL\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\nTo load a language pair which isn't part of the config, all you need to do is specify the language code as pairs.\nYou can find the valid pairs in Homepage section of Dataset Description: URL\nE.g.\n\n'dataset = load_dataset(\"tilde_model\", lang1=\"en\", lang2=\"lv\")'", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances\n\nHere are some examples of questions and facts:", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ 243, 16, 120, 28, 82, 10, 4, 6, 17, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 20 ]
[ "passage: TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-n<1K #source_datasets-original #language-Bulgarian #language-Czech #language-Danish #language-German #language-Modern Greek (1453-) #language-English #language-Spanish #language-Estonian #language-Finnish #language-French #language-Croatian #language-Hungarian #language-Icelandic #language-Italian #language-Lithuanian #language-Latvian #language-Maltese #language-Dutch #language-Norwegian #language-Polish #language-Portuguese #language-Romanian #language-Russian #language-Slovak #language-Slovenian #language-Albanian #language-Serbian #language-Swedish #language-Turkish #language-Ukrainian #license-cc-by-sa-4.0 #region-us \n# Dataset Card for Tilde Multilingual Open Data for European Languages## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: URL\n- Leaderboard: \n- Point of Contact:### Dataset Summary\n\nTo load a language pair which isn't part of the config, all you need to do is specify the language code as pairs.\nYou can find the valid pairs in Homepage section of Dataset Description: URL\nE.g.\n\n'dataset = load_dataset(\"tilde_model\", lang1=\"en\", lang2=\"lv\")'### Supported Tasks and Leaderboards### Languages" ]
cf456593a9d6d96cb02c109a3d0df62fde42defd
# Dataset Card for TimeDial: Temporal Commonsense Reasoning in Dialog ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [TimeDial](https://github.com/google-research-datasets/timedial) - **Paper:** [TimeDial: Temporal Commonsense Reasoning in Dialog](https://arxiv.org/abs/2106.04571) - **Point of Contact:** [Please create an issue in the official repository](https://github.com/google-research-datasets/timedial) ### Dataset Summary TimeDial presents a crowdsourced English challenge set, for temporal commonsense reasoning, formulated as a multiple choice cloze task with around 1.5k carefully curated dialogs. The dataset is derived from the DailyDialog ([Li et al., 2017](https://www.aclweb.org/anthology/I17-1099/)), which is a multi-turn dialog corpus. In order to establish strong baselines and provide information on future model development, the authors conducted extensive experiments with state-of-the-art LMs. While humans can easily answer these questions (97.8\%), the best T5 model variant struggles on this challenge set (73\%). Moreover, our qualitative error analyses show that the models often rely on shallow, spurious features (particularly text matching), instead of truly doing reasoning over the context. Detailed experiments and analyses can be found in their [paper](https://arxiv.org/pdf/2106.04571.pdf). ### Supported Tasks and Leaderboards To be updated soon. ### Languages The dataset is in English only. ## Dataset Structure ### Data Instances ``` { "id": 1, "conversation": [ "A: We need to take the accounts system offline to carry out the upgrade . But don't worry , it won't cause too much inconvenience . We're going to do it over the weekend .", "B: How long will the system be down for ?", "A: We'll be taking everything offline in about two hours ' time . It'll be down for a minimum of twelve hours . If everything goes according to plan , it should be up again by 6 pm on Saturday .", "B: That's fine . We've allowed <MASK> to be on the safe side ." ], "correct1": "forty-eight hours", "correct2": "50 hours ", "incorrect1": "two hours ", "incorrect1_rule": "Rule 1", "incorrect2": "12 days ", "incorrect2_rule": "Rule 2" } ``` ### Data Fields - "id": Unique identifier, as a integer - "conversation": Dialog context with <MASK> span, as a string - "correct1": Original <MASK> span, as a string - "correct2": Additional correct option provided by annotators, as a string - "incorrect1": Incorrect option #1 provided by annotators, as a string - "incorrect1_rule": One of phrase matching ("Rule 1"), numeral matching ("Rule 2"), or open ended ("Rule 3"), as a string - "incorrect2": Incorrect option #2 provided by annotators, as a string - "incorrect2_rule": One of phrase matching ("Rule 1"), numeral matching ("Rule 2"), or open ended ("Rule 3"), as a string ### Data Splits TimeDial dataset consists only of a test set of 1,104 dialog instances with 2 correct and 2 incorrect options with the following statistics: | | Avg. | |-----|-----| |Turns per Dialog | 11.7 | |Words per Turn | 16.5 | |Time Spans per Dialog | 3 | ## Dataset Creation ### Curation Rationale Although previous works have studied temporal reasoning in natural language, they have either focused on specific time-related concepts in isolation, such as temporal ordering and relation extraction, and/or dealt with limited context, such as single-sentence-based question answering and natural language inference. In this work, they make the first systematic study of temporal commonsense reasoning in a multi-turn dialog setting. The task involves complex reasoning that requires operations like comparison and arithmetic reasoning over temporal expressions and the need for commonsense and world knowledge. ### Source Data #### Initial Data Collection and Normalization The TIMEDIAL dataset is derived from DailyDialog data (Li et al., 2017), which is a multi-turn dialog corpus containing over 13K English dialogs. Dialogs in this dataset consist of turn-taking between two people on topics over 10 broad categories, ranging from daily lives to financial topics. #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process The data collection process involves two steps: (1) identifying dialogs that are rich in temporal expressions, and (2) asking human annotators to provide correct and incorrect options for cloze instances derived from these dialogs. More details about the two steps: 1) Temporal expression identification: Here, they select dialogs that are rich with temporal information, in order to focus on complex temporal reasoning that arises in natural dialogs. Temporal expressions are automatically identified with SU-Time, an off-the-shelf temporal expression detector. They keep only the dialogs with more than 3 temporal expressions and at least one expression that contains numerals like “two weeks” (as opposed to non-numeric spans, like “summer”, “right now”, and “later”). In their initial experiment, they observe that language models can often correctly predict these non-numerical temporal phrases. 2) Human annotated options: Next, they make spans in the dialogs. For a dialog, they mask out each temporal expression that contains numerals, each resulting in a cloze question that is then sent for human annotation. This resulted in 1,526 instances for annotation. For each masked span in each dialog, they obtain human annotation to derive a fixed set of correct and incorrect options given the context. Concretely, given a masked dialog and a seed correct answer (i.e., the original text) for the masked span, the annotators were asked to (1) come up with an alternative correct answer that makes sense in the dialog adhering to commonsense, and (2) formulate two incorrect answers that have no possibility of making sense in the dialog context. They highlight all time expressions in the context to make it easier for annotators to select reasonable time expressions. #### Who are the annotators? They are English linguists. ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations Dataset provided for research purposes only. Please check dataset license for additional information. ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information TimeDial dataset is licensed under CC BY-NC-SA 4.0. ### Citation Information ``` @inproceedings{qin-etal-2021-timedial, title = "{TimeDial: Temporal Commonsense Reasoning in Dialog}", author = "Qin, Lianhui and Gupta, Aditya and Upadhyay, Shyam and He, Luheng and Choi, Yejin and Faruqui, Manaal", booktitle = "Proc. of ACL", year = "2021" } ``` ### Contributions Thanks to [@bhavitvyamalik](https://github.com/bhavitvyamalik) for adding this dataset.
time_dial
[ "task_categories:text-classification", "task_ids:multi-label-classification", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:cc-by-nc-sa-4.0", "dialog-act-classification", "arxiv:2106.04571", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["expert-generated"], "language": ["en"], "license": ["cc-by-nc-sa-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["multi-label-classification"], "paperswithcode_id": "timedial", "pretty_name": "TimeDial: Temporal Commonsense Reasoning in Dialog", "tags": ["dialog-act-classification"], "dataset_info": {"features": [{"name": "id", "dtype": "int32"}, {"name": "conversation", "sequence": "string"}, {"name": "correct1", "dtype": "string"}, {"name": "correct2", "dtype": "string"}, {"name": "incorrect1", "dtype": "string"}, {"name": "incorrect1_rule", "dtype": "string"}, {"name": "incorrect2", "dtype": "string"}, {"name": "incorrect2_rule", "dtype": "string"}], "splits": [{"name": "test", "num_bytes": 1449879, "num_examples": 1446}], "download_size": 1613806, "dataset_size": 1449879}}
2024-01-18T11:17:11+00:00
[ "2106.04571" ]
[ "en" ]
TAGS #task_categories-text-classification #task_ids-multi-label-classification #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-nc-sa-4.0 #dialog-act-classification #arxiv-2106.04571 #region-us
Dataset Card for TimeDial: Temporal Commonsense Reasoning in Dialog =================================================================== Table of Contents ----------------- * Table of Contents * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: TimeDial * Paper: TimeDial: Temporal Commonsense Reasoning in Dialog * Point of Contact: Please create an issue in the official repository ### Dataset Summary TimeDial presents a crowdsourced English challenge set, for temporal commonsense reasoning, formulated as a multiple choice cloze task with around 1.5k carefully curated dialogs. The dataset is derived from the DailyDialog (Li et al., 2017), which is a multi-turn dialog corpus. In order to establish strong baselines and provide information on future model development, the authors conducted extensive experiments with state-of-the-art LMs. While humans can easily answer these questions (97.8%), the best T5 model variant struggles on this challenge set (73%). Moreover, our qualitative error analyses show that the models often rely on shallow, spurious features (particularly text matching), instead of truly doing reasoning over the context. Detailed experiments and analyses can be found in their paper. ### Supported Tasks and Leaderboards To be updated soon. ### Languages The dataset is in English only. Dataset Structure ----------------- ### Data Instances ### Data Fields * "id": Unique identifier, as a integer * "conversation": Dialog context with span, as a string * "correct1": Original span, as a string * "correct2": Additional correct option provided by annotators, as a string * "incorrect1": Incorrect option #1 provided by annotators, as a string * "incorrect1\_rule": One of phrase matching ("Rule 1"), numeral matching ("Rule 2"), or open ended ("Rule 3"), as a string * "incorrect2": Incorrect option #2 provided by annotators, as a string * "incorrect2\_rule": One of phrase matching ("Rule 1"), numeral matching ("Rule 2"), or open ended ("Rule 3"), as a string ### Data Splits TimeDial dataset consists only of a test set of 1,104 dialog instances with 2 correct and 2 incorrect options with the following statistics: Dataset Creation ---------------- ### Curation Rationale Although previous works have studied temporal reasoning in natural language, they have either focused on specific time-related concepts in isolation, such as temporal ordering and relation extraction, and/or dealt with limited context, such as single-sentence-based question answering and natural language inference. In this work, they make the first systematic study of temporal commonsense reasoning in a multi-turn dialog setting. The task involves complex reasoning that requires operations like comparison and arithmetic reasoning over temporal expressions and the need for commonsense and world knowledge. ### Source Data #### Initial Data Collection and Normalization The TIMEDIAL dataset is derived from DailyDialog data (Li et al., 2017), which is a multi-turn dialog corpus containing over 13K English dialogs. Dialogs in this dataset consist of turn-taking between two people on topics over 10 broad categories, ranging from daily lives to financial topics. #### Who are the source language producers? ### Annotations #### Annotation process The data collection process involves two steps: (1) identifying dialogs that are rich in temporal expressions, and (2) asking human annotators to provide correct and incorrect options for cloze instances derived from these dialogs. More details about the two steps: 1. Temporal expression identification: Here, they select dialogs that are rich with temporal information, in order to focus on complex temporal reasoning that arises in natural dialogs. Temporal expressions are automatically identified with SU-Time, an off-the-shelf temporal expression detector. They keep only the dialogs with more than 3 temporal expressions and at least one expression that contains numerals like “two weeks” (as opposed to non-numeric spans, like “summer”, “right now”, and “later”). In their initial experiment, they observe that language models can often correctly predict these non-numerical temporal phrases. 2. Human annotated options: Next, they make spans in the dialogs. For a dialog, they mask out each temporal expression that contains numerals, each resulting in a cloze question that is then sent for human annotation. This resulted in 1,526 instances for annotation. For each masked span in each dialog, they obtain human annotation to derive a fixed set of correct and incorrect options given the context. Concretely, given a masked dialog and a seed correct answer (i.e., the original text) for the masked span, the annotators were asked to (1) come up with an alternative correct answer that makes sense in the dialog adhering to commonsense, and (2) formulate two incorrect answers that have no possibility of making sense in the dialog context. They highlight all time expressions in the context to make it easier for annotators to select reasonable time expressions. #### Who are the annotators? They are English linguists. ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Dataset provided for research purposes only. Please check dataset license for additional information. Additional Information ---------------------- ### Dataset Curators ### Licensing Information TimeDial dataset is licensed under CC BY-NC-SA 4.0. ### Contributions Thanks to @bhavitvyamalik for adding this dataset.
[ "### Dataset Summary\n\n\nTimeDial presents a crowdsourced English challenge set, for temporal commonsense reasoning, formulated as a multiple choice cloze task with around 1.5k carefully curated dialogs. The dataset is derived from the DailyDialog (Li et al., 2017), which is a multi-turn dialog corpus.\n\n\nIn order to establish strong baselines and provide information on future model development, the authors conducted extensive experiments with state-of-the-art LMs. While humans can easily answer these questions (97.8%), the best T5 model variant struggles on this challenge set (73%). Moreover, our qualitative error analyses show that the models often rely on shallow, spurious features (particularly text matching), instead of truly doing reasoning over the context.\n\n\nDetailed experiments and analyses can be found in their paper.", "### Supported Tasks and Leaderboards\n\n\nTo be updated soon.", "### Languages\n\n\nThe dataset is in English only.\n\n\nDataset Structure\n-----------------", "### Data Instances", "### Data Fields\n\n\n* \"id\": Unique identifier, as a integer\n* \"conversation\": Dialog context with span, as a string\n* \"correct1\": Original span, as a string\n* \"correct2\": Additional correct option provided by annotators, as a string\n* \"incorrect1\": Incorrect option #1 provided by annotators, as a string\n* \"incorrect1\\_rule\": One of phrase matching (\"Rule 1\"), numeral matching (\"Rule 2\"), or open ended (\"Rule 3\"), as a string\n* \"incorrect2\": Incorrect option #2 provided by annotators, as a string\n* \"incorrect2\\_rule\": One of phrase matching (\"Rule 1\"), numeral matching (\"Rule 2\"), or open ended (\"Rule 3\"), as a string", "### Data Splits\n\n\nTimeDial dataset consists only of a test set of 1,104 dialog instances with 2 correct and 2 incorrect options with the following statistics:\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nAlthough previous works have studied temporal reasoning in natural language, they have either focused on specific time-related concepts in isolation, such as temporal ordering and relation extraction, and/or dealt with limited context, such as single-sentence-based question answering and natural language inference.\n\n\nIn this work, they make the first systematic study of temporal commonsense reasoning in a multi-turn dialog setting. The task involves complex reasoning that requires operations like comparison and arithmetic reasoning over temporal expressions and the need for commonsense and world knowledge.", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nThe TIMEDIAL dataset is derived from DailyDialog data (Li et al., 2017), which is a multi-turn dialog corpus containing over 13K English dialogs. Dialogs in this dataset consist of turn-taking between two people on topics over 10 broad categories, ranging from daily lives to financial topics.", "#### Who are the source language producers?", "### Annotations", "#### Annotation process\n\n\nThe data collection process involves two steps: (1) identifying dialogs that are rich in temporal expressions, and (2) asking human annotators to provide correct and incorrect options for cloze instances derived from these dialogs. More details about the two steps:\n\n\n1. Temporal expression identification: Here, they select dialogs that are rich with temporal information, in order to focus on complex temporal reasoning that arises in natural dialogs. Temporal expressions are automatically identified with SU-Time, an off-the-shelf temporal expression detector. They keep only the dialogs with more than 3 temporal expressions and at least one expression that contains numerals like “two weeks” (as opposed to non-numeric spans, like “summer”, “right now”, and “later”). In their initial experiment, they observe that language models can often correctly predict these non-numerical temporal phrases.\n2. Human annotated options: Next, they make spans in the dialogs. For a dialog, they mask out each temporal expression that contains numerals, each resulting in a cloze question that is then sent for human annotation.\nThis resulted in 1,526 instances for annotation. For each masked span in each dialog, they obtain human annotation to derive a fixed set of correct and incorrect options given the context. Concretely, given a masked dialog and a seed correct answer (i.e., the original text) for the masked span, the annotators were asked to (1) come up with an alternative correct answer that makes sense in the dialog adhering to commonsense, and (2) formulate two incorrect answers that have no possibility of making sense in the dialog context. They highlight all time expressions in the context to make it easier for annotators to select reasonable time expressions.", "#### Who are the annotators?\n\n\nThey are English linguists.", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nDataset provided for research purposes only. Please check dataset license for additional information.\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nTimeDial dataset is licensed under CC BY-NC-SA 4.0.", "### Contributions\n\n\nThanks to @bhavitvyamalik for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #task_ids-multi-label-classification #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-nc-sa-4.0 #dialog-act-classification #arxiv-2106.04571 #region-us \n", "### Dataset Summary\n\n\nTimeDial presents a crowdsourced English challenge set, for temporal commonsense reasoning, formulated as a multiple choice cloze task with around 1.5k carefully curated dialogs. The dataset is derived from the DailyDialog (Li et al., 2017), which is a multi-turn dialog corpus.\n\n\nIn order to establish strong baselines and provide information on future model development, the authors conducted extensive experiments with state-of-the-art LMs. While humans can easily answer these questions (97.8%), the best T5 model variant struggles on this challenge set (73%). Moreover, our qualitative error analyses show that the models often rely on shallow, spurious features (particularly text matching), instead of truly doing reasoning over the context.\n\n\nDetailed experiments and analyses can be found in their paper.", "### Supported Tasks and Leaderboards\n\n\nTo be updated soon.", "### Languages\n\n\nThe dataset is in English only.\n\n\nDataset Structure\n-----------------", "### Data Instances", "### Data Fields\n\n\n* \"id\": Unique identifier, as a integer\n* \"conversation\": Dialog context with span, as a string\n* \"correct1\": Original span, as a string\n* \"correct2\": Additional correct option provided by annotators, as a string\n* \"incorrect1\": Incorrect option #1 provided by annotators, as a string\n* \"incorrect1\\_rule\": One of phrase matching (\"Rule 1\"), numeral matching (\"Rule 2\"), or open ended (\"Rule 3\"), as a string\n* \"incorrect2\": Incorrect option #2 provided by annotators, as a string\n* \"incorrect2\\_rule\": One of phrase matching (\"Rule 1\"), numeral matching (\"Rule 2\"), or open ended (\"Rule 3\"), as a string", "### Data Splits\n\n\nTimeDial dataset consists only of a test set of 1,104 dialog instances with 2 correct and 2 incorrect options with the following statistics:\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nAlthough previous works have studied temporal reasoning in natural language, they have either focused on specific time-related concepts in isolation, such as temporal ordering and relation extraction, and/or dealt with limited context, such as single-sentence-based question answering and natural language inference.\n\n\nIn this work, they make the first systematic study of temporal commonsense reasoning in a multi-turn dialog setting. The task involves complex reasoning that requires operations like comparison and arithmetic reasoning over temporal expressions and the need for commonsense and world knowledge.", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nThe TIMEDIAL dataset is derived from DailyDialog data (Li et al., 2017), which is a multi-turn dialog corpus containing over 13K English dialogs. Dialogs in this dataset consist of turn-taking between two people on topics over 10 broad categories, ranging from daily lives to financial topics.", "#### Who are the source language producers?", "### Annotations", "#### Annotation process\n\n\nThe data collection process involves two steps: (1) identifying dialogs that are rich in temporal expressions, and (2) asking human annotators to provide correct and incorrect options for cloze instances derived from these dialogs. More details about the two steps:\n\n\n1. Temporal expression identification: Here, they select dialogs that are rich with temporal information, in order to focus on complex temporal reasoning that arises in natural dialogs. Temporal expressions are automatically identified with SU-Time, an off-the-shelf temporal expression detector. They keep only the dialogs with more than 3 temporal expressions and at least one expression that contains numerals like “two weeks” (as opposed to non-numeric spans, like “summer”, “right now”, and “later”). In their initial experiment, they observe that language models can often correctly predict these non-numerical temporal phrases.\n2. Human annotated options: Next, they make spans in the dialogs. For a dialog, they mask out each temporal expression that contains numerals, each resulting in a cloze question that is then sent for human annotation.\nThis resulted in 1,526 instances for annotation. For each masked span in each dialog, they obtain human annotation to derive a fixed set of correct and incorrect options given the context. Concretely, given a masked dialog and a seed correct answer (i.e., the original text) for the masked span, the annotators were asked to (1) come up with an alternative correct answer that makes sense in the dialog adhering to commonsense, and (2) formulate two incorrect answers that have no possibility of making sense in the dialog context. They highlight all time expressions in the context to make it easier for annotators to select reasonable time expressions.", "#### Who are the annotators?\n\n\nThey are English linguists.", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nDataset provided for research purposes only. Please check dataset license for additional information.\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nTimeDial dataset is licensed under CC BY-NC-SA 4.0.", "### Contributions\n\n\nThanks to @bhavitvyamalik for adding this dataset." ]
[ 114, 193, 15, 19, 6, 197, 43, 130, 4, 84, 10, 5, 392, 15, 18, 7, 8, 32, 6, 23, 19 ]
[ "passage: TAGS\n#task_categories-text-classification #task_ids-multi-label-classification #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-nc-sa-4.0 #dialog-act-classification #arxiv-2106.04571 #region-us \n### Dataset Summary\n\n\nTimeDial presents a crowdsourced English challenge set, for temporal commonsense reasoning, formulated as a multiple choice cloze task with around 1.5k carefully curated dialogs. The dataset is derived from the DailyDialog (Li et al., 2017), which is a multi-turn dialog corpus.\n\n\nIn order to establish strong baselines and provide information on future model development, the authors conducted extensive experiments with state-of-the-art LMs. While humans can easily answer these questions (97.8%), the best T5 model variant struggles on this challenge set (73%). Moreover, our qualitative error analyses show that the models often rely on shallow, spurious features (particularly text matching), instead of truly doing reasoning over the context.\n\n\nDetailed experiments and analyses can be found in their paper.### Supported Tasks and Leaderboards\n\n\nTo be updated soon.### Languages\n\n\nThe dataset is in English only.\n\n\nDataset Structure\n-----------------### Data Instances", "passage: ### Data Fields\n\n\n* \"id\": Unique identifier, as a integer\n* \"conversation\": Dialog context with span, as a string\n* \"correct1\": Original span, as a string\n* \"correct2\": Additional correct option provided by annotators, as a string\n* \"incorrect1\": Incorrect option #1 provided by annotators, as a string\n* \"incorrect1\\_rule\": One of phrase matching (\"Rule 1\"), numeral matching (\"Rule 2\"), or open ended (\"Rule 3\"), as a string\n* \"incorrect2\": Incorrect option #2 provided by annotators, as a string\n* \"incorrect2\\_rule\": One of phrase matching (\"Rule 1\"), numeral matching (\"Rule 2\"), or open ended (\"Rule 3\"), as a string### Data Splits\n\n\nTimeDial dataset consists only of a test set of 1,104 dialog instances with 2 correct and 2 incorrect options with the following statistics:\n\n\n\nDataset Creation\n----------------### Curation Rationale\n\n\nAlthough previous works have studied temporal reasoning in natural language, they have either focused on specific time-related concepts in isolation, such as temporal ordering and relation extraction, and/or dealt with limited context, such as single-sentence-based question answering and natural language inference.\n\n\nIn this work, they make the first systematic study of temporal commonsense reasoning in a multi-turn dialog setting. The task involves complex reasoning that requires operations like comparison and arithmetic reasoning over temporal expressions and the need for commonsense and world knowledge.### Source Data#### Initial Data Collection and Normalization\n\n\nThe TIMEDIAL dataset is derived from DailyDialog data (Li et al., 2017), which is a multi-turn dialog corpus containing over 13K English dialogs. Dialogs in this dataset consist of turn-taking between two people on topics over 10 broad categories, ranging from daily lives to financial topics.#### Who are the source language producers?### Annotations" ]
245d40b2520a2a2210a71930705816e720370889
# Dataset Card for Times of India News Headlines ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://dataverse.harvard.edu/dataset.xhtml?persistentId=doi:10.7910/DVN/J7BYRX - **Repository:** [More Information Needed] - **Paper:** [More Information Needed] - **Leaderboard:** [More Information Needed] - **Point of Contact:** [More Information Needed] ### Dataset Summary This news dataset is a persistent historical archive of noteable events in the Indian subcontinent from start-2001 to mid-2020, recorded in realtime by the journalists of India. It contains approximately 3.3 million events published by Times of India. Times Group as a news agency, reaches out a very wide audience across Asia and drawfs every other agency in the quantity of english articles published per day. Due to the heavy daily volume over multiple years, this data offers a deep insight into Indian society, its priorities, events, issues and talking points and how they have unfolded over time. It is possible to chop this dataset into a smaller piece for a more focused analysis, based on one or more facets. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The text in the dataset is in English. ## Dataset Structure ### Data Instances ``` { 'publish_date': '20010530', 'headline_category': city.kolkata, 'headline_text': "Malda fake notes" } ``` ### Data Fields - `publish_date`: Date of publishing in yyyyMMdd format - `headline_category`: Category of event in ascii, dot-delimited values - `headline_text`: Headline of article en la Engrezi (2020-07-10) ### Data Splits This dataset has no splits. ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators The dataset was created by Rohit Kulkarni. ### Licensing Information The data is under the [CC0: Public Domain](https://creativecommons.org/publicdomain/zero/1.0/) ### Citation Information ``` @data{DVN/DPQMQH_2020, author = {Kulkarni, Rohit}, publisher = {Harvard Dataverse}, title = {{Times of India News Headlines}}, year = {2020}, version = {V1}, doi = {10.7910/DVN/DPQMQH}, url = {https://doi.org/10.7910/DVN/DPQMQH} } ``` ### Contributions Thanks to [@tanmoyio](https://github.com/tanmoyio) for adding this dataset.
times_of_india_news_headlines
[ "task_categories:text2text-generation", "task_categories:text-retrieval", "task_ids:document-retrieval", "task_ids:fact-checking-retrieval", "task_ids:text-simplification", "annotations_creators:no-annotation", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:1M<n<10M", "source_datasets:original", "language:en", "license:cc0-1.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["expert-generated"], "language": ["en"], "license": ["cc0-1.0"], "multilinguality": ["monolingual"], "size_categories": ["1M<n<10M"], "source_datasets": ["original"], "task_categories": ["text2text-generation", "text-retrieval"], "task_ids": ["document-retrieval", "fact-checking-retrieval", "text-simplification"], "pretty_name": "Times of India News Headlines", "dataset_info": {"features": [{"name": "publish_date", "dtype": "string"}, {"name": "headline_category", "dtype": "string"}, {"name": "headline_text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 260939306, "num_examples": 3297173}], "download_size": 0, "dataset_size": 260939306}}
2024-01-18T11:17:12+00:00
[]
[ "en" ]
TAGS #task_categories-text2text-generation #task_categories-text-retrieval #task_ids-document-retrieval #task_ids-fact-checking-retrieval #task_ids-text-simplification #annotations_creators-no-annotation #language_creators-expert-generated #multilinguality-monolingual #size_categories-1M<n<10M #source_datasets-original #language-English #license-cc0-1.0 #region-us
# Dataset Card for Times of India News Headlines ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary This news dataset is a persistent historical archive of noteable events in the Indian subcontinent from start-2001 to mid-2020, recorded in realtime by the journalists of India. It contains approximately 3.3 million events published by Times of India. Times Group as a news agency, reaches out a very wide audience across Asia and drawfs every other agency in the quantity of english articles published per day. Due to the heavy daily volume over multiple years, this data offers a deep insight into Indian society, its priorities, events, issues and talking points and how they have unfolded over time. It is possible to chop this dataset into a smaller piece for a more focused analysis, based on one or more facets. ### Supported Tasks and Leaderboards ### Languages The text in the dataset is in English. ## Dataset Structure ### Data Instances ### Data Fields - 'publish_date': Date of publishing in yyyyMMdd format - 'headline_category': Category of event in ascii, dot-delimited values - 'headline_text': Headline of article en la Engrezi (2020-07-10) ### Data Splits This dataset has no splits. ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators The dataset was created by Rohit Kulkarni. ### Licensing Information The data is under the CC0: Public Domain ### Contributions Thanks to @tanmoyio for adding this dataset.
[ "# Dataset Card for Times of India News Headlines", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: \n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\nThis news dataset is a persistent historical archive of noteable events in the Indian subcontinent from start-2001 to mid-2020, recorded in realtime by the journalists of India. It contains approximately 3.3 million events published by Times of India. Times Group as a news agency, reaches out a very wide audience across Asia and drawfs every other agency in the quantity of english articles published per day. Due to the heavy daily volume over multiple years, this data offers a deep insight into Indian society, its priorities, events, issues and talking points and how they have unfolded over time. It is possible to chop this dataset into a smaller piece for a more focused analysis, based on one or more facets.", "### Supported Tasks and Leaderboards", "### Languages\n\nThe text in the dataset is in English.", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- 'publish_date': Date of publishing in yyyyMMdd format\n- 'headline_category': Category of event in ascii, dot-delimited values\n- 'headline_text': Headline of article en la Engrezi (2020-07-10)", "### Data Splits\n\nThis dataset has no splits.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThe dataset was created by Rohit Kulkarni.", "### Licensing Information\n\nThe data is under the CC0: Public Domain", "### Contributions\n\nThanks to @tanmoyio for adding this dataset." ]
[ "TAGS\n#task_categories-text2text-generation #task_categories-text-retrieval #task_ids-document-retrieval #task_ids-fact-checking-retrieval #task_ids-text-simplification #annotations_creators-no-annotation #language_creators-expert-generated #multilinguality-monolingual #size_categories-1M<n<10M #source_datasets-original #language-English #license-cc0-1.0 #region-us \n", "# Dataset Card for Times of India News Headlines", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: \n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\nThis news dataset is a persistent historical archive of noteable events in the Indian subcontinent from start-2001 to mid-2020, recorded in realtime by the journalists of India. It contains approximately 3.3 million events published by Times of India. Times Group as a news agency, reaches out a very wide audience across Asia and drawfs every other agency in the quantity of english articles published per day. Due to the heavy daily volume over multiple years, this data offers a deep insight into Indian society, its priorities, events, issues and talking points and how they have unfolded over time. It is possible to chop this dataset into a smaller piece for a more focused analysis, based on one or more facets.", "### Supported Tasks and Leaderboards", "### Languages\n\nThe text in the dataset is in English.", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- 'publish_date': Date of publishing in yyyyMMdd format\n- 'headline_category': Category of event in ascii, dot-delimited values\n- 'headline_text': Headline of article en la Engrezi (2020-07-10)", "### Data Splits\n\nThis dataset has no splits.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThe dataset was created by Rohit Kulkarni.", "### Licensing Information\n\nThe data is under the CC0: Public Domain", "### Contributions\n\nThanks to @tanmoyio for adding this dataset." ]
[ 131, 11, 120, 25, 160, 10, 14, 6, 6, 67, 13, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 18, 16, 18 ]
[ "passage: TAGS\n#task_categories-text2text-generation #task_categories-text-retrieval #task_ids-document-retrieval #task_ids-fact-checking-retrieval #task_ids-text-simplification #annotations_creators-no-annotation #language_creators-expert-generated #multilinguality-monolingual #size_categories-1M<n<10M #source_datasets-original #language-English #license-cc0-1.0 #region-us \n# Dataset Card for Times of India News Headlines## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: \n- Leaderboard: \n- Point of Contact:### Dataset Summary\n\nThis news dataset is a persistent historical archive of noteable events in the Indian subcontinent from start-2001 to mid-2020, recorded in realtime by the journalists of India. It contains approximately 3.3 million events published by Times of India. Times Group as a news agency, reaches out a very wide audience across Asia and drawfs every other agency in the quantity of english articles published per day. Due to the heavy daily volume over multiple years, this data offers a deep insight into Indian society, its priorities, events, issues and talking points and how they have unfolded over time. It is possible to chop this dataset into a smaller piece for a more focused analysis, based on one or more facets.### Supported Tasks and Leaderboards### Languages\n\nThe text in the dataset is in English.## Dataset Structure### Data Instances" ]
1d0cd09f9ca7c40158e7e5377f45c9c718e53c68
# Dataset Card for timit_asr ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [TIMIT Acoustic-Phonetic Continuous Speech Corpus](https://catalog.ldc.upenn.edu/LDC93S1) - **Repository:** [Needs More Information] - **Paper:** [TIMIT: Dataset designed to provide speech data for acoustic-phonetic studies and for the development and evaluation of automatic speech recognition systems.](https://catalog.ldc.upenn.edu/LDC93S1) - **Leaderboard:** [Paperswithcode Leaderboard](https://paperswithcode.com/sota/speech-recognition-on-timit) - **Point of Contact:** [Needs More Information] ### Dataset Summary The TIMIT corpus of read speech is designed to provide speech data for acoustic-phonetic studies and for the development and evaluation of automatic speech recognition systems. TIMIT contains broadband recordings of 630 speakers of eight major dialects of American English, each reading ten phonetically rich sentences. The TIMIT corpus includes time-aligned orthographic, phonetic and word transcriptions as well as a 16-bit, 16kHz speech waveform file for each utterance. Corpus design was a joint effort among the Massachusetts Institute of Technology (MIT), SRI International (SRI) and Texas Instruments, Inc. (TI). The speech was recorded at TI, transcribed at MIT and verified and prepared for CD-ROM production by the National Institute of Standards and Technology (NIST). The dataset needs to be downloaded manually from https://catalog.ldc.upenn.edu/LDC93S1: ``` To use TIMIT you have to download it manually. Please create an account and download the dataset from https://catalog.ldc.upenn.edu/LDC93S1 Then extract all files in one folder and load the dataset with: `datasets.load_dataset('timit_asr', data_dir='path/to/folder/folder_name')` ``` ### Supported Tasks and Leaderboards - `automatic-speech-recognition`, `speaker-identification`: The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). The task has an active leaderboard which can be found at https://paperswithcode.com/sota/speech-recognition-on-timit and ranks models based on their WER. ### Languages The audio is in English. The TIMIT corpus transcriptions have been hand verified. Test and training subsets, balanced for phonetic and dialectal coverage, are specified. Tabular computer-searchable information is included as well as written documentation. ## Dataset Structure ### Data Instances A typical data point comprises the path to the audio file, usually called `file` and its transcription, called `text`. Some additional information about the speaker and the passage which contains the transcription is provided. ``` { 'file': '/data/TRAIN/DR4/MMDM0/SI681.WAV', 'audio': {'path': '/data/TRAIN/DR4/MMDM0/SI681.WAV', 'array': array([-0.00048828, -0.00018311, -0.00137329, ..., 0.00079346, 0.00091553, 0.00085449], dtype=float32), 'sampling_rate': 16000}, 'text': 'Would such an act of refusal be useful?', 'phonetic_detail': [{'start': '0', 'stop': '1960', 'utterance': 'h#'}, {'start': '1960', 'stop': '2466', 'utterance': 'w'}, {'start': '2466', 'stop': '3480', 'utterance': 'ix'}, {'start': '3480', 'stop': '4000', 'utterance': 'dcl'}, {'start': '4000', 'stop': '5960', 'utterance': 's'}, {'start': '5960', 'stop': '7480', 'utterance': 'ah'}, {'start': '7480', 'stop': '7880', 'utterance': 'tcl'}, {'start': '7880', 'stop': '9400', 'utterance': 'ch'}, {'start': '9400', 'stop': '9960', 'utterance': 'ix'}, {'start': '9960', 'stop': '10680', 'utterance': 'n'}, {'start': '10680', 'stop': '13480', 'utterance': 'ae'}, {'start': '13480', 'stop': '15680', 'utterance': 'kcl'}, {'start': '15680', 'stop': '15880', 'utterance': 't'}, {'start': '15880', 'stop': '16920', 'utterance': 'ix'}, {'start': '16920', 'stop': '18297', 'utterance': 'v'}, {'start': '18297', 'stop': '18882', 'utterance': 'r'}, {'start': '18882', 'stop': '19480', 'utterance': 'ix'}, {'start': '19480', 'stop': '21723', 'utterance': 'f'}, {'start': '21723', 'stop': '22516', 'utterance': 'y'}, {'start': '22516', 'stop': '24040', 'utterance': 'ux'}, {'start': '24040', 'stop': '25190', 'utterance': 'zh'}, {'start': '25190', 'stop': '27080', 'utterance': 'el'}, {'start': '27080', 'stop': '28160', 'utterance': 'bcl'}, {'start': '28160', 'stop': '28560', 'utterance': 'b'}, {'start': '28560', 'stop': '30120', 'utterance': 'iy'}, {'start': '30120', 'stop': '31832', 'utterance': 'y'}, {'start': '31832', 'stop': '33240', 'utterance': 'ux'}, {'start': '33240', 'stop': '34640', 'utterance': 's'}, {'start': '34640', 'stop': '35968', 'utterance': 'f'}, {'start': '35968', 'stop': '37720', 'utterance': 'el'}, {'start': '37720', 'stop': '39920', 'utterance': 'h#'}], 'word_detail': [{'start': '1960', 'stop': '4000', 'utterance': 'would'}, {'start': '4000', 'stop': '9400', 'utterance': 'such'}, {'start': '9400', 'stop': '10680', 'utterance': 'an'}, {'start': '10680', 'stop': '15880', 'utterance': 'act'}, {'start': '15880', 'stop': '18297', 'utterance': 'of'}, {'start': '18297', 'stop': '27080', 'utterance': 'refusal'}, {'start': '27080', 'stop': '30120', 'utterance': 'be'}, {'start': '30120', 'stop': '37720', 'utterance': 'useful'}], 'dialect_region': 'DR4', 'sentence_type': 'SI', 'speaker_id': 'MMDM0', 'id': 'SI681' } ``` ### Data Fields - file: A path to the downloaded audio file in .wav format. - audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: `dataset[0]["audio"]` the audio file is automatically decoded and resampled to `dataset.features["audio"].sampling_rate`. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the `"audio"` column, *i.e.* `dataset[0]["audio"]` should **always** be preferred over `dataset["audio"][0]`. - text: The transcription of the audio file. - phonetic_detail: The phonemes that make up the sentence. The PHONCODE.DOC contains a table of all the phonemic and phonetic symbols used in TIMIT lexicon. - word_detail: Word level split of the transcript. - dialect_region: The dialect code of the recording. - sentence_type: The type of the sentence - 'SA':'Dialect', 'SX':'Compact' or 'SI':'Diverse'. - speaker_id: Unique id of the speaker. The same speaker id can be found for multiple data samples. - id: ID of the data sample. Contains the <SENTENCE_TYPE><SENTENCE_NUMBER>. ### Data Splits The speech material has been subdivided into portions for training and testing. The default train-test split will be made available on data download. The test data alone has a core portion containing 24 speakers, 2 male and 1 female from each dialect region. More information about the test set can be found [here](https://catalog.ldc.upenn.edu/docs/LDC93S1/TESTSET.TXT) ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in this dataset. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations Dataset provided for research purposes only. Please check dataset license for additional information. ## Additional Information ### Dataset Curators The dataset was created by John S. Garofolo, Lori F. Lamel, William M. Fisher, Jonathan G. Fiscus, David S. Pallett, Nancy L. Dahlgren, Victor Zue ### Licensing Information [LDC User Agreement for Non-Members](https://catalog.ldc.upenn.edu/license/ldc-non-members-agreement.pdf) ### Citation Information ``` @inproceedings{ title={TIMIT Acoustic-Phonetic Continuous Speech Corpus}, author={Garofolo, John S., et al}, ldc_catalog_no={LDC93S1}, DOI={https://doi.org/10.35111/17gk-bn40}, journal={Linguistic Data Consortium, Philadelphia}, year={1983} } ``` ### Contributions Thanks to [@vrindaprabhu](https://github.com/vrindaprabhu) for adding this dataset.
timit_asr
[ "task_categories:automatic-speech-recognition", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:other", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["expert-generated"], "language": ["en"], "license": ["other"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["automatic-speech-recognition"], "task_ids": [], "paperswithcode_id": "timit", "pretty_name": "TIMIT", "license_details": "LDC-User-Agreement-for-Non-Members", "train-eval-index": [{"config": "clean", "task": "automatic-speech-recognition", "task_id": "speech_recognition", "splits": {"train_split": "train", "eval_split": "test"}, "col_mapping": {"file": "path", "text": "text"}, "metrics": [{"type": "wer", "name": "WER"}, {"type": "cer", "name": "CER"}]}]}
2022-10-28T15:41:41+00:00
[]
[ "en" ]
TAGS #task_categories-automatic-speech-recognition #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-other #region-us
# Dataset Card for timit_asr ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: TIMIT Acoustic-Phonetic Continuous Speech Corpus - Repository: - Paper: TIMIT: Dataset designed to provide speech data for acoustic-phonetic studies and for the development and evaluation of automatic speech recognition systems. - Leaderboard: Paperswithcode Leaderboard - Point of Contact: ### Dataset Summary The TIMIT corpus of read speech is designed to provide speech data for acoustic-phonetic studies and for the development and evaluation of automatic speech recognition systems. TIMIT contains broadband recordings of 630 speakers of eight major dialects of American English, each reading ten phonetically rich sentences. The TIMIT corpus includes time-aligned orthographic, phonetic and word transcriptions as well as a 16-bit, 16kHz speech waveform file for each utterance. Corpus design was a joint effort among the Massachusetts Institute of Technology (MIT), SRI International (SRI) and Texas Instruments, Inc. (TI). The speech was recorded at TI, transcribed at MIT and verified and prepared for CD-ROM production by the National Institute of Standards and Technology (NIST). The dataset needs to be downloaded manually from URL ### Supported Tasks and Leaderboards - 'automatic-speech-recognition', 'speaker-identification': The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). The task has an active leaderboard which can be found at URL and ranks models based on their WER. ### Languages The audio is in English. The TIMIT corpus transcriptions have been hand verified. Test and training subsets, balanced for phonetic and dialectal coverage, are specified. Tabular computer-searchable information is included as well as written documentation. ## Dataset Structure ### Data Instances A typical data point comprises the path to the audio file, usually called 'file' and its transcription, called 'text'. Some additional information about the speaker and the passage which contains the transcription is provided. ### Data Fields - file: A path to the downloaded audio file in .wav format. - audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: 'dataset[0]["audio"]' the audio file is automatically decoded and resampled to 'dataset.features["audio"].sampling_rate'. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the '"audio"' column, *i.e.* 'dataset[0]["audio"]' should always be preferred over 'dataset["audio"][0]'. - text: The transcription of the audio file. - phonetic_detail: The phonemes that make up the sentence. The PHONCODE.DOC contains a table of all the phonemic and phonetic symbols used in TIMIT lexicon. - word_detail: Word level split of the transcript. - dialect_region: The dialect code of the recording. - sentence_type: The type of the sentence - 'SA':'Dialect', 'SX':'Compact' or 'SI':'Diverse'. - speaker_id: Unique id of the speaker. The same speaker id can be found for multiple data samples. - id: ID of the data sample. Contains the <SENTENCE_TYPE><SENTENCE_NUMBER>. ### Data Splits The speech material has been subdivided into portions for training and testing. The default train-test split will be made available on data download. The test data alone has a core portion containing 24 speakers, 2 male and 1 female from each dialect region. More information about the test set can be found here ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in this dataset. ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Dataset provided for research purposes only. Please check dataset license for additional information. ## Additional Information ### Dataset Curators The dataset was created by John S. Garofolo, Lori F. Lamel, William M. Fisher, Jonathan G. Fiscus, David S. Pallett, Nancy L. Dahlgren, Victor Zue ### Licensing Information LDC User Agreement for Non-Members ### Contributions Thanks to @vrindaprabhu for adding this dataset.
[ "# Dataset Card for timit_asr", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: TIMIT Acoustic-Phonetic Continuous Speech Corpus\n- Repository: \n- Paper: TIMIT: Dataset designed to provide speech data for acoustic-phonetic studies and for the development and evaluation of automatic speech recognition systems.\n- Leaderboard: Paperswithcode Leaderboard\n- Point of Contact:", "### Dataset Summary\n\nThe TIMIT corpus of read speech is designed to provide speech data for acoustic-phonetic studies and for the development and evaluation of automatic speech recognition systems. TIMIT contains broadband recordings of 630 speakers of eight major dialects of American English, each reading ten phonetically rich sentences. The TIMIT corpus includes time-aligned orthographic, phonetic and word transcriptions as well as a 16-bit, 16kHz speech waveform file for each utterance. Corpus design was a joint effort among the Massachusetts Institute of Technology (MIT), SRI International (SRI) and Texas Instruments, Inc. (TI). The speech was recorded at TI, transcribed at MIT and verified and prepared for CD-ROM production by the National Institute of Standards and Technology (NIST).\n\nThe dataset needs to be downloaded manually from URL", "### Supported Tasks and Leaderboards\n\n- 'automatic-speech-recognition', 'speaker-identification': The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). The task has an active leaderboard which can be found at URL and ranks models based on their WER.", "### Languages\n\nThe audio is in English.\nThe TIMIT corpus transcriptions have been hand verified. Test and training subsets, balanced for phonetic and dialectal coverage, are specified. Tabular computer-searchable information is included as well as written documentation.", "## Dataset Structure", "### Data Instances\n\nA typical data point comprises the path to the audio file, usually called 'file' and its transcription, called 'text'. Some additional information about the speaker and the passage which contains the transcription is provided.", "### Data Fields\n\n- file: A path to the downloaded audio file in .wav format.\n\n- audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: 'dataset[0][\"audio\"]' the audio file is automatically decoded and resampled to 'dataset.features[\"audio\"].sampling_rate'. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the '\"audio\"' column, *i.e.* 'dataset[0][\"audio\"]' should always be preferred over 'dataset[\"audio\"][0]'.\n\n- text: The transcription of the audio file.\n\n- phonetic_detail: The phonemes that make up the sentence. The PHONCODE.DOC contains a table of all the phonemic and phonetic symbols used in TIMIT lexicon.\n\n- word_detail: Word level split of the transcript.\n\n- dialect_region: The dialect code of the recording.\n\n- sentence_type: The type of the sentence - 'SA':'Dialect', 'SX':'Compact' or 'SI':'Diverse'.\n\n- speaker_id: Unique id of the speaker. The same speaker id can be found for multiple data samples.\n\n- id: ID of the data sample. Contains the <SENTENCE_TYPE><SENTENCE_NUMBER>.", "### Data Splits\n\nThe speech material has been subdivided into portions for training and\ntesting. The default train-test split will be made available on data download.\n\nThe test data alone has a core portion containing 24 speakers, 2 male and 1 female\nfrom each dialect region. More information about the test set can\nbe found here", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\nThe dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in this dataset.", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\nDataset provided for research purposes only. Please check dataset license for additional information.", "## Additional Information", "### Dataset Curators\n\nThe dataset was created by John S. Garofolo, Lori F. Lamel, William M. Fisher, Jonathan G. Fiscus, David S. Pallett, Nancy L. Dahlgren, Victor Zue", "### Licensing Information\n\nLDC User Agreement for Non-Members", "### Contributions\nThanks to @vrindaprabhu for adding this dataset." ]
[ "TAGS\n#task_categories-automatic-speech-recognition #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-other #region-us \n", "# Dataset Card for timit_asr", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: TIMIT Acoustic-Phonetic Continuous Speech Corpus\n- Repository: \n- Paper: TIMIT: Dataset designed to provide speech data for acoustic-phonetic studies and for the development and evaluation of automatic speech recognition systems.\n- Leaderboard: Paperswithcode Leaderboard\n- Point of Contact:", "### Dataset Summary\n\nThe TIMIT corpus of read speech is designed to provide speech data for acoustic-phonetic studies and for the development and evaluation of automatic speech recognition systems. TIMIT contains broadband recordings of 630 speakers of eight major dialects of American English, each reading ten phonetically rich sentences. The TIMIT corpus includes time-aligned orthographic, phonetic and word transcriptions as well as a 16-bit, 16kHz speech waveform file for each utterance. Corpus design was a joint effort among the Massachusetts Institute of Technology (MIT), SRI International (SRI) and Texas Instruments, Inc. (TI). The speech was recorded at TI, transcribed at MIT and verified and prepared for CD-ROM production by the National Institute of Standards and Technology (NIST).\n\nThe dataset needs to be downloaded manually from URL", "### Supported Tasks and Leaderboards\n\n- 'automatic-speech-recognition', 'speaker-identification': The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). The task has an active leaderboard which can be found at URL and ranks models based on their WER.", "### Languages\n\nThe audio is in English.\nThe TIMIT corpus transcriptions have been hand verified. Test and training subsets, balanced for phonetic and dialectal coverage, are specified. Tabular computer-searchable information is included as well as written documentation.", "## Dataset Structure", "### Data Instances\n\nA typical data point comprises the path to the audio file, usually called 'file' and its transcription, called 'text'. Some additional information about the speaker and the passage which contains the transcription is provided.", "### Data Fields\n\n- file: A path to the downloaded audio file in .wav format.\n\n- audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: 'dataset[0][\"audio\"]' the audio file is automatically decoded and resampled to 'dataset.features[\"audio\"].sampling_rate'. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the '\"audio\"' column, *i.e.* 'dataset[0][\"audio\"]' should always be preferred over 'dataset[\"audio\"][0]'.\n\n- text: The transcription of the audio file.\n\n- phonetic_detail: The phonemes that make up the sentence. The PHONCODE.DOC contains a table of all the phonemic and phonetic symbols used in TIMIT lexicon.\n\n- word_detail: Word level split of the transcript.\n\n- dialect_region: The dialect code of the recording.\n\n- sentence_type: The type of the sentence - 'SA':'Dialect', 'SX':'Compact' or 'SI':'Diverse'.\n\n- speaker_id: Unique id of the speaker. The same speaker id can be found for multiple data samples.\n\n- id: ID of the data sample. Contains the <SENTENCE_TYPE><SENTENCE_NUMBER>.", "### Data Splits\n\nThe speech material has been subdivided into portions for training and\ntesting. The default train-test split will be made available on data download.\n\nThe test data alone has a core portion containing 24 speakers, 2 male and 1 female\nfrom each dialect region. More information about the test set can\nbe found here", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\nThe dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in this dataset.", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\nDataset provided for research purposes only. Please check dataset license for additional information.", "## Additional Information", "### Dataset Curators\n\nThe dataset was created by John S. Garofolo, Lori F. Lamel, William M. Fisher, Jonathan G. Fiscus, David S. Pallett, Nancy L. Dahlgren, Victor Zue", "### Licensing Information\n\nLDC User Agreement for Non-Members", "### Contributions\nThanks to @vrindaprabhu for adding this dataset." ]
[ 83, 10, 120, 73, 190, 110, 59, 6, 52, 368, 69, 5, 7, 4, 10, 10, 5, 5, 9, 40, 8, 7, 8, 25, 5, 52, 16, 18 ]
[ "passage: TAGS\n#task_categories-automatic-speech-recognition #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-other #region-us \n# Dataset Card for timit_asr## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: TIMIT Acoustic-Phonetic Continuous Speech Corpus\n- Repository: \n- Paper: TIMIT: Dataset designed to provide speech data for acoustic-phonetic studies and for the development and evaluation of automatic speech recognition systems.\n- Leaderboard: Paperswithcode Leaderboard\n- Point of Contact:### Dataset Summary\n\nThe TIMIT corpus of read speech is designed to provide speech data for acoustic-phonetic studies and for the development and evaluation of automatic speech recognition systems. TIMIT contains broadband recordings of 630 speakers of eight major dialects of American English, each reading ten phonetically rich sentences. The TIMIT corpus includes time-aligned orthographic, phonetic and word transcriptions as well as a 16-bit, 16kHz speech waveform file for each utterance. Corpus design was a joint effort among the Massachusetts Institute of Technology (MIT), SRI International (SRI) and Texas Instruments, Inc. (TI). The speech was recorded at TI, transcribed at MIT and verified and prepared for CD-ROM production by the National Institute of Standards and Technology (NIST).\n\nThe dataset needs to be downloaded manually from URL", "passage: ### Supported Tasks and Leaderboards\n\n- 'automatic-speech-recognition', 'speaker-identification': The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). The task has an active leaderboard which can be found at URL and ranks models based on their WER.### Languages\n\nThe audio is in English.\nThe TIMIT corpus transcriptions have been hand verified. Test and training subsets, balanced for phonetic and dialectal coverage, are specified. Tabular computer-searchable information is included as well as written documentation.## Dataset Structure### Data Instances\n\nA typical data point comprises the path to the audio file, usually called 'file' and its transcription, called 'text'. Some additional information about the speaker and the passage which contains the transcription is provided.### Data Fields\n\n- file: A path to the downloaded audio file in .wav format.\n\n- audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: 'dataset[0][\"audio\"]' the audio file is automatically decoded and resampled to 'dataset.features[\"audio\"].sampling_rate'. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the '\"audio\"' column, *i.e.* 'dataset[0][\"audio\"]' should always be preferred over 'dataset[\"audio\"][0]'.\n\n- text: The transcription of the audio file.\n\n- phonetic_detail: The phonemes that make up the sentence. The PHONCODE.DOC contains a table of all the phonemic and phonetic symbols used in TIMIT lexicon.\n\n- word_detail: Word level split of the transcript.\n\n- dialect_region: The dialect code of the recording.\n\n- sentence_type: The type of the sentence - 'SA':'Dialect', 'SX':'Compact' or 'SI':'Diverse'.\n\n- speaker_id: Unique id of the speaker. The same speaker id can be found for multiple data samples.\n\n- id: ID of the data sample. Contains the <SENTENCE_TYPE><SENTENCE_NUMBER>." ]
c7a7ff3e41cda4f190ec575d180e764eb7f7f4ba
# Dataset Card for "tiny_shakespeare" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://github.com/karpathy/char-rnn/blob/master/data/tinyshakespeare/input.txt](https://github.com/karpathy/char-rnn/blob/master/data/tinyshakespeare/input.txt) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 1.11 MB - **Size of the generated dataset:** 1.11 MB - **Total amount of disk used:** 2.23 MB ### Dataset Summary 40,000 lines of Shakespeare from a variety of Shakespeare's plays. Featured in Andrej Karpathy's blog post 'The Unreasonable Effectiveness of Recurrent Neural Networks': http://karpathy.github.io/2015/05/21/rnn-effectiveness/. To use for e.g. character modelling: ``` d = datasets.load_dataset(name='tiny_shakespeare')['train'] d = d.map(lambda x: datasets.Value('strings').unicode_split(x['text'], 'UTF-8')) # train split includes vocabulary for other splits vocabulary = sorted(set(next(iter(d)).numpy())) d = d.map(lambda x: {'cur_char': x[:-1], 'next_char': x[1:]}) d = d.unbatch() seq_len = 100 batch_size = 2 d = d.batch(seq_len) d = d.batch(batch_size) ``` ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### default - **Size of downloaded dataset files:** 1.11 MB - **Size of the generated dataset:** 1.11 MB - **Total amount of disk used:** 2.23 MB An example of 'train' looks as follows. ``` { "text": "First Citizen:\nBefore we proceed any further, hear me " } ``` ### Data Fields The data fields are the same among all splits. #### default - `text`: a `string` feature. ### Data Splits | name |train|validation|test| |-------|----:|---------:|---:| |default| 1| 1| 1| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @misc{ author={Karpathy, Andrej}, title={char-rnn}, year={2015}, howpublished={\url{https://github.com/karpathy/char-rnn}} } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@lewtun](https://github.com/lewtun), [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset.
tiny_shakespeare
[ "region:us" ]
2022-03-02T23:29:22+00:00
{"pretty_name": "TinyShakespeare", "dataset_info": {"features": [{"name": "text", "dtype": "string"}], "splits": [{"name": "test", "num_bytes": 55780, "num_examples": 1}, {"name": "train", "num_bytes": 1003864, "num_examples": 1}, {"name": "validation", "num_bytes": 55780, "num_examples": 1}], "download_size": 1115394, "dataset_size": 1115424}}
2024-01-18T11:17:14+00:00
[]
[]
TAGS #region-us
Dataset Card for "tiny\_shakespeare" ==================================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: * Point of Contact: * Size of downloaded dataset files: 1.11 MB * Size of the generated dataset: 1.11 MB * Total amount of disk used: 2.23 MB ### Dataset Summary 40,000 lines of Shakespeare from a variety of Shakespeare's plays. Featured in Andrej Karpathy's blog post 'The Unreasonable Effectiveness of Recurrent Neural Networks': URL To use for e.g. character modelling: ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- ### Data Instances #### default * Size of downloaded dataset files: 1.11 MB * Size of the generated dataset: 1.11 MB * Total amount of disk used: 2.23 MB An example of 'train' looks as follows. ### Data Fields The data fields are the same among all splits. #### default * 'text': a 'string' feature. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @thomwolf, @lewtun, @patrickvonplaten for adding this dataset.
[ "### Dataset Summary\n\n\n40,000 lines of Shakespeare from a variety of Shakespeare's plays. Featured in Andrej Karpathy's blog post 'The Unreasonable Effectiveness of Recurrent Neural Networks': URL\n\n\nTo use for e.g. character modelling:", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### default\n\n\n* Size of downloaded dataset files: 1.11 MB\n* Size of the generated dataset: 1.11 MB\n* Total amount of disk used: 2.23 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### default\n\n\n* 'text': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @thomwolf, @lewtun, @patrickvonplaten for adding this dataset." ]
[ "TAGS\n#region-us \n", "### Dataset Summary\n\n\n40,000 lines of Shakespeare from a variety of Shakespeare's plays. Featured in Andrej Karpathy's blog post 'The Unreasonable Effectiveness of Recurrent Neural Networks': URL\n\n\nTo use for e.g. character modelling:", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### default\n\n\n* Size of downloaded dataset files: 1.11 MB\n* Size of the generated dataset: 1.11 MB\n* Total amount of disk used: 2.23 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### default\n\n\n* 'text': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @thomwolf, @lewtun, @patrickvonplaten for adding this dataset." ]
[ 6, 60, 10, 11, 6, 49, 17, 14, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 6, 28 ]
[ "passage: TAGS\n#region-us \n### Dataset Summary\n\n\n40,000 lines of Shakespeare from a variety of Shakespeare's plays. Featured in Andrej Karpathy's blog post 'The Unreasonable Effectiveness of Recurrent Neural Networks': URL\n\n\nTo use for e.g. character modelling:### Supported Tasks and Leaderboards### Languages\n\n\nDataset Structure\n-----------------### Data Instances#### default\n\n\n* Size of downloaded dataset files: 1.11 MB\n* Size of the generated dataset: 1.11 MB\n* Total amount of disk used: 2.23 MB\n\n\nAn example of 'train' looks as follows.### Data Fields\n\n\nThe data fields are the same among all splits.#### default\n\n\n* 'text': a 'string' feature.### Data Splits\n\n\n\nDataset Creation\n----------------### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------### Social Impact of Dataset### Discussion of Biases### Other Known Limitations\n\n\nAdditional Information\n----------------------### Dataset Curators### Licensing Information### Contributions\n\n\nThanks to @thomwolf, @lewtun, @patrickvonplaten for adding this dataset." ]
35f9d33020aa73f5ad123efd5402ad9bec1e856b
# Dataset Card for Thai Literature Corpora (TLC) ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://attapol.github.io/tlc.html - **Leaderboard:** https://www.kaggle.com/c/wisesight-sentiment/ - **Paper:** - **Leaderboard:** - **Point of Contact:** Jitkapat Sawatphol, Attapol Rutherford; attapolrutherford at gmail.com ### Dataset Summary Thai Literature Corpora (TLC): Corpora of machine-ingestible Thai classical literature texts. It consists of two datasets: ## TLC set It is texts from [Vajirayana Digital Library](https://vajirayana.org/), stored by chapters and stanzas (non-tokenized). tlc v.2.0 (6/17/19 : a total of 34 documents, 292,270 lines, 31,790,734 characters) tlc v.1.0 (6/11/19 : a total of 25 documents, 113,981 lines, 28,775,761 characters) ## TNHC set It is texts from Thai National Historical Corpus, stored by lines (manually tokenized). tnhc v.1.0 (6/25/19 : a total of 47 documents, 756,478 lines, 13,361,142 characters) ### Supported Tasks and Leaderboards Language Modeling, Language Generation ### Languages Thai ## Dataset Structure ### Data Instances ``` { "ch_num": "๑", "title": "กากี กลอนสุภาพ", "text": [ [ "๏ จักกล่าวอดีตนิทานแต่ปางก่อน\n", "เมื่อครั้งองค์สมเด็จพระชินวร\tยังสัญจรแสวงหาโพธิญาณ\n", "เสวยชาติเป็นสกุณาพระยานก\tจึงชักเรื่องชาดกมาบรรหาร\n", "หวังแสดงแห่งจิตหญิงพาล\tให้ชายชาญรู้เชิงกระสัตรี ฯ\n" ] } ``` ### Data Fields - `ch_num`: chapter number in Thai Numerals (๑, ๒, ๓, ๔, ๕, ๖, ๗, ๘, ๙, ๑๐, ...) - `title`: chapter name - `text`: each item corresponds to one stanzas, each line is a couplet which can be seperated by `\t` ### Data Splits tlc v.2.0 (6/17/19 : a total of 34 documents, 292,270 lines, 31,790,734 characters) tlc v.1.0 (6/11/19 : a total of 25 documents, 113,981 lines, 28,775,761 characters) ## TNHC set It is texts from Thai National Historical Corpus, stored by lines (manually tokenized). tnhc v.1.0 (6/25/19 : a total of 47 documents, 756,478 lines, 13,361,142 characters) | | tlc2.0 | tlc1.0 | tnhc | |-----------|-------|-------|-------| | # documents | 34 | 25 | 47 | | # lines | 292,270 | 113,981 | 756,478 | ## Dataset Creation ### Curation Rationale Originally, the dataset was compiled for the [Thai Poetry Generator](https://github.com/jitkapat/thaipoetrygenerator) at Chulalongkorn university as the Final project for `2209372 Introduction to Computational Linguistics` by [Jitkapat Sawatphol](https://jitkapat.github.io/) (Faculty of Engineering, Chulalongkorn University). ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information There is no personal information. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators Thanks [Jitkapat Sawatphol](https://jitkapat.github.io/) (Faculty of Arts, Chulalongkorn University), and [Attapol Rutherford](https://attapol.github.io/) (Faculty of Arts, Chulalongkorn University) ### Licensing Information [More Information Needed] ### Citation Information Please cite the following if you make use of the dataset: Jitkapat Sawatphol, and Attapol Rutherford. 2019. **Thai Literature Corpora (TLC)**. BibTeX: ``` @misc{ author={Sawatphol, Jitkapat}, title={Thai Literature Corpora}, year={2019}, howpublished={\\url{https://attapol.github.io/tlc.html}} } ``` ### Contributions Thanks to [@chameleonTK](https://github.com/chameleonTK) for adding this dataset.
tlc
[ "task_categories:text-generation", "task_categories:fill-mask", "task_ids:language-modeling", "task_ids:masked-language-modeling", "annotations_creators:expert-generated", "annotations_creators:no-annotation", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:n<1K", "source_datasets:original", "language:th", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated", "no-annotation"], "language_creators": ["expert-generated"], "language": ["th"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["n<1K"], "source_datasets": ["original"], "task_categories": ["text-generation", "fill-mask"], "task_ids": ["language-modeling", "masked-language-modeling"], "pretty_name": "Thai Literature Corpora (TLC)", "dataset_info": [{"config_name": "tlcv1.0", "features": [{"name": "ch_num", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "text", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 32498, "num_examples": 1}], "download_size": 2904472, "dataset_size": 32498}, {"config_name": "tlcv2.0", "features": [{"name": "ch_num", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "text", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 32498, "num_examples": 1}], "download_size": 5551710, "dataset_size": 32498}, {"config_name": "tnhcv1.0", "features": [{"name": "text", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 25198, "num_examples": 152}], "download_size": 1465403, "dataset_size": 25198}]}
2024-01-18T11:17:15+00:00
[]
[ "th" ]
TAGS #task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-expert-generated #annotations_creators-no-annotation #language_creators-expert-generated #multilinguality-monolingual #size_categories-n<1K #source_datasets-original #language-Thai #license-unknown #region-us
Dataset Card for Thai Literature Corpora (TLC) ============================================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Leaderboard: URL * Paper: * Leaderboard: * Point of Contact: Jitkapat Sawatphol, Attapol Rutherford; attapolrutherford at URL ### Dataset Summary Thai Literature Corpora (TLC): Corpora of machine-ingestible Thai classical literature texts. It consists of two datasets: TLC set ------- It is texts from Vajirayana Digital Library, stored by chapters and stanzas (non-tokenized). tlc v.2.0 (6/17/19 : a total of 34 documents, 292,270 lines, 31,790,734 characters) tlc v.1.0 (6/11/19 : a total of 25 documents, 113,981 lines, 28,775,761 characters) TNHC set -------- It is texts from Thai National Historical Corpus, stored by lines (manually tokenized). tnhc v.1.0 (6/25/19 : a total of 47 documents, 756,478 lines, 13,361,142 characters) ### Supported Tasks and Leaderboards Language Modeling, Language Generation ### Languages Thai Dataset Structure ----------------- ### Data Instances ### Data Fields * 'ch\_num': chapter number in Thai Numerals (๑, ๒, ๓, ๔, ๕, ๖, ๗, ๘, ๙, ๑๐, ...) * 'title': chapter name * 'text': each item corresponds to one stanzas, each line is a couplet which can be seperated by '\t' ### Data Splits tlc v.2.0 (6/17/19 : a total of 34 documents, 292,270 lines, 31,790,734 characters) tlc v.1.0 (6/11/19 : a total of 25 documents, 113,981 lines, 28,775,761 characters) TNHC set -------- It is texts from Thai National Historical Corpus, stored by lines (manually tokenized). tnhc v.1.0 (6/25/19 : a total of 47 documents, 756,478 lines, 13,361,142 characters) Dataset Creation ---------------- ### Curation Rationale Originally, the dataset was compiled for the Thai Poetry Generator at Chulalongkorn university as the Final project for '2209372 Introduction to Computational Linguistics' by Jitkapat Sawatphol (Faculty of Engineering, Chulalongkorn University). ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information There is no personal information. Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators Thanks Jitkapat Sawatphol (Faculty of Arts, Chulalongkorn University), and Attapol Rutherford (Faculty of Arts, Chulalongkorn University) ### Licensing Information Please cite the following if you make use of the dataset: Jitkapat Sawatphol, and Attapol Rutherford. 2019. Thai Literature Corpora (TLC). BibTeX: ### Contributions Thanks to @chameleonTK for adding this dataset.
[ "### Dataset Summary\n\n\nThai Literature Corpora (TLC): Corpora of machine-ingestible Thai classical literature texts.\n\n\nIt consists of two datasets:\n\n\nTLC set\n-------\n\n\nIt is texts from Vajirayana Digital Library, stored by chapters and stanzas (non-tokenized).\n\n\ntlc v.2.0 (6/17/19 : a total of 34 documents, 292,270 lines, 31,790,734 characters)\ntlc v.1.0 (6/11/19 : a total of 25 documents, 113,981 lines, 28,775,761 characters)\n\n\nTNHC set\n--------\n\n\nIt is texts from Thai National Historical Corpus, stored by lines (manually tokenized).\n\n\ntnhc v.1.0 (6/25/19 : a total of 47 documents, 756,478 lines, 13,361,142 characters)", "### Supported Tasks and Leaderboards\n\n\nLanguage Modeling, Language Generation", "### Languages\n\n\nThai\n\n\nDataset Structure\n-----------------", "### Data Instances", "### Data Fields\n\n\n* 'ch\\_num': chapter number in Thai Numerals (๑, ๒, ๓, ๔, ๕, ๖, ๗, ๘, ๙, ๑๐, ...)\n* 'title': chapter name\n* 'text': each item corresponds to one stanzas, each line is a couplet which can be seperated by '\\t'", "### Data Splits\n\n\ntlc v.2.0 (6/17/19 : a total of 34 documents, 292,270 lines, 31,790,734 characters)\ntlc v.1.0 (6/11/19 : a total of 25 documents, 113,981 lines, 28,775,761 characters)\n\n\nTNHC set\n--------\n\n\nIt is texts from Thai National Historical Corpus, stored by lines (manually tokenized).\n\n\ntnhc v.1.0 (6/25/19 : a total of 47 documents, 756,478 lines, 13,361,142 characters)\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nOriginally, the dataset was compiled for the Thai Poetry Generator at Chulalongkorn university as the Final project for '2209372 Introduction to Computational Linguistics' by Jitkapat Sawatphol (Faculty of Engineering, Chulalongkorn University).", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nThere is no personal information.\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThanks Jitkapat Sawatphol (Faculty of Arts, Chulalongkorn University), and Attapol Rutherford (Faculty of Arts, Chulalongkorn University)", "### Licensing Information\n\n\nPlease cite the following if you make use of the dataset:\n\n\nJitkapat Sawatphol, and Attapol Rutherford. 2019. Thai Literature Corpora (TLC).\n\n\nBibTeX:", "### Contributions\n\n\nThanks to @chameleonTK for adding this dataset." ]
[ "TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-expert-generated #annotations_creators-no-annotation #language_creators-expert-generated #multilinguality-monolingual #size_categories-n<1K #source_datasets-original #language-Thai #license-unknown #region-us \n", "### Dataset Summary\n\n\nThai Literature Corpora (TLC): Corpora of machine-ingestible Thai classical literature texts.\n\n\nIt consists of two datasets:\n\n\nTLC set\n-------\n\n\nIt is texts from Vajirayana Digital Library, stored by chapters and stanzas (non-tokenized).\n\n\ntlc v.2.0 (6/17/19 : a total of 34 documents, 292,270 lines, 31,790,734 characters)\ntlc v.1.0 (6/11/19 : a total of 25 documents, 113,981 lines, 28,775,761 characters)\n\n\nTNHC set\n--------\n\n\nIt is texts from Thai National Historical Corpus, stored by lines (manually tokenized).\n\n\ntnhc v.1.0 (6/25/19 : a total of 47 documents, 756,478 lines, 13,361,142 characters)", "### Supported Tasks and Leaderboards\n\n\nLanguage Modeling, Language Generation", "### Languages\n\n\nThai\n\n\nDataset Structure\n-----------------", "### Data Instances", "### Data Fields\n\n\n* 'ch\\_num': chapter number in Thai Numerals (๑, ๒, ๓, ๔, ๕, ๖, ๗, ๘, ๙, ๑๐, ...)\n* 'title': chapter name\n* 'text': each item corresponds to one stanzas, each line is a couplet which can be seperated by '\\t'", "### Data Splits\n\n\ntlc v.2.0 (6/17/19 : a total of 34 documents, 292,270 lines, 31,790,734 characters)\ntlc v.1.0 (6/11/19 : a total of 25 documents, 113,981 lines, 28,775,761 characters)\n\n\nTNHC set\n--------\n\n\nIt is texts from Thai National Historical Corpus, stored by lines (manually tokenized).\n\n\ntnhc v.1.0 (6/25/19 : a total of 47 documents, 756,478 lines, 13,361,142 characters)\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nOriginally, the dataset was compiled for the Thai Poetry Generator at Chulalongkorn university as the Final project for '2209372 Introduction to Computational Linguistics' by Jitkapat Sawatphol (Faculty of Engineering, Chulalongkorn University).", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nThere is no personal information.\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThanks Jitkapat Sawatphol (Faculty of Arts, Chulalongkorn University), and Attapol Rutherford (Faculty of Arts, Chulalongkorn University)", "### Licensing Information\n\n\nPlease cite the following if you make use of the dataset:\n\n\nJitkapat Sawatphol, and Attapol Rutherford. 2019. Thai Literature Corpora (TLC).\n\n\nBibTeX:", "### Contributions\n\n\nThanks to @chameleonTK for adding this dataset." ]
[ 126, 177, 16, 12, 6, 88, 120, 71, 4, 10, 10, 5, 5, 9, 24, 7, 8, 14, 49, 49, 18 ]
[ "passage: TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-expert-generated #annotations_creators-no-annotation #language_creators-expert-generated #multilinguality-monolingual #size_categories-n<1K #source_datasets-original #language-Thai #license-unknown #region-us \n### Dataset Summary\n\n\nThai Literature Corpora (TLC): Corpora of machine-ingestible Thai classical literature texts.\n\n\nIt consists of two datasets:\n\n\nTLC set\n-------\n\n\nIt is texts from Vajirayana Digital Library, stored by chapters and stanzas (non-tokenized).\n\n\ntlc v.2.0 (6/17/19 : a total of 34 documents, 292,270 lines, 31,790,734 characters)\ntlc v.1.0 (6/11/19 : a total of 25 documents, 113,981 lines, 28,775,761 characters)\n\n\nTNHC set\n--------\n\n\nIt is texts from Thai National Historical Corpus, stored by lines (manually tokenized).\n\n\ntnhc v.1.0 (6/25/19 : a total of 47 documents, 756,478 lines, 13,361,142 characters)### Supported Tasks and Leaderboards\n\n\nLanguage Modeling, Language Generation### Languages\n\n\nThai\n\n\nDataset Structure\n-----------------### Data Instances### Data Fields\n\n\n* 'ch\\_num': chapter number in Thai Numerals (๑, ๒, ๓, ๔, ๕, ๖, ๗, ๘, ๙, ๑๐, ...)\n* 'title': chapter name\n* 'text': each item corresponds to one stanzas, each line is a couplet which can be seperated by '\\t'" ]
a05ffb8f51e3823a6fd16eb9e8af7646887a8db7
# Dataset Card for TMU-GFM-Dataset ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [N/A] - **Repository:** https://github.com/tmu-nlp/TMU-GFM-Dataset - **Paper:** [SOME: Reference-less Sub-Metrics Optimized for Manual Evaluations of Grammatical Error Correction](https://www.aclweb.org/anthology/2020.coling-main.573.pdf) - **Leaderboard:** [N/A] - **Point of Contact:** Check the paper. ### Dataset Summary Authors collected manual evaluations for the grammaticality, fluency, and meaning preservation of the system outputs of 1,381 sentences from CoNLL 2013. To collect the manual evaluations for various system outputs, each source sentence was corrected by the following five typical systems: statistical machine translation (SMT) (Grundkiewicz and Junczys-Dowmunt, 2018), recurrent neural network (RNN) (Luong et al., 2015), convolutional neural network (CNN) (Chollampatt and Ng, 2018), self-attention network (SAN) (Vaswani et al., 2017), and SAN with copy mechanism (SAN+Copy) (Zhao et al., 2019). Manual evaluation for the grammaticality, fluency, and meaning preservation were assigned to a total of 4,223 sentences. ### Supported Tasks and Leaderboards Grammatical Error Correction ### Languages English ## Dataset Structure ### Data Instances An example from the TMU-GFM-Dataset looks as follows: ``` {'ave_f': 3.4000000953674316, 'ave_g': 3.4000000953674316, 'ave_m': 3.5999999046325684, 'fluency': [3, 4, 3, 4, 3], 'grammer': [3, 4, 3, 4, 3], 'meaning': [3, 4, 4, 4, 3], 'output': 'After all, there will be an endless battle between the technology and human mentality.', 'source': 'Afterall there will be an endless battle between the technology and human mentality.', 'system': 'lstm,cnn'} ``` ### Data Fields The are 9 columns in the tmu-gfm-dataset. - source: source sentence. - output: system output sentence. - grammer: Grammaticaliry annotations by 5 annotators. - fluency: Fluency annotations by 5 annotators. - meaning: Meaning Preservation annotations by 5 annotators. - system: Which system the output sentence is from. - ave_g: Average grammer score. - ave_f: Average fluency score. - ave_m: Average meaning score. ### Data Splits Authors divided the dataset into train/dev/test with 3,376/422/423 sentences and used for fine-tuning BERT in thier paper. ## Dataset Creation ### Curation Rationale The authors proposed a reference-less metric trained on manual evaluations of system outputs for grammatical error correction (GEC). They said that previous studies have shown that reference-less metrics are promising; however, existing metrics are not optimized for manual evaluation of the system output because there is no dataset of system output with manual evaluation. To achieve a better correlation with manual evaluation, they created a dataset to optimize each sub-metric to the manual evaluation of GEC systems. Their annotators evaluated the output of five typical GEC systems. ### Source Data #### Initial Data Collection and Normalization Authors collected manual evaluations for the grammaticality, fluency, and meaning preservation of the system outputs of 1,381 sentences from CoNLL 2013. To collect the manual evaluations for various system outputs, each source sentence was corrected by the following five typical systems: statistical machine translation (SMT) (Grundkiewicz and Junczys-Dowmunt, 2018), recurrent neural network (RNN) (Luong et al., 2015), convolutional neural network (CNN) (Chollampatt and Ng, 2018), self-attention network (SAN) (Vaswani et al., 2017), and SAN with copy mechanism (SAN+Copy) (Zhao et al., 2019). #### Who are the source language producers? machine-generated ### Annotations #### Annotation process By excluding duplicate corrected sentences, manual evaluation for the grammaticality, fluency, and meaning preservation were assigned to a total of 4,223 sentences, as follows: - Grammaticality: Annotators evaluated the grammatical correctness of the system output. The authors followed the five-point scale evaluation criteria (4: Perfect, 3: Comprehensible, 2: Somewhat comprehensible, 1: Incomprehensible, and 0: Other) proposed by Heilman et al. (2014). - Fluency: Annotators evaluated how natural the sentence sounds for native speakers. The authors followed the criteria (4: Extremely natural, 3: Somewhat natural, 2: Somewhat unnatural, and 1: Extremely unnatural) proposed by Lau et al. (2015). - Meaning preservation: Annotators evaluated the extent to which the meaning of source sentences is preserved in system output. The authors followed the criteria (4: Identical, 3: Minor differences, 2: Moderate differences, 1: Sub- stantially different, and 0: Other) proposed by Xu et al. (2016). Finally, the authors created a dataset with manual evaluations for a total of 4,221 sentences, excluding sentences in which three or more annotators answered “0: Other.” #### Who are the annotators? Five native English annotators reqruited by using Amazon Mechaincal turk ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information @inproceedings{yoshimura-etal-2020-reference, title = "{SOME}: Reference-less Sub-Metrics Optimized for Manual Evaluations of Grammatical Error Correction", author = "Yoshimura, Ryoma and Kaneko, Masahiro and Kajiwara, Tomoyuki and Komachi, Mamoru", booktitle = "Proceedings of the 28th International Conference on Computational Linguistics", month = dec, year = "2020", address = "Barcelona, Spain (Online)", publisher = "International Committee on Computational Linguistics", url = "https://www.aclweb.org/anthology/2020.coling-main.573", pages = "6516--6522", abstract = "We propose a reference-less metric trained on manual evaluations of system outputs for grammatical error correction (GEC). Previous studies have shown that reference-less metrics are promising; however, existing metrics are not optimized for manual evaluations of the system outputs because no dataset of the system output exists with manual evaluation. This study manually evaluates outputs of GEC systems to optimize the metrics. Experimental results show that the proposed metric improves correlation with the manual evaluation in both system- and sentence-level meta-evaluation. Our dataset and metric will be made publicly available.", } ### Contributions Thanks to [@forest1988](https://github.com/forest1988) for adding this dataset.
tmu_gfm_dataset
[ "task_categories:text2text-generation", "annotations_creators:crowdsourced", "language_creators:machine-generated", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:unknown", "grammatical-error-correction", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["machine-generated"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text2text-generation"], "task_ids": [], "pretty_name": "TMU-GFM-Dataset", "tags": ["grammatical-error-correction"], "dataset_info": {"features": [{"name": "source", "dtype": "string"}, {"name": "output", "dtype": "string"}, {"name": "grammer", "sequence": "int32"}, {"name": "fluency", "sequence": "int32"}, {"name": "meaning", "sequence": "int32"}, {"name": "system", "dtype": "string"}, {"name": "ave_g", "dtype": "float32"}, {"name": "ave_f", "dtype": "float32"}, {"name": "ave_m", "dtype": "float32"}], "splits": [{"name": "train", "num_bytes": 1446144, "num_examples": 4221}], "download_size": 1270197, "dataset_size": 1446144}}
2024-01-18T11:17:16+00:00
[]
[ "en" ]
TAGS #task_categories-text2text-generation #annotations_creators-crowdsourced #language_creators-machine-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-unknown #grammatical-error-correction #region-us
# Dataset Card for TMU-GFM-Dataset ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: [N/A] - Repository: URL - Paper: SOME: Reference-less Sub-Metrics Optimized for Manual Evaluations of Grammatical Error Correction - Leaderboard: [N/A] - Point of Contact: Check the paper. ### Dataset Summary Authors collected manual evaluations for the grammaticality, fluency, and meaning preservation of the system outputs of 1,381 sentences from CoNLL 2013. To collect the manual evaluations for various system outputs, each source sentence was corrected by the following five typical systems: statistical machine translation (SMT) (Grundkiewicz and Junczys-Dowmunt, 2018), recurrent neural network (RNN) (Luong et al., 2015), convolutional neural network (CNN) (Chollampatt and Ng, 2018), self-attention network (SAN) (Vaswani et al., 2017), and SAN with copy mechanism (SAN+Copy) (Zhao et al., 2019). Manual evaluation for the grammaticality, fluency, and meaning preservation were assigned to a total of 4,223 sentences. ### Supported Tasks and Leaderboards Grammatical Error Correction ### Languages English ## Dataset Structure ### Data Instances An example from the TMU-GFM-Dataset looks as follows: ### Data Fields The are 9 columns in the tmu-gfm-dataset. - source: source sentence. - output: system output sentence. - grammer: Grammaticaliry annotations by 5 annotators. - fluency: Fluency annotations by 5 annotators. - meaning: Meaning Preservation annotations by 5 annotators. - system: Which system the output sentence is from. - ave_g: Average grammer score. - ave_f: Average fluency score. - ave_m: Average meaning score. ### Data Splits Authors divided the dataset into train/dev/test with 3,376/422/423 sentences and used for fine-tuning BERT in thier paper. ## Dataset Creation ### Curation Rationale The authors proposed a reference-less metric trained on manual evaluations of system outputs for grammatical error correction (GEC). They said that previous studies have shown that reference-less metrics are promising; however, existing metrics are not optimized for manual evaluation of the system output because there is no dataset of system output with manual evaluation. To achieve a better correlation with manual evaluation, they created a dataset to optimize each sub-metric to the manual evaluation of GEC systems. Their annotators evaluated the output of five typical GEC systems. ### Source Data #### Initial Data Collection and Normalization Authors collected manual evaluations for the grammaticality, fluency, and meaning preservation of the system outputs of 1,381 sentences from CoNLL 2013. To collect the manual evaluations for various system outputs, each source sentence was corrected by the following five typical systems: statistical machine translation (SMT) (Grundkiewicz and Junczys-Dowmunt, 2018), recurrent neural network (RNN) (Luong et al., 2015), convolutional neural network (CNN) (Chollampatt and Ng, 2018), self-attention network (SAN) (Vaswani et al., 2017), and SAN with copy mechanism (SAN+Copy) (Zhao et al., 2019). #### Who are the source language producers? machine-generated ### Annotations #### Annotation process By excluding duplicate corrected sentences, manual evaluation for the grammaticality, fluency, and meaning preservation were assigned to a total of 4,223 sentences, as follows: - Grammaticality: Annotators evaluated the grammatical correctness of the system output. The authors followed the five-point scale evaluation criteria (4: Perfect, 3: Comprehensible, 2: Somewhat comprehensible, 1: Incomprehensible, and 0: Other) proposed by Heilman et al. (2014). - Fluency: Annotators evaluated how natural the sentence sounds for native speakers. The authors followed the criteria (4: Extremely natural, 3: Somewhat natural, 2: Somewhat unnatural, and 1: Extremely unnatural) proposed by Lau et al. (2015). - Meaning preservation: Annotators evaluated the extent to which the meaning of source sentences is preserved in system output. The authors followed the criteria (4: Identical, 3: Minor differences, 2: Moderate differences, 1: Sub- stantially different, and 0: Other) proposed by Xu et al. (2016). Finally, the authors created a dataset with manual evaluations for a total of 4,221 sentences, excluding sentences in which three or more annotators answered “0: Other.” #### Who are the annotators? Five native English annotators reqruited by using Amazon Mechaincal turk ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information @inproceedings{yoshimura-etal-2020-reference, title = "{SOME}: Reference-less Sub-Metrics Optimized for Manual Evaluations of Grammatical Error Correction", author = "Yoshimura, Ryoma and Kaneko, Masahiro and Kajiwara, Tomoyuki and Komachi, Mamoru", booktitle = "Proceedings of the 28th International Conference on Computational Linguistics", month = dec, year = "2020", address = "Barcelona, Spain (Online)", publisher = "International Committee on Computational Linguistics", url = "URL pages = "6516--6522", abstract = "We propose a reference-less metric trained on manual evaluations of system outputs for grammatical error correction (GEC). Previous studies have shown that reference-less metrics are promising; however, existing metrics are not optimized for manual evaluations of the system outputs because no dataset of the system output exists with manual evaluation. This study manually evaluates outputs of GEC systems to optimize the metrics. Experimental results show that the proposed metric improves correlation with the manual evaluation in both system- and sentence-level meta-evaluation. Our dataset and metric will be made publicly available.", } ### Contributions Thanks to @forest1988 for adding this dataset.
[ "# Dataset Card for TMU-GFM-Dataset", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: [N/A]\n- Repository: URL\n- Paper: SOME: Reference-less Sub-Metrics Optimized for Manual Evaluations of Grammatical Error Correction\n- Leaderboard: [N/A]\n- Point of Contact: Check the paper.", "### Dataset Summary\n\nAuthors collected manual evaluations for the grammaticality, fluency, and meaning preservation of the system outputs of 1,381 sentences from CoNLL 2013.\nTo collect the manual evaluations for various system outputs, each source sentence was corrected by the following five typical systems: statistical machine translation (SMT) (Grundkiewicz and Junczys-Dowmunt, 2018), recurrent neural network (RNN) (Luong et al., 2015), convolutional neural network (CNN) (Chollampatt and Ng, 2018), self-attention network (SAN) (Vaswani et al., 2017), and SAN with copy mechanism (SAN+Copy) (Zhao et al., 2019).\nManual evaluation for the grammaticality, fluency, and meaning preservation were assigned to a total of 4,223 sentences.", "### Supported Tasks and Leaderboards\n\nGrammatical Error Correction", "### Languages\n\nEnglish", "## Dataset Structure", "### Data Instances\n\nAn example from the TMU-GFM-Dataset looks as follows:", "### Data Fields\n\nThe are 9 columns in the tmu-gfm-dataset.\n\n- source: source sentence.\n- output: system output sentence.\n- grammer: Grammaticaliry annotations by 5 annotators.\n- fluency: Fluency annotations by 5 annotators.\n- meaning: Meaning Preservation annotations by 5 annotators.\n- system: Which system the output sentence is from.\n- ave_g: Average grammer score.\n- ave_f: Average fluency score.\n- ave_m: Average meaning score.", "### Data Splits\n\nAuthors divided the dataset into train/dev/test with 3,376/422/423 sentences and used for fine-tuning BERT in thier paper.", "## Dataset Creation", "### Curation Rationale\n\nThe authors proposed a reference-less metric trained on manual evaluations of system outputs for grammatical error correction (GEC). \nThey said that previous studies have shown that reference-less metrics are promising; however, existing metrics are not optimized for manual evaluation of the system output because there is no dataset of system output with manual evaluation.\nTo achieve a better correlation with manual evaluation, they created a dataset to optimize each sub-metric to the manual evaluation of GEC systems. Their annotators evaluated the output of five typical GEC systems.", "### Source Data", "#### Initial Data Collection and Normalization\n\nAuthors collected manual evaluations for the grammaticality, fluency, and meaning preservation of the system outputs of 1,381 sentences from CoNLL 2013.\nTo collect the manual evaluations for various system outputs, each source sentence was corrected by the following five typical systems: statistical machine translation (SMT) (Grundkiewicz and Junczys-Dowmunt, 2018), recurrent neural network (RNN) (Luong et al., 2015), convolutional neural network (CNN) (Chollampatt and Ng, 2018), self-attention network (SAN) (Vaswani et al., 2017), and SAN with copy mechanism (SAN+Copy) (Zhao et al., 2019).", "#### Who are the source language producers?\n\nmachine-generated", "### Annotations", "#### Annotation process\n\nBy excluding duplicate corrected sentences, manual evaluation for the grammaticality, fluency, and meaning preservation were assigned to a total of 4,223 sentences, as follows: \n- Grammaticality: Annotators evaluated the grammatical correctness of the system output. The authors followed the five-point scale evaluation criteria (4: Perfect, 3: Comprehensible, 2: Somewhat comprehensible, 1: Incomprehensible, and 0: Other) proposed by Heilman et al. (2014). \n- Fluency: Annotators evaluated how natural the sentence sounds for native speakers. The authors followed the criteria (4: Extremely natural, 3: Somewhat natural, 2: Somewhat unnatural, and 1: Extremely unnatural) proposed by Lau et al. (2015). \n- Meaning preservation: Annotators evaluated the extent to which the meaning of source sentences is preserved in system output. The authors followed the criteria (4: Identical, 3: Minor differences, 2: Moderate differences, 1: Sub- stantially different, and 0: Other) proposed by Xu et al. (2016).\n\nFinally, the authors created a dataset with manual evaluations for a total of 4,221 sentences, excluding sentences in which three or more annotators answered “0: Other.”", "#### Who are the annotators?\n\nFive native English annotators reqruited by using Amazon Mechaincal turk", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\n\n\n\n\n@inproceedings{yoshimura-etal-2020-reference,\n title = \"{SOME}: Reference-less Sub-Metrics Optimized for Manual Evaluations of Grammatical Error Correction\",\n author = \"Yoshimura, Ryoma and\n Kaneko, Masahiro and\n Kajiwara, Tomoyuki and\n Komachi, Mamoru\",\n booktitle = \"Proceedings of the 28th International Conference on Computational Linguistics\",\n month = dec,\n year = \"2020\",\n address = \"Barcelona, Spain (Online)\",\n publisher = \"International Committee on Computational Linguistics\",\n url = \"URL\n pages = \"6516--6522\",\n abstract = \"We propose a reference-less metric trained on manual evaluations of system outputs for grammatical error correction (GEC). Previous studies have shown that reference-less metrics are promising; however, existing metrics are not optimized for manual evaluations of the system outputs because no dataset of the system output exists with manual evaluation. This study manually evaluates outputs of GEC systems to optimize the metrics. Experimental results show that the proposed metric improves correlation with the manual evaluation in both system- and sentence-level meta-evaluation. Our dataset and metric will be made publicly available.\",\n}", "### Contributions\n\nThanks to @forest1988 for adding this dataset." ]
[ "TAGS\n#task_categories-text2text-generation #annotations_creators-crowdsourced #language_creators-machine-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-unknown #grammatical-error-correction #region-us \n", "# Dataset Card for TMU-GFM-Dataset", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: [N/A]\n- Repository: URL\n- Paper: SOME: Reference-less Sub-Metrics Optimized for Manual Evaluations of Grammatical Error Correction\n- Leaderboard: [N/A]\n- Point of Contact: Check the paper.", "### Dataset Summary\n\nAuthors collected manual evaluations for the grammaticality, fluency, and meaning preservation of the system outputs of 1,381 sentences from CoNLL 2013.\nTo collect the manual evaluations for various system outputs, each source sentence was corrected by the following five typical systems: statistical machine translation (SMT) (Grundkiewicz and Junczys-Dowmunt, 2018), recurrent neural network (RNN) (Luong et al., 2015), convolutional neural network (CNN) (Chollampatt and Ng, 2018), self-attention network (SAN) (Vaswani et al., 2017), and SAN with copy mechanism (SAN+Copy) (Zhao et al., 2019).\nManual evaluation for the grammaticality, fluency, and meaning preservation were assigned to a total of 4,223 sentences.", "### Supported Tasks and Leaderboards\n\nGrammatical Error Correction", "### Languages\n\nEnglish", "## Dataset Structure", "### Data Instances\n\nAn example from the TMU-GFM-Dataset looks as follows:", "### Data Fields\n\nThe are 9 columns in the tmu-gfm-dataset.\n\n- source: source sentence.\n- output: system output sentence.\n- grammer: Grammaticaliry annotations by 5 annotators.\n- fluency: Fluency annotations by 5 annotators.\n- meaning: Meaning Preservation annotations by 5 annotators.\n- system: Which system the output sentence is from.\n- ave_g: Average grammer score.\n- ave_f: Average fluency score.\n- ave_m: Average meaning score.", "### Data Splits\n\nAuthors divided the dataset into train/dev/test with 3,376/422/423 sentences and used for fine-tuning BERT in thier paper.", "## Dataset Creation", "### Curation Rationale\n\nThe authors proposed a reference-less metric trained on manual evaluations of system outputs for grammatical error correction (GEC). \nThey said that previous studies have shown that reference-less metrics are promising; however, existing metrics are not optimized for manual evaluation of the system output because there is no dataset of system output with manual evaluation.\nTo achieve a better correlation with manual evaluation, they created a dataset to optimize each sub-metric to the manual evaluation of GEC systems. Their annotators evaluated the output of five typical GEC systems.", "### Source Data", "#### Initial Data Collection and Normalization\n\nAuthors collected manual evaluations for the grammaticality, fluency, and meaning preservation of the system outputs of 1,381 sentences from CoNLL 2013.\nTo collect the manual evaluations for various system outputs, each source sentence was corrected by the following five typical systems: statistical machine translation (SMT) (Grundkiewicz and Junczys-Dowmunt, 2018), recurrent neural network (RNN) (Luong et al., 2015), convolutional neural network (CNN) (Chollampatt and Ng, 2018), self-attention network (SAN) (Vaswani et al., 2017), and SAN with copy mechanism (SAN+Copy) (Zhao et al., 2019).", "#### Who are the source language producers?\n\nmachine-generated", "### Annotations", "#### Annotation process\n\nBy excluding duplicate corrected sentences, manual evaluation for the grammaticality, fluency, and meaning preservation were assigned to a total of 4,223 sentences, as follows: \n- Grammaticality: Annotators evaluated the grammatical correctness of the system output. The authors followed the five-point scale evaluation criteria (4: Perfect, 3: Comprehensible, 2: Somewhat comprehensible, 1: Incomprehensible, and 0: Other) proposed by Heilman et al. (2014). \n- Fluency: Annotators evaluated how natural the sentence sounds for native speakers. The authors followed the criteria (4: Extremely natural, 3: Somewhat natural, 2: Somewhat unnatural, and 1: Extremely unnatural) proposed by Lau et al. (2015). \n- Meaning preservation: Annotators evaluated the extent to which the meaning of source sentences is preserved in system output. The authors followed the criteria (4: Identical, 3: Minor differences, 2: Moderate differences, 1: Sub- stantially different, and 0: Other) proposed by Xu et al. (2016).\n\nFinally, the authors created a dataset with manual evaluations for a total of 4,221 sentences, excluding sentences in which three or more annotators answered “0: Other.”", "#### Who are the annotators?\n\nFive native English annotators reqruited by using Amazon Mechaincal turk", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\n\n\n\n\n@inproceedings{yoshimura-etal-2020-reference,\n title = \"{SOME}: Reference-less Sub-Metrics Optimized for Manual Evaluations of Grammatical Error Correction\",\n author = \"Yoshimura, Ryoma and\n Kaneko, Masahiro and\n Kajiwara, Tomoyuki and\n Komachi, Mamoru\",\n booktitle = \"Proceedings of the 28th International Conference on Computational Linguistics\",\n month = dec,\n year = \"2020\",\n address = \"Barcelona, Spain (Online)\",\n publisher = \"International Committee on Computational Linguistics\",\n url = \"URL\n pages = \"6516--6522\",\n abstract = \"We propose a reference-less metric trained on manual evaluations of system outputs for grammatical error correction (GEC). Previous studies have shown that reference-less metrics are promising; however, existing metrics are not optimized for manual evaluations of the system outputs because no dataset of the system output exists with manual evaluation. This study manually evaluates outputs of GEC systems to optimize the metrics. Experimental results show that the proposed metric improves correlation with the manual evaluation in both system- and sentence-level meta-evaluation. Our dataset and metric will be made publicly available.\",\n}", "### Contributions\n\nThanks to @forest1988 for adding this dataset." ]
[ 91, 13, 120, 63, 197, 16, 5, 6, 23, 135, 43, 5, 132, 4, 171, 14, 5, 298, 27, 8, 8, 7, 8, 7, 5, 6, 300, 17 ]
[ "passage: TAGS\n#task_categories-text2text-generation #annotations_creators-crowdsourced #language_creators-machine-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-unknown #grammatical-error-correction #region-us \n# Dataset Card for TMU-GFM-Dataset## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: [N/A]\n- Repository: URL\n- Paper: SOME: Reference-less Sub-Metrics Optimized for Manual Evaluations of Grammatical Error Correction\n- Leaderboard: [N/A]\n- Point of Contact: Check the paper.### Dataset Summary\n\nAuthors collected manual evaluations for the grammaticality, fluency, and meaning preservation of the system outputs of 1,381 sentences from CoNLL 2013.\nTo collect the manual evaluations for various system outputs, each source sentence was corrected by the following five typical systems: statistical machine translation (SMT) (Grundkiewicz and Junczys-Dowmunt, 2018), recurrent neural network (RNN) (Luong et al., 2015), convolutional neural network (CNN) (Chollampatt and Ng, 2018), self-attention network (SAN) (Vaswani et al., 2017), and SAN with copy mechanism (SAN+Copy) (Zhao et al., 2019).\nManual evaluation for the grammaticality, fluency, and meaning preservation were assigned to a total of 4,223 sentences.### Supported Tasks and Leaderboards\n\nGrammatical Error Correction### Languages\n\nEnglish", "passage: ## Dataset Structure### Data Instances\n\nAn example from the TMU-GFM-Dataset looks as follows:### Data Fields\n\nThe are 9 columns in the tmu-gfm-dataset.\n\n- source: source sentence.\n- output: system output sentence.\n- grammer: Grammaticaliry annotations by 5 annotators.\n- fluency: Fluency annotations by 5 annotators.\n- meaning: Meaning Preservation annotations by 5 annotators.\n- system: Which system the output sentence is from.\n- ave_g: Average grammer score.\n- ave_f: Average fluency score.\n- ave_m: Average meaning score.### Data Splits\n\nAuthors divided the dataset into train/dev/test with 3,376/422/423 sentences and used for fine-tuning BERT in thier paper.## Dataset Creation### Curation Rationale\n\nThe authors proposed a reference-less metric trained on manual evaluations of system outputs for grammatical error correction (GEC). \nThey said that previous studies have shown that reference-less metrics are promising; however, existing metrics are not optimized for manual evaluation of the system output because there is no dataset of system output with manual evaluation.\nTo achieve a better correlation with manual evaluation, they created a dataset to optimize each sub-metric to the manual evaluation of GEC systems. Their annotators evaluated the output of five typical GEC systems.### Source Data", "passage: #### Initial Data Collection and Normalization\n\nAuthors collected manual evaluations for the grammaticality, fluency, and meaning preservation of the system outputs of 1,381 sentences from CoNLL 2013.\nTo collect the manual evaluations for various system outputs, each source sentence was corrected by the following five typical systems: statistical machine translation (SMT) (Grundkiewicz and Junczys-Dowmunt, 2018), recurrent neural network (RNN) (Luong et al., 2015), convolutional neural network (CNN) (Chollampatt and Ng, 2018), self-attention network (SAN) (Vaswani et al., 2017), and SAN with copy mechanism (SAN+Copy) (Zhao et al., 2019).#### Who are the source language producers?\n\nmachine-generated### Annotations#### Annotation process\n\nBy excluding duplicate corrected sentences, manual evaluation for the grammaticality, fluency, and meaning preservation were assigned to a total of 4,223 sentences, as follows: \n- Grammaticality: Annotators evaluated the grammatical correctness of the system output. The authors followed the five-point scale evaluation criteria (4: Perfect, 3: Comprehensible, 2: Somewhat comprehensible, 1: Incomprehensible, and 0: Other) proposed by Heilman et al. (2014). \n- Fluency: Annotators evaluated how natural the sentence sounds for native speakers. The authors followed the criteria (4: Extremely natural, 3: Somewhat natural, 2: Somewhat unnatural, and 1: Extremely unnatural) proposed by Lau et al. (2015). \n- Meaning preservation: Annotators evaluated the extent to which the meaning of source sentences is preserved in system output. The authors followed the criteria (4: Identical, 3: Minor differences, 2: Moderate differences, 1: Sub- stantially different, and 0: Other) proposed by Xu et al. (2016).\n\nFinally, the authors created a dataset with manual evaluations for a total of 4,221 sentences, excluding sentences in which three or more annotators answered “0: Other.”#### Who are the annotators?\n\nFive native English annotators reqruited by using Amazon Mechaincal turk### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators" ]
fb4f11a5bc68b99891852d20f1ec074be6289768
# Dataset Card for "ToLD-Br" ## Table of Contents - [Dataset Card for "ToLD-Br"](#dataset-card-for-told-br) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Initial Data Collection and Normalization](#initial-data-collection-and-normalization) - [Who are the source language producers?](#who-are-the-source-language-producers) - [Annotations](#annotations) - [Annotation process](#annotation-process) - [Who are the annotators?](#who-are-the-annotators) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://paperswithcode.com/dataset/told-br - **Repository:** https://github.com/JAugusto97/ToLD-Br - **Paper:** https://arxiv.org/abs/2010.04543 - **Leaderboard:** https://paperswithcode.com/sota/hate-speech-detection-on-told-br - **Point of Contact:** joao.leite@estudante.ufscar.br ### Dataset Summary ToLD-Br is the biggest dataset for toxic tweets in Brazilian Portuguese, crowdsourced by 42 annotators selected from a pool of 129 volunteers. Annotators were selected aiming to create a plural group in terms of demographics (ethnicity, sexual orientation, age, gender). Each tweet was labeled by three annotators in 6 possible categories: LGBTQ+phobia, Xenophobia, Obscene, Insult, Misogyny and Racism. ### Supported Tasks and Leaderboards -`text-classification-other-hate-speech-detection`: The dataset can be used to train a model for Hate Speech Detection, either using it's multi-label classes or by grouping them into a binary Hate vs. Non-Hate class. A [BERT](https://huggingface.co/docs/transformers/model_doc/bert) model can be fine-tuned to perform this task and achieve 0.75 F1-Score for it's binary version. ### Languages The text in the dataset is in Brazilian Portuguese, as spoken by Tweet users. The associated BCP-47 code is `pt-BR`. ## Dataset Structure ### Data Instances ToLD-Br has two versions: binary and multilabel. Multilabel: A data point consists of the tweet text (string) followed by 6 categories that have values ranging from 0 to 3, meaning the amount of votes from annotators for that specific class on homophobia, obscene, insult, racism, misogyny and xenophobia. An example from multilabel ToLD-Br looks as follows: ``` {'text': '@user bandido dissimulado. esse sérgio moro é uma espécie de mal carater com ditadura e pitadas de atraso' 'homophobia': 0 'obscene': 0 'insult': 2 'racism': 0 'misogyny': 0 'xenophobia': 0} ``` Binary: A data point consists of the tweet text (string) followed by a binary class "toxic" with values 0 or 1. An example from binary ToLD-Br looks as follows: ``` {'text': '@user bandido dissimulado. esse sérgio moro é uma espécie de mal carater com ditadura e pitadas de atraso' 'toxic': 1} ``` ### Data Fields Multilabel: - text: A string representing the tweet posted by a user. Mentions to other users are anonymized by replacing the mention with a @user tag. - homophobia: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as homophobic. - obscene: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as obscene. - insult: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as insult. - racism: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as racism. - misogyny: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as misogyny. - xenophobia: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as xenophobia. Binary: - text: A string representing the tweet posted by a user. Mentions to other users are anonymized by replacing the mention with a @user tag. - label: numerical binary value {0, 1} representing if the respective text is toxic/abusive or not. ### Data Splits Multilabel: The entire dataset consists of 21.000 examples. Binary: The train set consists of 16.800 examples, validation set consists of 2.100 examples and test set consists of 2.100 examples. ## Dataset Creation ### Curation Rationale Despite Portuguese being the 5th most spoken language in the world and Brazil being the 4th country with most unique users, Brazilian Portuguese was underrepresented in the hate-speech detection task. Only two other datasets were available, one of them being European Portuguese. ToLD-Br is 4x bigger than both these datasets combined. Also, none of them had multiple annotators per instance. Also, this work proposes a plural and diverse group of annotators carefully selected to avoid inserting bias into the annotation. ### Source Data #### Initial Data Collection and Normalization Data was collected in 15 days in August 2019 using Gate Cloud's Tweet Collector. Ten million tweets were collected using two methods: a keyword-based method and a user-mention method. The first method collected tweets mentioning the following keywords: viado,veado,viadinho,veadinho,viadao,veadao,bicha,bixa,bichinha,bixinha,bichona,bixona,baitola,sapatão,sapatao,traveco,bambi,biba,boiola,marica,gayzão,gayzao,flor,florzinha,vagabundo,vagaba,desgraçada,desgraçado,desgracado,arrombado,arrombada,foder,fuder,fudido,fodido,cú,cu,pinto,pau,pal,caralho,caraio,carai,pica,cacete,rola,porra,escroto,buceta,fdp,pqp,vsf,tnc,vtnc,puto,putinho,acéfalo,acefalo,burro,idiota,trouxa,estúpido,estupido,estúpida,canalha,demente,retardado,retardada,verme,maldito,maldita,ridículo,ridiculo,ridícula,ridicula,morfético,morfetico,morfética,morfetica,lazarento,lazarenta,lixo,mongolóide,mongoloide,mongol,asqueroso,asquerosa,cretino,cretina,babaca,pilantra,neguinho,neguinha,pretinho,pretinha,escurinho,escurinha,pretinha,pretinho,crioulo,criolo,crioula,criola,macaco,macaca,gorila,puta,vagabunda,vagaba,mulherzinha,piranha,feminazi,putinha,piriguete,vaca,putinha,bahiano,baiano,baianagem,xingling,xing ling,xing-ling,carioca,paulista,sulista,mineiro,gringo The list of most followed Brazilian Twitter accounts can be found [here](https://assuperlistas.com/2022/01/21/os-100-brasileiros-mais-seguidos-do-twitter/). #### Who are the source language producers? The language producers are Twitter users from Brazil, speakers of Portuguese. ### Annotations #### Annotation process A form was published at the Federal University of São Carlos asking for volunteers to annotate our dataset. 129 people volunteered and 42 were selected according to their demographics in order to create a diverse and plural annotation group. Guidelines were produced and presented to the annotators. The entire process was done asynchronously because of the Covid-19 pandemic. The tool used was Google Sheets. Annotators were grouped into 14 teams of three annotators each. Each group annotated a respective file containing 1500 tweets. Annotators didn't have contact with each other, nor did they know that other annotators were labelling the same tweets as they were. #### Who are the annotators? Annotators were people from the Federal University of São Carlos' Facebook group. Their demographics are described below: | Gender | | |--------|--------| | Male | 18 | | Female | 24 | | Sexual Orientation | | |--------------------|----| | Heterosexual | 22 | | Bisexual | 12 | | Homosexual | 5 | | Pansexual | 3 | | Ethnicity | | |--------------|----| | White | 25 | | Brown | 9 | | Black | 5 | | Asian | 2 | | Non-Declared | 1 | Ages range from 18 to 37 years old. Annotators were paid R$50 ($10) to label 1500 examples each. ### Personal and Sensitive Information The dataset contains sensitive information for homophobia, obscene, insult, racism, misogyny and xenophobia. Tweets were anonymized by replacing user mentions with a @user tag. ## Considerations for Using the Data ### Social Impact of Dataset The purpose of this dataset is to help develop better hate speech detection systems. A system that succeeds at this task would be able to identify hate speech tweets associated with the classes available in the dataset. ### Discussion of Biases An effort was made to reduce annotation bias by selecting annotators with a diverse demographic background. In terms of data collection, by using keywords and user mentions, we are introducing some bias to the data, restricting our scope to the list of keywords and users we created. ### Other Known Limitations Because of the massive data skew for the multilabel classes, it is extremely hard to train a robust model for this version of the dataset. We advise using it for analysis and experimentation only. The binary version of the dataset is robust enough to train a classifier with up to 76% F1-score. ## Additional Information ### Dataset Curators The dataset was created by João Augusto Leite, Diego Furtado Silva, both from the Federal University of São Carlos (BR), Carolina Scarton and Kalina Bontcheva both from the University of Sheffield (UK) ### Licensing Information ToLD-Br is licensed under a Creative Commons BY-SA 4.0 ### Citation Information ``` @article{DBLP:journals/corr/abs-2010-04543, author = {Joao Augusto Leite and Diego F. Silva and Kalina Bontcheva and Carolina Scarton}, title = {Toxic Language Detection in Social Media for Brazilian Portuguese: New Dataset and Multilingual Analysis}, journal = {CoRR}, volume = {abs/2010.04543}, year = {2020}, url = {https://arxiv.org/abs/2010.04543}, eprinttype = {arXiv}, eprint = {2010.04543}, timestamp = {Tue, 15 Dec 2020 16:10:16 +0100}, biburl = {https://dblp.org/rec/journals/corr/abs-2010-04543.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ``` ### Contributions Thanks to [@JAugusto97](https://github.com/JAugusto97) for adding this dataset.
told-br
[ "task_categories:text-classification", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:pt", "license:cc-by-sa-4.0", "hate-speech-detection", "arxiv:2010.04543", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["pt"], "license": ["cc-by-sa-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": [], "paperswithcode_id": "told-br", "pretty_name": "ToLD-Br", "language_bcp47": ["pt-BR"], "tags": ["hate-speech-detection"], "dataset_info": [{"config_name": "multilabel", "features": [{"name": "text", "dtype": "string"}, {"name": "homophobia", "dtype": {"class_label": {"names": {"0": "zero_votes", "1": "one_vote", "2": "two_votes", "3": "three_votes"}}}}, {"name": "obscene", "dtype": {"class_label": {"names": {"0": "zero_votes", "1": "one_vote", "2": "two_votes", "3": "three_votes"}}}}, {"name": "insult", "dtype": {"class_label": {"names": {"0": "zero_votes", "1": "one_vote", "2": "two_votes", "3": "three_votes"}}}}, {"name": "racism", "dtype": {"class_label": {"names": {"0": "zero_votes", "1": "one_vote", "2": "two_votes", "3": "three_votes"}}}}, {"name": "misogyny", "dtype": {"class_label": {"names": {"0": "zero_votes", "1": "one_vote", "2": "two_votes", "3": "three_votes"}}}}, {"name": "xenophobia", "dtype": {"class_label": {"names": {"0": "zero_votes", "1": "one_vote", "2": "two_votes", "3": "three_votes"}}}}], "splits": [{"name": "train", "num_bytes": 2978006, "num_examples": 21000}], "download_size": 2430416, "dataset_size": 2978006}, {"config_name": "binary", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "not-toxic", "1": "toxic"}}}}], "splits": [{"name": "train", "num_bytes": 1709560, "num_examples": 16800}, {"name": "test", "num_bytes": 216297, "num_examples": 2100}, {"name": "validation", "num_bytes": 212153, "num_examples": 2100}], "download_size": 853322, "dataset_size": 2138010}]}
2024-01-18T11:17:17+00:00
[ "2010.04543" ]
[ "pt" ]
TAGS #task_categories-text-classification #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Portuguese #license-cc-by-sa-4.0 #hate-speech-detection #arxiv-2010.04543 #region-us
Dataset Card for "ToLD-Br" ========================== Table of Contents ----------------- * Dataset Card for "ToLD-Br" + Table of Contents + Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages + Dataset Structure - Data Instances - Data Fields - Data Splits + Dataset Creation - Curation Rationale - Source Data * Initial Data Collection and Normalization * Who are the source language producers? - Annotations * Annotation process * Who are the annotators? - Personal and Sensitive Information + Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations + Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions Dataset Description ------------------- * Homepage: URL * Repository: URL * Paper: URL * Leaderboard: URL * Point of Contact: URL@URL ### Dataset Summary ToLD-Br is the biggest dataset for toxic tweets in Brazilian Portuguese, crowdsourced by 42 annotators selected from a pool of 129 volunteers. Annotators were selected aiming to create a plural group in terms of demographics (ethnicity, sexual orientation, age, gender). Each tweet was labeled by three annotators in 6 possible categories: LGBTQ+phobia, Xenophobia, Obscene, Insult, Misogyny and Racism. ### Supported Tasks and Leaderboards -'text-classification-other-hate-speech-detection': The dataset can be used to train a model for Hate Speech Detection, either using it's multi-label classes or by grouping them into a binary Hate vs. Non-Hate class. A BERT model can be fine-tuned to perform this task and achieve 0.75 F1-Score for it's binary version. ### Languages The text in the dataset is in Brazilian Portuguese, as spoken by Tweet users. The associated BCP-47 code is 'pt-BR'. Dataset Structure ----------------- ### Data Instances ToLD-Br has two versions: binary and multilabel. Multilabel: A data point consists of the tweet text (string) followed by 6 categories that have values ranging from 0 to 3, meaning the amount of votes from annotators for that specific class on homophobia, obscene, insult, racism, misogyny and xenophobia. An example from multilabel ToLD-Br looks as follows: Binary: A data point consists of the tweet text (string) followed by a binary class "toxic" with values 0 or 1. An example from binary ToLD-Br looks as follows: ### Data Fields Multilabel: * text: A string representing the tweet posted by a user. Mentions to other users are anonymized by replacing the mention with a @user tag. * homophobia: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as homophobic. * obscene: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as obscene. * insult: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as insult. * racism: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as racism. * misogyny: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as misogyny. * xenophobia: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as xenophobia. Binary: * text: A string representing the tweet posted by a user. Mentions to other users are anonymized by replacing the mention with a @user tag. * label: numerical binary value {0, 1} representing if the respective text is toxic/abusive or not. ### Data Splits Multilabel: The entire dataset consists of 21.000 examples. Binary: The train set consists of 16.800 examples, validation set consists of 2.100 examples and test set consists of 2.100 examples. Dataset Creation ---------------- ### Curation Rationale Despite Portuguese being the 5th most spoken language in the world and Brazil being the 4th country with most unique users, Brazilian Portuguese was underrepresented in the hate-speech detection task. Only two other datasets were available, one of them being European Portuguese. ToLD-Br is 4x bigger than both these datasets combined. Also, none of them had multiple annotators per instance. Also, this work proposes a plural and diverse group of annotators carefully selected to avoid inserting bias into the annotation. ### Source Data #### Initial Data Collection and Normalization Data was collected in 15 days in August 2019 using Gate Cloud's Tweet Collector. Ten million tweets were collected using two methods: a keyword-based method and a user-mention method. The first method collected tweets mentioning the following keywords: viado,veado,viadinho,veadinho,viadao,veadao,bicha,bixa,bichinha,bixinha,bichona,bixona,baitola,sapatão,sapatao,traveco,bambi,biba,boiola,marica,gayzão,gayzao,flor,florzinha,vagabundo,vagaba,desgraçada,desgraçado,desgracado,arrombado,arrombada,foder,fuder,fudido,fodido,cú,cu,pinto,pau,pal,caralho,caraio,carai,pica,cacete,rola,porra,escroto,buceta,fdp,pqp,vsf,tnc,vtnc,puto,putinho,acéfalo,acefalo,burro,idiota,trouxa,estúpido,estupido,estúpida,canalha,demente,retardado,retardada,verme,maldito,maldita,ridículo,ridiculo,ridícula,ridicula,morfético,morfetico,morfética,morfetica,lazarento,lazarenta,lixo,mongolóide,mongoloide,mongol,asqueroso,asquerosa,cretino,cretina,babaca,pilantra,neguinho,neguinha,pretinho,pretinha,escurinho,escurinha,pretinha,pretinho,crioulo,criolo,crioula,criola,macaco,macaca,gorila,puta,vagabunda,vagaba,mulherzinha,piranha,feminazi,putinha,piriguete,vaca,putinha,bahiano,baiano,baianagem,xingling,xing ling,xing-ling,carioca,paulista,sulista,mineiro,gringo The list of most followed Brazilian Twitter accounts can be found here. #### Who are the source language producers? The language producers are Twitter users from Brazil, speakers of Portuguese. ### Annotations #### Annotation process A form was published at the Federal University of São Carlos asking for volunteers to annotate our dataset. 129 people volunteered and 42 were selected according to their demographics in order to create a diverse and plural annotation group. Guidelines were produced and presented to the annotators. The entire process was done asynchronously because of the Covid-19 pandemic. The tool used was Google Sheets. Annotators were grouped into 14 teams of three annotators each. Each group annotated a respective file containing 1500 tweets. Annotators didn't have contact with each other, nor did they know that other annotators were labelling the same tweets as they were. #### Who are the annotators? Annotators were people from the Federal University of São Carlos' Facebook group. Their demographics are described below: Ages range from 18 to 37 years old. Annotators were paid R$50 ($10) to label 1500 examples each. ### Personal and Sensitive Information The dataset contains sensitive information for homophobia, obscene, insult, racism, misogyny and xenophobia. Tweets were anonymized by replacing user mentions with a @user tag. Considerations for Using the Data --------------------------------- ### Social Impact of Dataset The purpose of this dataset is to help develop better hate speech detection systems. A system that succeeds at this task would be able to identify hate speech tweets associated with the classes available in the dataset. ### Discussion of Biases An effort was made to reduce annotation bias by selecting annotators with a diverse demographic background. In terms of data collection, by using keywords and user mentions, we are introducing some bias to the data, restricting our scope to the list of keywords and users we created. ### Other Known Limitations Because of the massive data skew for the multilabel classes, it is extremely hard to train a robust model for this version of the dataset. We advise using it for analysis and experimentation only. The binary version of the dataset is robust enough to train a classifier with up to 76% F1-score. Additional Information ---------------------- ### Dataset Curators The dataset was created by João Augusto Leite, Diego Furtado Silva, both from the Federal University of São Carlos (BR), Carolina Scarton and Kalina Bontcheva both from the University of Sheffield (UK) ### Licensing Information ToLD-Br is licensed under a Creative Commons BY-SA 4.0 ### Contributions Thanks to @JAugusto97 for adding this dataset.
[ "### Dataset Summary\n\n\nToLD-Br is the biggest dataset for toxic tweets in Brazilian Portuguese, crowdsourced by 42 annotators selected from a pool of 129 volunteers. Annotators were selected aiming to create a plural group in terms of demographics (ethnicity, sexual orientation, age, gender). Each tweet was labeled by three annotators in 6 possible categories: LGBTQ+phobia, Xenophobia, Obscene, Insult, Misogyny and Racism.", "### Supported Tasks and Leaderboards\n\n\n-'text-classification-other-hate-speech-detection': The dataset can be used to train a model for Hate Speech Detection, either using it's multi-label classes or by grouping them into a binary Hate vs. Non-Hate class. A BERT model can be fine-tuned to perform this task and achieve 0.75 F1-Score for it's binary version.", "### Languages\n\n\nThe text in the dataset is in Brazilian Portuguese, as spoken by Tweet users. The associated BCP-47 code is 'pt-BR'.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nToLD-Br has two versions: binary and multilabel.\n\n\nMultilabel:\nA data point consists of the tweet text (string) followed by 6 categories that have values ranging from 0 to 3, meaning the amount of votes from annotators for that specific class on homophobia, obscene, insult, racism, misogyny and xenophobia.\n\n\nAn example from multilabel ToLD-Br looks as follows:\n\n\nBinary: \n\nA data point consists of the tweet text (string) followed by a binary class \"toxic\" with values 0 or 1.\n\n\nAn example from binary ToLD-Br looks as follows:", "### Data Fields\n\n\nMultilabel:\n\n\n* text: A string representing the tweet posted by a user. Mentions to other users are anonymized by replacing the mention with a @user tag.\n* homophobia: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as homophobic.\n* obscene: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as obscene.\n* insult: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as insult.\n* racism: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as racism.\n* misogyny: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as misogyny.\n* xenophobia: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as xenophobia.\n\n\nBinary:\n\n\n* text: A string representing the tweet posted by a user. Mentions to other users are anonymized by replacing the mention with a @user tag.\n* label: numerical binary value {0, 1} representing if the respective text is toxic/abusive or not.", "### Data Splits\n\n\nMultilabel: \n\nThe entire dataset consists of 21.000 examples.\n\n\nBinary: \n\nThe train set consists of 16.800 examples, validation set consists of 2.100 examples and test set consists of 2.100 examples.\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nDespite Portuguese being the 5th most spoken language in the world and Brazil being the 4th country with most unique users, Brazilian Portuguese was underrepresented in the hate-speech detection task. Only two other datasets were available, one of them being European Portuguese. ToLD-Br is 4x bigger than both these datasets combined. Also, none of them had multiple annotators per instance. Also, this work proposes a plural and diverse group of annotators carefully selected to avoid inserting bias into the annotation.", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nData was collected in 15 days in August 2019 using Gate Cloud's Tweet Collector. Ten million tweets were collected using two methods: a keyword-based method and a user-mention method. The first method collected tweets mentioning the following keywords:\n\n\nviado,veado,viadinho,veadinho,viadao,veadao,bicha,bixa,bichinha,bixinha,bichona,bixona,baitola,sapatão,sapatao,traveco,bambi,biba,boiola,marica,gayzão,gayzao,flor,florzinha,vagabundo,vagaba,desgraçada,desgraçado,desgracado,arrombado,arrombada,foder,fuder,fudido,fodido,cú,cu,pinto,pau,pal,caralho,caraio,carai,pica,cacete,rola,porra,escroto,buceta,fdp,pqp,vsf,tnc,vtnc,puto,putinho,acéfalo,acefalo,burro,idiota,trouxa,estúpido,estupido,estúpida,canalha,demente,retardado,retardada,verme,maldito,maldita,ridículo,ridiculo,ridícula,ridicula,morfético,morfetico,morfética,morfetica,lazarento,lazarenta,lixo,mongolóide,mongoloide,mongol,asqueroso,asquerosa,cretino,cretina,babaca,pilantra,neguinho,neguinha,pretinho,pretinha,escurinho,escurinha,pretinha,pretinho,crioulo,criolo,crioula,criola,macaco,macaca,gorila,puta,vagabunda,vagaba,mulherzinha,piranha,feminazi,putinha,piriguete,vaca,putinha,bahiano,baiano,baianagem,xingling,xing ling,xing-ling,carioca,paulista,sulista,mineiro,gringo\n\n\nThe list of most followed Brazilian Twitter accounts can be found here.", "#### Who are the source language producers?\n\n\nThe language producers are Twitter users from Brazil, speakers of Portuguese.", "### Annotations", "#### Annotation process\n\n\nA form was published at the Federal University of São Carlos asking for volunteers to annotate our dataset. 129 people volunteered and 42 were selected according to their demographics in order to create a diverse and plural annotation group. Guidelines were produced and presented to the annotators. The entire process was done asynchronously because of the Covid-19 pandemic. The tool used was Google Sheets. Annotators were grouped into 14 teams of three annotators each. Each group annotated a respective file containing 1500 tweets. Annotators didn't have contact with each other, nor did they know that other annotators were labelling the same tweets as they were.", "#### Who are the annotators?\n\n\nAnnotators were people from the Federal University of São Carlos' Facebook group. Their demographics are described below:\n\n\n\n\n\nAges range from 18 to 37 years old.\n\n\nAnnotators were paid R$50 ($10) to label 1500 examples each.", "### Personal and Sensitive Information\n\n\nThe dataset contains sensitive information for homophobia, obscene, insult, racism, misogyny and xenophobia.\n\n\nTweets were anonymized by replacing user mentions with a @user tag.\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset\n\n\nThe purpose of this dataset is to help develop better hate speech detection systems.\n\n\nA system that succeeds at this task would be able to identify hate speech tweets associated with the classes available in the dataset.", "### Discussion of Biases\n\n\nAn effort was made to reduce annotation bias by selecting annotators with a diverse demographic background. In terms of data collection, by using keywords and user mentions, we are introducing some bias to the data, restricting our scope to the list of keywords and users we created.", "### Other Known Limitations\n\n\nBecause of the massive data skew for the multilabel classes, it is extremely hard to train a robust model for this version of the dataset. We advise using it for analysis and experimentation only. The binary version of the dataset is robust enough to train a classifier with up to 76% F1-score.\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThe dataset was created by João Augusto Leite, Diego Furtado Silva, both from the Federal University of São Carlos (BR), Carolina Scarton and Kalina Bontcheva both from the University of Sheffield (UK)", "### Licensing Information\n\n\nToLD-Br is licensed under a Creative Commons BY-SA 4.0", "### Contributions\n\n\nThanks to @JAugusto97 for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Portuguese #license-cc-by-sa-4.0 #hate-speech-detection #arxiv-2010.04543 #region-us \n", "### Dataset Summary\n\n\nToLD-Br is the biggest dataset for toxic tweets in Brazilian Portuguese, crowdsourced by 42 annotators selected from a pool of 129 volunteers. Annotators were selected aiming to create a plural group in terms of demographics (ethnicity, sexual orientation, age, gender). Each tweet was labeled by three annotators in 6 possible categories: LGBTQ+phobia, Xenophobia, Obscene, Insult, Misogyny and Racism.", "### Supported Tasks and Leaderboards\n\n\n-'text-classification-other-hate-speech-detection': The dataset can be used to train a model for Hate Speech Detection, either using it's multi-label classes or by grouping them into a binary Hate vs. Non-Hate class. A BERT model can be fine-tuned to perform this task and achieve 0.75 F1-Score for it's binary version.", "### Languages\n\n\nThe text in the dataset is in Brazilian Portuguese, as spoken by Tweet users. The associated BCP-47 code is 'pt-BR'.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nToLD-Br has two versions: binary and multilabel.\n\n\nMultilabel:\nA data point consists of the tweet text (string) followed by 6 categories that have values ranging from 0 to 3, meaning the amount of votes from annotators for that specific class on homophobia, obscene, insult, racism, misogyny and xenophobia.\n\n\nAn example from multilabel ToLD-Br looks as follows:\n\n\nBinary: \n\nA data point consists of the tweet text (string) followed by a binary class \"toxic\" with values 0 or 1.\n\n\nAn example from binary ToLD-Br looks as follows:", "### Data Fields\n\n\nMultilabel:\n\n\n* text: A string representing the tweet posted by a user. Mentions to other users are anonymized by replacing the mention with a @user tag.\n* homophobia: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as homophobic.\n* obscene: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as obscene.\n* insult: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as insult.\n* racism: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as racism.\n* misogyny: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as misogyny.\n* xenophobia: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as xenophobia.\n\n\nBinary:\n\n\n* text: A string representing the tweet posted by a user. Mentions to other users are anonymized by replacing the mention with a @user tag.\n* label: numerical binary value {0, 1} representing if the respective text is toxic/abusive or not.", "### Data Splits\n\n\nMultilabel: \n\nThe entire dataset consists of 21.000 examples.\n\n\nBinary: \n\nThe train set consists of 16.800 examples, validation set consists of 2.100 examples and test set consists of 2.100 examples.\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nDespite Portuguese being the 5th most spoken language in the world and Brazil being the 4th country with most unique users, Brazilian Portuguese was underrepresented in the hate-speech detection task. Only two other datasets were available, one of them being European Portuguese. ToLD-Br is 4x bigger than both these datasets combined. Also, none of them had multiple annotators per instance. Also, this work proposes a plural and diverse group of annotators carefully selected to avoid inserting bias into the annotation.", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nData was collected in 15 days in August 2019 using Gate Cloud's Tweet Collector. Ten million tweets were collected using two methods: a keyword-based method and a user-mention method. The first method collected tweets mentioning the following keywords:\n\n\nviado,veado,viadinho,veadinho,viadao,veadao,bicha,bixa,bichinha,bixinha,bichona,bixona,baitola,sapatão,sapatao,traveco,bambi,biba,boiola,marica,gayzão,gayzao,flor,florzinha,vagabundo,vagaba,desgraçada,desgraçado,desgracado,arrombado,arrombada,foder,fuder,fudido,fodido,cú,cu,pinto,pau,pal,caralho,caraio,carai,pica,cacete,rola,porra,escroto,buceta,fdp,pqp,vsf,tnc,vtnc,puto,putinho,acéfalo,acefalo,burro,idiota,trouxa,estúpido,estupido,estúpida,canalha,demente,retardado,retardada,verme,maldito,maldita,ridículo,ridiculo,ridícula,ridicula,morfético,morfetico,morfética,morfetica,lazarento,lazarenta,lixo,mongolóide,mongoloide,mongol,asqueroso,asquerosa,cretino,cretina,babaca,pilantra,neguinho,neguinha,pretinho,pretinha,escurinho,escurinha,pretinha,pretinho,crioulo,criolo,crioula,criola,macaco,macaca,gorila,puta,vagabunda,vagaba,mulherzinha,piranha,feminazi,putinha,piriguete,vaca,putinha,bahiano,baiano,baianagem,xingling,xing ling,xing-ling,carioca,paulista,sulista,mineiro,gringo\n\n\nThe list of most followed Brazilian Twitter accounts can be found here.", "#### Who are the source language producers?\n\n\nThe language producers are Twitter users from Brazil, speakers of Portuguese.", "### Annotations", "#### Annotation process\n\n\nA form was published at the Federal University of São Carlos asking for volunteers to annotate our dataset. 129 people volunteered and 42 were selected according to their demographics in order to create a diverse and plural annotation group. Guidelines were produced and presented to the annotators. The entire process was done asynchronously because of the Covid-19 pandemic. The tool used was Google Sheets. Annotators were grouped into 14 teams of three annotators each. Each group annotated a respective file containing 1500 tweets. Annotators didn't have contact with each other, nor did they know that other annotators were labelling the same tweets as they were.", "#### Who are the annotators?\n\n\nAnnotators were people from the Federal University of São Carlos' Facebook group. Their demographics are described below:\n\n\n\n\n\nAges range from 18 to 37 years old.\n\n\nAnnotators were paid R$50 ($10) to label 1500 examples each.", "### Personal and Sensitive Information\n\n\nThe dataset contains sensitive information for homophobia, obscene, insult, racism, misogyny and xenophobia.\n\n\nTweets were anonymized by replacing user mentions with a @user tag.\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset\n\n\nThe purpose of this dataset is to help develop better hate speech detection systems.\n\n\nA system that succeeds at this task would be able to identify hate speech tweets associated with the classes available in the dataset.", "### Discussion of Biases\n\n\nAn effort was made to reduce annotation bias by selecting annotators with a diverse demographic background. In terms of data collection, by using keywords and user mentions, we are introducing some bias to the data, restricting our scope to the list of keywords and users we created.", "### Other Known Limitations\n\n\nBecause of the massive data skew for the multilabel classes, it is extremely hard to train a robust model for this version of the dataset. We advise using it for analysis and experimentation only. The binary version of the dataset is robust enough to train a classifier with up to 76% F1-score.\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThe dataset was created by João Augusto Leite, Diego Furtado Silva, both from the Federal University of São Carlos (BR), Carolina Scarton and Kalina Bontcheva both from the University of Sheffield (UK)", "### Licensing Information\n\n\nToLD-Br is licensed under a Creative Commons BY-SA 4.0", "### Contributions\n\n\nThanks to @JAugusto97 for adding this dataset." ]
[ 103, 119, 105, 45, 149, 333, 61, 133, 4, 517, 27, 5, 154, 62, 67, 51, 73, 81, 55, 22, 18 ]
[ "passage: TAGS\n#task_categories-text-classification #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Portuguese #license-cc-by-sa-4.0 #hate-speech-detection #arxiv-2010.04543 #region-us \n### Dataset Summary\n\n\nToLD-Br is the biggest dataset for toxic tweets in Brazilian Portuguese, crowdsourced by 42 annotators selected from a pool of 129 volunteers. Annotators were selected aiming to create a plural group in terms of demographics (ethnicity, sexual orientation, age, gender). Each tweet was labeled by three annotators in 6 possible categories: LGBTQ+phobia, Xenophobia, Obscene, Insult, Misogyny and Racism.### Supported Tasks and Leaderboards\n\n\n-'text-classification-other-hate-speech-detection': The dataset can be used to train a model for Hate Speech Detection, either using it's multi-label classes or by grouping them into a binary Hate vs. Non-Hate class. A BERT model can be fine-tuned to perform this task and achieve 0.75 F1-Score for it's binary version.### Languages\n\n\nThe text in the dataset is in Brazilian Portuguese, as spoken by Tweet users. The associated BCP-47 code is 'pt-BR'.\n\n\nDataset Structure\n-----------------", "passage: ### Data Instances\n\n\nToLD-Br has two versions: binary and multilabel.\n\n\nMultilabel:\nA data point consists of the tweet text (string) followed by 6 categories that have values ranging from 0 to 3, meaning the amount of votes from annotators for that specific class on homophobia, obscene, insult, racism, misogyny and xenophobia.\n\n\nAn example from multilabel ToLD-Br looks as follows:\n\n\nBinary: \n\nA data point consists of the tweet text (string) followed by a binary class \"toxic\" with values 0 or 1.\n\n\nAn example from binary ToLD-Br looks as follows:### Data Fields\n\n\nMultilabel:\n\n\n* text: A string representing the tweet posted by a user. Mentions to other users are anonymized by replacing the mention with a @user tag.\n* homophobia: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as homophobic.\n* obscene: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as obscene.\n* insult: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as insult.\n* racism: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as racism.\n* misogyny: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as misogyny.\n* xenophobia: numerical value {0, 1, 2, 3) representing the number of votes given by annotators flagging the respective tweet as xenophobia.\n\n\nBinary:\n\n\n* text: A string representing the tweet posted by a user. Mentions to other users are anonymized by replacing the mention with a @user tag.\n* label: numerical binary value {0, 1} representing if the respective text is toxic/abusive or not.### Data Splits\n\n\nMultilabel: \n\nThe entire dataset consists of 21.000 examples.\n\n\nBinary: \n\nThe train set consists of 16.800 examples, validation set consists of 2.100 examples and test set consists of 2.100 examples.\n\n\nDataset Creation\n----------------", "passage: ### Curation Rationale\n\n\nDespite Portuguese being the 5th most spoken language in the world and Brazil being the 4th country with most unique users, Brazilian Portuguese was underrepresented in the hate-speech detection task. Only two other datasets were available, one of them being European Portuguese. ToLD-Br is 4x bigger than both these datasets combined. Also, none of them had multiple annotators per instance. Also, this work proposes a plural and diverse group of annotators carefully selected to avoid inserting bias into the annotation.### Source Data", "passage: #### Initial Data Collection and Normalization\n\n\nData was collected in 15 days in August 2019 using Gate Cloud's Tweet Collector. Ten million tweets were collected using two methods: a keyword-based method and a user-mention method. The first method collected tweets mentioning the following keywords:\n\n\nviado,veado,viadinho,veadinho,viadao,veadao,bicha,bixa,bichinha,bixinha,bichona,bixona,baitola,sapatão,sapatao,traveco,bambi,biba,boiola,marica,gayzão,gayzao,flor,florzinha,vagabundo,vagaba,desgraçada,desgraçado,desgracado,arrombado,arrombada,foder,fuder,fudido,fodido,cú,cu,pinto,pau,pal,caralho,caraio,carai,pica,cacete,rola,porra,escroto,buceta,fdp,pqp,vsf,tnc,vtnc,puto,putinho,acéfalo,acefalo,burro,idiota,trouxa,estúpido,estupido,estúpida,canalha,demente,retardado,retardada,verme,maldito,maldita,ridículo,ridiculo,ridícula,ridicula,morfético,morfetico,morfética,morfetica,lazarento,lazarenta,lixo,mongolóide,mongoloide,mongol,asqueroso,asquerosa,cretino,cretina,babaca,pilantra,neguinho,neguinha,pretinho,pretinha,escurinho,escurinha,pretinha,pretinho,crioulo,criolo,crioula,criola,macaco,macaca,gorila,puta,vagabunda,vagaba,mulherzinha,piranha,feminazi,putinha,piriguete,vaca,putinha,bahiano,baiano,baianagem,xingling,xing ling,xing-ling,carioca,paulista,sulista,mineiro,gringo\n\n\nThe list of most followed Brazilian Twitter accounts can be found here.#### Who are the source language producers?\n\n\nThe language producers are Twitter users from Brazil, speakers of Portuguese.### Annotations#### Annotation process\n\n\nA form was published at the Federal University of São Carlos asking for volunteers to annotate our dataset. 129 people volunteered and 42 were selected according to their demographics in order to create a diverse and plural annotation group. Guidelines were produced and presented to the annotators. The entire process was done asynchronously because of the Covid-19 pandemic. The tool used was Google Sheets. Annotators were grouped into 14 teams of three annotators each. Each group annotated a respective file containing 1500 tweets. Annotators didn't have contact with each other, nor did they know that other annotators were labelling the same tweets as they were.#### Who are the annotators?\n\n\nAnnotators were people from the Federal University of São Carlos' Facebook group. Their demographics are described below:\n\n\n\n\n\nAges range from 18 to 37 years old.\n\n\nAnnotators were paid R$50 ($10) to label 1500 examples each.### Personal and Sensitive Information\n\n\nThe dataset contains sensitive information for homophobia, obscene, insult, racism, misogyny and xenophobia.\n\n\nTweets were anonymized by replacing user mentions with a @user tag.\n\n\nConsiderations for Using the Data\n---------------------------------### Social Impact of Dataset\n\n\nThe purpose of this dataset is to help develop better hate speech detection systems.\n\n\nA system that succeeds at this task would be able to identify hate speech tweets associated with the classes available in the dataset.### Discussion of Biases\n\n\nAn effort was made to reduce annotation bias by selecting annotators with a diverse demographic background. In terms of data collection, by using keywords and user mentions, we are introducing some bias to the data, restricting our scope to the list of keywords and users we created." ]
c118342026b0dd3974093d50d1ab3f67be05a2bf
# Dataset Card for ToTTo ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** None - **Repository:** https://github.com/google-research-datasets/ToTTo - **Paper:** https://arxiv.org/abs/2004.14373 - **Leaderboard:** https://github.com/google-research-datasets/ToTTo#leaderboard - **Point of Contact:** [totto@google.com](mailto:totto@google.com) ### Dataset Summary ToTTo is an open-domain English table-to-text dataset with over 120,000 training examples that proposes a controlled generation task: given a Wikipedia table and a set of highlighted table cells, produce a one-sentence description. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances A sample training set is provided below ``` {'example_id': '1762238357686640028', 'highlighted_cells': [[13, 2]], 'id': 0, 'overlap_subset': 'none', 'sentence_annotations': {'final_sentence': ['A Favorita is the telenovela aired in the 9 pm timeslot.'], 'original_sentence': ['It is also the first telenovela by the writer to air in the 9 pm timeslot.'], 'sentence_after_ambiguity': ['A Favorita is the telenovela aired in the 9 pm timeslot.'], 'sentence_after_deletion': ['It is the telenovela air in the 9 pm timeslot.']}, 'table': [[{'column_span': 1, 'is_header': True, 'row_span': 1, 'value': '#'}, {'column_span': 1, 'is_header': True, 'row_span': 1, 'value': 'Run'}, {'column_span': 1, 'is_header': True, 'row_span': 1, 'value': 'Title'}, {'column_span': 1, 'is_header': True, 'row_span': 1, 'value': 'Chapters'}, {'column_span': 1, 'is_header': True, 'row_span': 1, 'value': 'Author'}, {'column_span': 1, 'is_header': True, 'row_span': 1, 'value': 'Director'}, {'column_span': 1, 'is_header': True, 'row_span': 1, 'value': 'Ibope Rating'}], [{'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '59'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'June 5, 2000— February 2, 2001'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Laços de Família'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '209'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Manoel Carlos'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Ricardo Waddington'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '44.9'}], [{'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '60'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'February 5, 2001— September 28, 2001'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Porto dos Milagres'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '203'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Aguinaldo Silva Ricardo Linhares'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Marcos Paulo Simões'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '44.6'}], [{'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '61'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'October 1, 2001— June 14, 2002'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'O Clone'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '221'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Glória Perez'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Jayme Monjardim'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '47.0'}], [{'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '62'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'June 17, 2002— February 14, 2003'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Esperança'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '209'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Benedito Ruy Barbosa'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Luiz Fernando'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '37.7'}], [{'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '63'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'February 17, 2003— October 10, 2003'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Mulheres Apaixonadas'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '203'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Manoel Carlos'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Ricardo Waddington'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '46.6'}], [{'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '64'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'October 13, 2003— June 25, 2004'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Celebridade'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '221'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Gilberto Braga'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Dennis Carvalho'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '46.0'}], [{'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '65'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'June 28, 2004— March 11, 2005'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Senhora do Destino'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '221'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Aguinaldo Silva'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Wolf Maya'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '50.4'}], [{'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '66'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'March 14, 2005— November 4, 2005'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'América'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '203'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Glória Perez'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Jayme Monjardim Marcos Schechtman'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '49.4'}], [{'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '67'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'November 7, 2005— July 7, 2006'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Belíssima'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '209'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Sílvio de Abreu'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Denise Saraceni'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '48.5'}], [{'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '68'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'July 10, 2006— March 2, 2007'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Páginas da Vida'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '203'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Manoel Carlos'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Jayme Monjardim'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '46.8'}], [{'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '69'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'March 5, 2007— September 28, 2007'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Paraíso Tropical'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '179'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Gilberto Braga Ricardo Linhares'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Dennis Carvalho'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '42.8'}], [{'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '70'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'October 1, 2007— May 31, 2008'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Duas Caras'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '210'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Aguinaldo Silva'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Wolf Maya'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '41.1'}], [{'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '71'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'June 2, 2008— January 16, 2009'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'A Favorita'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '197'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'João Emanuel Carneiro'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Ricardo Waddington'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '39.5'}], [{'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '72'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'January 19, 2009— September 11, 2009'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Caminho das Índias'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '203'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Glória Perez'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Marcos Schechtman'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '38.8'}], [{'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '73'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'September 14, 2009— May 14, 2010'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Viver a Vida'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '209'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Manoel Carlos'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Jayme Monjardim'}, {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '35.6'}]], 'table_page_title': 'List of 8/9 PM telenovelas of Rede Globo', 'table_section_text': '', 'table_section_title': '2000s', 'table_webpage_url': 'http://en.wikipedia.org/wiki/List_of_8/9_PM_telenovelas_of_Rede_Globo'} ``` Please note that in test set sentence annotations are not available and thus values inside `sentence_annotations` can be safely ignored. ### Data Fields - `table_webpage_url` (`str`): Table webpage URL. - `table_page_title` (`str`): Table metadata with context about the table. - `table_section_title` (`str`): Table metadata with context about the table. - `table_section_text` (`str`): Table metadata with context about the table. - `table` (`List[List[Dict]]`): The outer lists represents rows and the inner lists columns. Each Dict has the fields: - `column_span` (`int`) - `is_header` (`bool`) - `row_span` (`int`) - `value` (`str`) - `highlighted_cells` (`List[[row_index, column_index]]`): Where each `[row_index, column_index]` pair indicates that `table[row_index][column_index]` is highlighted. - `example_id` (`int`): A unique id for this example. - `sentence_annotations`: Consists of the `original_sentence` and the sequence of revised sentences performed in order to produce the `final_sentence`. ### Data Splits ``` DatasetDict({ train: Dataset({ features: ['id', 'table_page_title', 'table_webpage_url', 'table_section_title', 'table_section_text', 'table', 'highlighted_cells', 'example_id', 'sentence_annotations', 'overlap_subset'], num_rows: 120761 }) validation: Dataset({ features: ['id', 'table_page_title', 'table_webpage_url', 'table_section_title', 'table_section_text', 'table', 'highlighted_cells', 'example_id', 'sentence_annotations', 'overlap_subset'], num_rows: 7700 }) test: Dataset({ features: ['id', 'table_page_title', 'table_webpage_url', 'table_section_title', 'table_section_text', 'table', 'highlighted_cells', 'example_id', 'sentence_annotations', 'overlap_subset'], num_rows: 7700 }) }) ``` ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ``` @inproceedings{parikh2020totto, title={{ToTTo}: A Controlled Table-To-Text Generation Dataset}, author={Parikh, Ankur P and Wang, Xuezhi and Gehrmann, Sebastian and Faruqui, Manaal and Dhingra, Bhuwan and Yang, Diyi and Das, Dipanjan}, booktitle={Proceedings of EMNLP}, year={2020} } ``` ### Contributions Thanks to [@abhishekkrthakur](https://github.com/abhishekkrthakur) for adding this dataset.
totto
[ "task_categories:table-to-text", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:en", "license:cc-by-sa-3.0", "arxiv:2004.14373", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found"], "language": ["en"], "license": ["cc-by-sa-3.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["table-to-text"], "task_ids": [], "paperswithcode_id": "totto", "pretty_name": "ToTTo", "dataset_info": {"features": [{"name": "id", "dtype": "int32"}, {"name": "table_page_title", "dtype": "string"}, {"name": "table_webpage_url", "dtype": "string"}, {"name": "table_section_title", "dtype": "string"}, {"name": "table_section_text", "dtype": "string"}, {"name": "table", "list": {"list": [{"name": "column_span", "dtype": "int32"}, {"name": "is_header", "dtype": "bool"}, {"name": "row_span", "dtype": "int32"}, {"name": "value", "dtype": "string"}]}}, {"name": "highlighted_cells", "sequence": {"sequence": "int32"}}, {"name": "example_id", "dtype": "string"}, {"name": "sentence_annotations", "sequence": [{"name": "original_sentence", "dtype": "string"}, {"name": "sentence_after_deletion", "dtype": "string"}, {"name": "sentence_after_ambiguity", "dtype": "string"}, {"name": "final_sentence", "dtype": "string"}]}, {"name": "overlap_subset", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 652754806, "num_examples": 120761}, {"name": "validation", "num_bytes": 47277039, "num_examples": 7700}, {"name": "test", "num_bytes": 40883586, "num_examples": 7700}], "download_size": 187724372, "dataset_size": 740915431}}
2024-01-18T11:17:18+00:00
[ "2004.14373" ]
[ "en" ]
TAGS #task_categories-table-to-text #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-sa-3.0 #arxiv-2004.14373 #region-us
# Dataset Card for ToTTo ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: None - Repository: URL - Paper: URL - Leaderboard: URL - Point of Contact: totto@URL ### Dataset Summary ToTTo is an open-domain English table-to-text dataset with over 120,000 training examples that proposes a controlled generation task: given a Wikipedia table and a set of highlighted table cells, produce a one-sentence description. ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances A sample training set is provided below Please note that in test set sentence annotations are not available and thus values inside 'sentence_annotations' can be safely ignored. ### Data Fields - 'table_webpage_url' ('str'): Table webpage URL. - 'table_page_title' ('str'): Table metadata with context about the table. - 'table_section_title' ('str'): Table metadata with context about the table. - 'table_section_text' ('str'): Table metadata with context about the table. - 'table' ('List[List[Dict]]'): The outer lists represents rows and the inner lists columns. Each Dict has the fields: - 'column_span' ('int') - 'is_header' ('bool') - 'row_span' ('int') - 'value' ('str') - 'highlighted_cells' ('List[[row_index, column_index]]'): Where each '[row_index, column_index]' pair indicates that 'table[row_index][column_index]' is highlighted. - 'example_id' ('int'): A unique id for this example. - 'sentence_annotations': Consists of the 'original_sentence' and the sequence of revised sentences performed in order to produce the 'final_sentence'. ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @abhishekkrthakur for adding this dataset.
[ "# Dataset Card for ToTTo", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: None\n- Repository: URL\n- Paper: URL\n- Leaderboard: URL\n- Point of Contact: totto@URL", "### Dataset Summary\n\nToTTo is an open-domain English table-to-text dataset with over 120,000 training examples that proposes a controlled \ngeneration task: given a Wikipedia table and a set of highlighted table cells, produce a one-sentence description.", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances\n\nA sample training set is provided below\n\n\n\nPlease note that in test set sentence annotations are not available and thus values inside 'sentence_annotations' can be safely ignored.", "### Data Fields\n\n- 'table_webpage_url' ('str'): Table webpage URL.\n- 'table_page_title' ('str'): Table metadata with context about the table.\n- 'table_section_title' ('str'): Table metadata with context about the table.\n- 'table_section_text' ('str'): Table metadata with context about the table.\n- 'table' ('List[List[Dict]]'): The outer lists represents rows and the inner lists columns. Each Dict has the fields:\n - 'column_span' ('int')\n - 'is_header' ('bool')\n - 'row_span' ('int')\n - 'value' ('str')\n- 'highlighted_cells' ('List[[row_index, column_index]]'): Where each '[row_index, column_index]' pair indicates that 'table[row_index][column_index]' is highlighted.\n- 'example_id' ('int'): A unique id for this example.\n- 'sentence_annotations': Consists of the 'original_sentence' and the sequence of revised sentences performed in order to produce the 'final_sentence'.", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ "TAGS\n#task_categories-table-to-text #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-sa-3.0 #arxiv-2004.14373 #region-us \n", "# Dataset Card for ToTTo", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: None\n- Repository: URL\n- Paper: URL\n- Leaderboard: URL\n- Point of Contact: totto@URL", "### Dataset Summary\n\nToTTo is an open-domain English table-to-text dataset with over 120,000 training examples that proposes a controlled \ngeneration task: given a Wikipedia table and a set of highlighted table cells, produce a one-sentence description.", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances\n\nA sample training set is provided below\n\n\n\nPlease note that in test set sentence annotations are not available and thus values inside 'sentence_annotations' can be safely ignored.", "### Data Fields\n\n- 'table_webpage_url' ('str'): Table webpage URL.\n- 'table_page_title' ('str'): Table metadata with context about the table.\n- 'table_section_title' ('str'): Table metadata with context about the table.\n- 'table_section_text' ('str'): Table metadata with context about the table.\n- 'table' ('List[List[Dict]]'): The outer lists represents rows and the inner lists columns. Each Dict has the fields:\n - 'column_span' ('int')\n - 'is_header' ('bool')\n - 'row_span' ('int')\n - 'value' ('str')\n- 'highlighted_cells' ('List[[row_index, column_index]]'): Where each '[row_index, column_index]' pair indicates that 'table[row_index][column_index]' is highlighted.\n- 'example_id' ('int'): A unique id for this example.\n- 'sentence_annotations': Consists of the 'original_sentence' and the sequence of revised sentences performed in order to produce the 'final_sentence'.", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ 91, 8, 120, 33, 61, 10, 4, 6, 46, 307, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 20 ]
[ "passage: TAGS\n#task_categories-table-to-text #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-sa-3.0 #arxiv-2004.14373 #region-us \n# Dataset Card for ToTTo## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: None\n- Repository: URL\n- Paper: URL\n- Leaderboard: URL\n- Point of Contact: totto@URL### Dataset Summary\n\nToTTo is an open-domain English table-to-text dataset with over 120,000 training examples that proposes a controlled \ngeneration task: given a Wikipedia table and a set of highlighted table cells, produce a one-sentence description.### Supported Tasks and Leaderboards### Languages## Dataset Structure### Data Instances\n\nA sample training set is provided below\n\n\n\nPlease note that in test set sentence annotations are not available and thus values inside 'sentence_annotations' can be safely ignored." ]
eb1e45c1ba990fecca7cf84b67ce845edbcf49bf
# Dataset Card for "trec" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://cogcomp.seas.upenn.edu/Data/QA/QC/](https://cogcomp.seas.upenn.edu/Data/QA/QC/) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 0.36 MB - **Size of the generated dataset:** 0.41 MB - **Total amount of disk used:** 0.78 MB ### Dataset Summary The Text REtrieval Conference (TREC) Question Classification dataset contains 5500 labeled questions in training set and another 500 for test set. The dataset has 6 coarse class labels and 50 fine class labels. Average length of each sentence is 10, vocabulary size of 8700. Data are collected from four sources: 4,500 English questions published by USC (Hovy et al., 2001), about 500 manually constructed questions for a few rare classes, 894 TREC 8 and TREC 9 questions, and also 500 questions from TREC 10 which serves as the test set. These questions were manually labeled. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages The language in this dataset is English (`en`). ## Dataset Structure ### Data Instances - **Size of downloaded dataset files:** 0.36 MB - **Size of the generated dataset:** 0.41 MB - **Total amount of disk used:** 0.78 MB An example of 'train' looks as follows. ``` { 'text': 'How did serfdom develop in and then leave Russia ?', 'coarse_label': 2, 'fine_label': 26 } ``` ### Data Fields The data fields are the same among all splits. - `text` (`str`): Text of the question. - `coarse_label` (`ClassLabel`): Coarse class label. Possible values are: - 'ABBR' (0): Abbreviation. - 'ENTY' (1): Entity. - 'DESC' (2): Description and abstract concept. - 'HUM' (3): Human being. - 'LOC' (4): Location. - 'NUM' (5): Numeric value. - `fine_label` (`ClassLabel`): Fine class label. Possible values are: - ABBREVIATION: - 'ABBR:abb' (0): Abbreviation. - 'ABBR:exp' (1): Expression abbreviated. - ENTITY: - 'ENTY:animal' (2): Animal. - 'ENTY:body' (3): Organ of body. - 'ENTY:color' (4): Color. - 'ENTY:cremat' (5): Invention, book and other creative piece. - 'ENTY:currency' (6): Currency name. - 'ENTY:dismed' (7): Disease and medicine. - 'ENTY:event' (8): Event. - 'ENTY:food' (9): Food. - 'ENTY:instru' (10): Musical instrument. - 'ENTY:lang' (11): Language. - 'ENTY:letter' (12): Letter like a-z. - 'ENTY:other' (13): Other entity. - 'ENTY:plant' (14): Plant. - 'ENTY:product' (15): Product. - 'ENTY:religion' (16): Religion. - 'ENTY:sport' (17): Sport. - 'ENTY:substance' (18): Element and substance. - 'ENTY:symbol' (19): Symbols and sign. - 'ENTY:techmeth' (20): Techniques and method. - 'ENTY:termeq' (21): Equivalent term. - 'ENTY:veh' (22): Vehicle. - 'ENTY:word' (23): Word with a special property. - DESCRIPTION: - 'DESC:def' (24): Definition of something. - 'DESC:desc' (25): Description of something. - 'DESC:manner' (26): Manner of an action. - 'DESC:reason' (27): Reason. - HUMAN: - 'HUM:gr' (28): Group or organization of persons - 'HUM:ind' (29): Individual. - 'HUM:title' (30): Title of a person. - 'HUM:desc' (31): Description of a person. - LOCATION: - 'LOC:city' (32): City. - 'LOC:country' (33): Country. - 'LOC:mount' (34): Mountain. - 'LOC:other' (35): Other location. - 'LOC:state' (36): State. - NUMERIC: - 'NUM:code' (37): Postcode or other code. - 'NUM:count' (38): Number of something. - 'NUM:date' (39): Date. - 'NUM:dist' (40): Distance, linear measure. - 'NUM:money' (41): Price. - 'NUM:ord' (42): Order, rank. - 'NUM:other' (43): Other number. - 'NUM:period' (44): Lasting time of something - 'NUM:perc' (45): Percent, fraction. - 'NUM:speed' (46): Speed. - 'NUM:temp' (47): Temperature. - 'NUM:volsize' (48): Size, area and volume. - 'NUM:weight' (49): Weight. ### Data Splits | name | train | test | |---------|------:|-----:| | default | 5452 | 500 | ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @inproceedings{li-roth-2002-learning, title = "Learning Question Classifiers", author = "Li, Xin and Roth, Dan", booktitle = "{COLING} 2002: The 19th International Conference on Computational Linguistics", year = "2002", url = "https://www.aclweb.org/anthology/C02-1150", } @inproceedings{hovy-etal-2001-toward, title = "Toward Semantics-Based Answer Pinpointing", author = "Hovy, Eduard and Gerber, Laurie and Hermjakob, Ulf and Lin, Chin-Yew and Ravichandran, Deepak", booktitle = "Proceedings of the First International Conference on Human Language Technology Research", year = "2001", url = "https://www.aclweb.org/anthology/H01-1069", } ``` ### Contributions Thanks to [@lhoestq](https://github.com/lhoestq), [@thomwolf](https://github.com/thomwolf) for adding this dataset.
trec
[ "task_categories:text-classification", "task_ids:multi-class-classification", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["expert-generated"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["multi-class-classification"], "paperswithcode_id": "trecqa", "pretty_name": "Text Retrieval Conference Question Answering", "dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "coarse_label", "dtype": {"class_label": {"names": {"0": "ABBR", "1": "ENTY", "2": "DESC", "3": "HUM", "4": "LOC", "5": "NUM"}}}}, {"name": "fine_label", "dtype": {"class_label": {"names": {"0": "ABBR:abb", "1": "ABBR:exp", "2": "ENTY:animal", "3": "ENTY:body", "4": "ENTY:color", "5": "ENTY:cremat", "6": "ENTY:currency", "7": "ENTY:dismed", "8": "ENTY:event", "9": "ENTY:food", "10": "ENTY:instru", "11": "ENTY:lang", "12": "ENTY:letter", "13": "ENTY:other", "14": "ENTY:plant", "15": "ENTY:product", "16": "ENTY:religion", "17": "ENTY:sport", "18": "ENTY:substance", "19": "ENTY:symbol", "20": "ENTY:techmeth", "21": "ENTY:termeq", "22": "ENTY:veh", "23": "ENTY:word", "24": "DESC:def", "25": "DESC:desc", "26": "DESC:manner", "27": "DESC:reason", "28": "HUM:gr", "29": "HUM:ind", "30": "HUM:title", "31": "HUM:desc", "32": "LOC:city", "33": "LOC:country", "34": "LOC:mount", "35": "LOC:other", "36": "LOC:state", "37": "NUM:code", "38": "NUM:count", "39": "NUM:date", "40": "NUM:dist", "41": "NUM:money", "42": "NUM:ord", "43": "NUM:other", "44": "NUM:period", "45": "NUM:perc", "46": "NUM:speed", "47": "NUM:temp", "48": "NUM:volsize", "49": "NUM:weight"}}}}], "splits": [{"name": "train", "num_bytes": 385090, "num_examples": 5452}, {"name": "test", "num_bytes": 27983, "num_examples": 500}], "download_size": 359212, "dataset_size": 413073}}
2024-01-18T11:17:19+00:00
[]
[ "en" ]
TAGS #task_categories-text-classification #task_ids-multi-class-classification #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-unknown #region-us
Dataset Card for "trec" ======================= Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: * Point of Contact: * Size of downloaded dataset files: 0.36 MB * Size of the generated dataset: 0.41 MB * Total amount of disk used: 0.78 MB ### Dataset Summary The Text REtrieval Conference (TREC) Question Classification dataset contains 5500 labeled questions in training set and another 500 for test set. The dataset has 6 coarse class labels and 50 fine class labels. Average length of each sentence is 10, vocabulary size of 8700. Data are collected from four sources: 4,500 English questions published by USC (Hovy et al., 2001), about 500 manually constructed questions for a few rare classes, 894 TREC 8 and TREC 9 questions, and also 500 questions from TREC 10 which serves as the test set. These questions were manually labeled. ### Supported Tasks and Leaderboards ### Languages The language in this dataset is English ('en'). Dataset Structure ----------------- ### Data Instances * Size of downloaded dataset files: 0.36 MB * Size of the generated dataset: 0.41 MB * Total amount of disk used: 0.78 MB An example of 'train' looks as follows. ### Data Fields The data fields are the same among all splits. * 'text' ('str'): Text of the question. * 'coarse\_label' ('ClassLabel'): Coarse class label. Possible values are: + 'ABBR' (0): Abbreviation. + 'ENTY' (1): Entity. + 'DESC' (2): Description and abstract concept. + 'HUM' (3): Human being. + 'LOC' (4): Location. + 'NUM' (5): Numeric value. * 'fine\_label' ('ClassLabel'): Fine class label. Possible values are: + ABBREVIATION: - 'ABBR:abb' (0): Abbreviation. - 'ABBR:exp' (1): Expression abbreviated. + ENTITY: - 'ENTY:animal' (2): Animal. - 'ENTY:body' (3): Organ of body. - 'ENTY:color' (4): Color. - 'ENTY:cremat' (5): Invention, book and other creative piece. - 'ENTY:currency' (6): Currency name. - 'ENTY:dismed' (7): Disease and medicine. - 'ENTY:event' (8): Event. - 'ENTY:food' (9): Food. - 'ENTY:instru' (10): Musical instrument. - 'ENTY:lang' (11): Language. - 'ENTY:letter' (12): Letter like a-z. - 'ENTY:other' (13): Other entity. - 'ENTY:plant' (14): Plant. - 'ENTY:product' (15): Product. - 'ENTY:religion' (16): Religion. - 'ENTY:sport' (17): Sport. - 'ENTY:substance' (18): Element and substance. - 'ENTY:symbol' (19): Symbols and sign. - 'ENTY:techmeth' (20): Techniques and method. - 'ENTY:termeq' (21): Equivalent term. - 'ENTY:veh' (22): Vehicle. - 'ENTY:word' (23): Word with a special property. + DESCRIPTION: - 'DESC:def' (24): Definition of something. - 'DESC:desc' (25): Description of something. - 'DESC:manner' (26): Manner of an action. - 'DESC:reason' (27): Reason. + HUMAN: - 'HUM:gr' (28): Group or organization of persons - 'HUM:ind' (29): Individual. - 'HUM:title' (30): Title of a person. - 'HUM:desc' (31): Description of a person. + LOCATION: - 'LOC:city' (32): City. - 'LOC:country' (33): Country. - 'LOC:mount' (34): Mountain. - 'LOC:other' (35): Other location. - 'LOC:state' (36): State. + NUMERIC: - 'NUM:code' (37): Postcode or other code. - 'NUM:count' (38): Number of something. - 'NUM:date' (39): Date. - 'NUM:dist' (40): Distance, linear measure. - 'NUM:money' (41): Price. - 'NUM:ord' (42): Order, rank. - 'NUM:other' (43): Other number. - 'NUM:period' (44): Lasting time of something - 'NUM:perc' (45): Percent, fraction. - 'NUM:speed' (46): Speed. - 'NUM:temp' (47): Temperature. - 'NUM:volsize' (48): Size, area and volume. - 'NUM:weight' (49): Weight. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @lhoestq, @thomwolf for adding this dataset.
[ "### Dataset Summary\n\n\nThe Text REtrieval Conference (TREC) Question Classification dataset contains 5500 labeled questions in training set and another 500 for test set.\n\n\nThe dataset has 6 coarse class labels and 50 fine class labels. Average length of each sentence is 10, vocabulary size of 8700.\n\n\nData are collected from four sources: 4,500 English questions published by USC (Hovy et al., 2001), about 500 manually constructed questions for a few rare classes, 894 TREC 8 and TREC 9 questions, and also 500 questions from TREC 10 which serves as the test set. These questions were manually labeled.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe language in this dataset is English ('en').\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\n* Size of downloaded dataset files: 0.36 MB\n* Size of the generated dataset: 0.41 MB\n* Total amount of disk used: 0.78 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.\n\n\n* 'text' ('str'): Text of the question.\n* 'coarse\\_label' ('ClassLabel'): Coarse class label. Possible values are:\n\t+ 'ABBR' (0): Abbreviation.\n\t+ 'ENTY' (1): Entity.\n\t+ 'DESC' (2): Description and abstract concept.\n\t+ 'HUM' (3): Human being.\n\t+ 'LOC' (4): Location.\n\t+ 'NUM' (5): Numeric value.\n* 'fine\\_label' ('ClassLabel'): Fine class label. Possible values are:\n\t+ ABBREVIATION:\n\t\t- 'ABBR:abb' (0): Abbreviation.\n\t\t- 'ABBR:exp' (1): Expression abbreviated.\n\t+ ENTITY:\n\t\t- 'ENTY:animal' (2): Animal.\n\t\t- 'ENTY:body' (3): Organ of body.\n\t\t- 'ENTY:color' (4): Color.\n\t\t- 'ENTY:cremat' (5): Invention, book and other creative piece.\n\t\t- 'ENTY:currency' (6): Currency name.\n\t\t- 'ENTY:dismed' (7): Disease and medicine.\n\t\t- 'ENTY:event' (8): Event.\n\t\t- 'ENTY:food' (9): Food.\n\t\t- 'ENTY:instru' (10): Musical instrument.\n\t\t- 'ENTY:lang' (11): Language.\n\t\t- 'ENTY:letter' (12): Letter like a-z.\n\t\t- 'ENTY:other' (13): Other entity.\n\t\t- 'ENTY:plant' (14): Plant.\n\t\t- 'ENTY:product' (15): Product.\n\t\t- 'ENTY:religion' (16): Religion.\n\t\t- 'ENTY:sport' (17): Sport.\n\t\t- 'ENTY:substance' (18): Element and substance.\n\t\t- 'ENTY:symbol' (19): Symbols and sign.\n\t\t- 'ENTY:techmeth' (20): Techniques and method.\n\t\t- 'ENTY:termeq' (21): Equivalent term.\n\t\t- 'ENTY:veh' (22): Vehicle.\n\t\t- 'ENTY:word' (23): Word with a special property.\n\t+ DESCRIPTION:\n\t\t- 'DESC:def' (24): Definition of something.\n\t\t- 'DESC:desc' (25): Description of something.\n\t\t- 'DESC:manner' (26): Manner of an action.\n\t\t- 'DESC:reason' (27): Reason.\n\t+ HUMAN:\n\t\t- 'HUM:gr' (28): Group or organization of persons\n\t\t- 'HUM:ind' (29): Individual.\n\t\t- 'HUM:title' (30): Title of a person.\n\t\t- 'HUM:desc' (31): Description of a person.\n\t+ LOCATION:\n\t\t- 'LOC:city' (32): City.\n\t\t- 'LOC:country' (33): Country.\n\t\t- 'LOC:mount' (34): Mountain.\n\t\t- 'LOC:other' (35): Other location.\n\t\t- 'LOC:state' (36): State.\n\t+ NUMERIC:\n\t\t- 'NUM:code' (37): Postcode or other code.\n\t\t- 'NUM:count' (38): Number of something.\n\t\t- 'NUM:date' (39): Date.\n\t\t- 'NUM:dist' (40): Distance, linear measure.\n\t\t- 'NUM:money' (41): Price.\n\t\t- 'NUM:ord' (42): Order, rank.\n\t\t- 'NUM:other' (43): Other number.\n\t\t- 'NUM:period' (44): Lasting time of something\n\t\t- 'NUM:perc' (45): Percent, fraction.\n\t\t- 'NUM:speed' (46): Speed.\n\t\t- 'NUM:temp' (47): Temperature.\n\t\t- 'NUM:volsize' (48): Size, area and volume.\n\t\t- 'NUM:weight' (49): Weight.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @lhoestq, @thomwolf for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #task_ids-multi-class-classification #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-unknown #region-us \n", "### Dataset Summary\n\n\nThe Text REtrieval Conference (TREC) Question Classification dataset contains 5500 labeled questions in training set and another 500 for test set.\n\n\nThe dataset has 6 coarse class labels and 50 fine class labels. Average length of each sentence is 10, vocabulary size of 8700.\n\n\nData are collected from four sources: 4,500 English questions published by USC (Hovy et al., 2001), about 500 manually constructed questions for a few rare classes, 894 TREC 8 and TREC 9 questions, and also 500 questions from TREC 10 which serves as the test set. These questions were manually labeled.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe language in this dataset is English ('en').\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\n* Size of downloaded dataset files: 0.36 MB\n* Size of the generated dataset: 0.41 MB\n* Total amount of disk used: 0.78 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.\n\n\n* 'text' ('str'): Text of the question.\n* 'coarse\\_label' ('ClassLabel'): Coarse class label. Possible values are:\n\t+ 'ABBR' (0): Abbreviation.\n\t+ 'ENTY' (1): Entity.\n\t+ 'DESC' (2): Description and abstract concept.\n\t+ 'HUM' (3): Human being.\n\t+ 'LOC' (4): Location.\n\t+ 'NUM' (5): Numeric value.\n* 'fine\\_label' ('ClassLabel'): Fine class label. Possible values are:\n\t+ ABBREVIATION:\n\t\t- 'ABBR:abb' (0): Abbreviation.\n\t\t- 'ABBR:exp' (1): Expression abbreviated.\n\t+ ENTITY:\n\t\t- 'ENTY:animal' (2): Animal.\n\t\t- 'ENTY:body' (3): Organ of body.\n\t\t- 'ENTY:color' (4): Color.\n\t\t- 'ENTY:cremat' (5): Invention, book and other creative piece.\n\t\t- 'ENTY:currency' (6): Currency name.\n\t\t- 'ENTY:dismed' (7): Disease and medicine.\n\t\t- 'ENTY:event' (8): Event.\n\t\t- 'ENTY:food' (9): Food.\n\t\t- 'ENTY:instru' (10): Musical instrument.\n\t\t- 'ENTY:lang' (11): Language.\n\t\t- 'ENTY:letter' (12): Letter like a-z.\n\t\t- 'ENTY:other' (13): Other entity.\n\t\t- 'ENTY:plant' (14): Plant.\n\t\t- 'ENTY:product' (15): Product.\n\t\t- 'ENTY:religion' (16): Religion.\n\t\t- 'ENTY:sport' (17): Sport.\n\t\t- 'ENTY:substance' (18): Element and substance.\n\t\t- 'ENTY:symbol' (19): Symbols and sign.\n\t\t- 'ENTY:techmeth' (20): Techniques and method.\n\t\t- 'ENTY:termeq' (21): Equivalent term.\n\t\t- 'ENTY:veh' (22): Vehicle.\n\t\t- 'ENTY:word' (23): Word with a special property.\n\t+ DESCRIPTION:\n\t\t- 'DESC:def' (24): Definition of something.\n\t\t- 'DESC:desc' (25): Description of something.\n\t\t- 'DESC:manner' (26): Manner of an action.\n\t\t- 'DESC:reason' (27): Reason.\n\t+ HUMAN:\n\t\t- 'HUM:gr' (28): Group or organization of persons\n\t\t- 'HUM:ind' (29): Individual.\n\t\t- 'HUM:title' (30): Title of a person.\n\t\t- 'HUM:desc' (31): Description of a person.\n\t+ LOCATION:\n\t\t- 'LOC:city' (32): City.\n\t\t- 'LOC:country' (33): Country.\n\t\t- 'LOC:mount' (34): Mountain.\n\t\t- 'LOC:other' (35): Other location.\n\t\t- 'LOC:state' (36): State.\n\t+ NUMERIC:\n\t\t- 'NUM:code' (37): Postcode or other code.\n\t\t- 'NUM:count' (38): Number of something.\n\t\t- 'NUM:date' (39): Date.\n\t\t- 'NUM:dist' (40): Distance, linear measure.\n\t\t- 'NUM:money' (41): Price.\n\t\t- 'NUM:ord' (42): Order, rank.\n\t\t- 'NUM:other' (43): Other number.\n\t\t- 'NUM:period' (44): Lasting time of something\n\t\t- 'NUM:perc' (45): Percent, fraction.\n\t\t- 'NUM:speed' (46): Speed.\n\t\t- 'NUM:temp' (47): Temperature.\n\t\t- 'NUM:volsize' (48): Size, area and volume.\n\t\t- 'NUM:weight' (49): Weight.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @lhoestq, @thomwolf for adding this dataset." ]
[ 92, 145, 10, 24, 54, 816, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 6, 22 ]
[ "passage: TAGS\n#task_categories-text-classification #task_ids-multi-class-classification #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-unknown #region-us \n### Dataset Summary\n\n\nThe Text REtrieval Conference (TREC) Question Classification dataset contains 5500 labeled questions in training set and another 500 for test set.\n\n\nThe dataset has 6 coarse class labels and 50 fine class labels. Average length of each sentence is 10, vocabulary size of 8700.\n\n\nData are collected from four sources: 4,500 English questions published by USC (Hovy et al., 2001), about 500 manually constructed questions for a few rare classes, 894 TREC 8 and TREC 9 questions, and also 500 questions from TREC 10 which serves as the test set. These questions were manually labeled.### Supported Tasks and Leaderboards### Languages\n\n\nThe language in this dataset is English ('en').\n\n\nDataset Structure\n-----------------### Data Instances\n\n\n* Size of downloaded dataset files: 0.36 MB\n* Size of the generated dataset: 0.41 MB\n* Total amount of disk used: 0.78 MB\n\n\nAn example of 'train' looks as follows." ]
0f7faf33a3908546c6fd5b73a660e0f8ff173c2f
# Dataset Card for "trivia_qa" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [http://nlp.cs.washington.edu/triviaqa/](http://nlp.cs.washington.edu/triviaqa/) - **Repository:** [https://github.com/mandarjoshi90/triviaqa](https://github.com/mandarjoshi90/triviaqa) - **Paper:** [TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension](https://arxiv.org/abs/1705.03551) - **Leaderboard:** [CodaLab Leaderboard](https://competitions.codalab.org/competitions/17208#results) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 9.26 GB - **Size of the generated dataset:** 45.46 GB - **Total amount of disk used:** 54.72 GB ### Dataset Summary TriviaqQA is a reading comprehension dataset containing over 650K question-answer-evidence triples. TriviaqQA includes 95K question-answer pairs authored by trivia enthusiasts and independently gathered evidence documents, six per question on average, that provide high quality distant supervision for answering the questions. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages English. ## Dataset Structure ### Data Instances #### rc - **Size of downloaded dataset files:** 2.67 GB - **Size of the generated dataset:** 16.02 GB - **Total amount of disk used:** 18.68 GB An example of 'train' looks as follows. ``` ``` #### rc.nocontext - **Size of downloaded dataset files:** 2.67 GB - **Size of the generated dataset:** 126.27 MB - **Total amount of disk used:** 2.79 GB An example of 'train' looks as follows. ``` ``` #### unfiltered - **Size of downloaded dataset files:** 3.30 GB - **Size of the generated dataset:** 29.24 GB - **Total amount of disk used:** 32.54 GB An example of 'validation' looks as follows. ``` ``` #### unfiltered.nocontext - **Size of downloaded dataset files:** 632.55 MB - **Size of the generated dataset:** 74.56 MB - **Total amount of disk used:** 707.11 MB An example of 'train' looks as follows. ``` ``` ### Data Fields The data fields are the same among all splits. #### rc - `question`: a `string` feature. - `question_id`: a `string` feature. - `question_source`: a `string` feature. - `entity_pages`: a dictionary feature containing: - `doc_source`: a `string` feature. - `filename`: a `string` feature. - `title`: a `string` feature. - `wiki_context`: a `string` feature. - `search_results`: a dictionary feature containing: - `description`: a `string` feature. - `filename`: a `string` feature. - `rank`: a `int32` feature. - `title`: a `string` feature. - `url`: a `string` feature. - `search_context`: a `string` feature. - `aliases`: a `list` of `string` features. - `normalized_aliases`: a `list` of `string` features. - `matched_wiki_entity_name`: a `string` feature. - `normalized_matched_wiki_entity_name`: a `string` feature. - `normalized_value`: a `string` feature. - `type`: a `string` feature. - `value`: a `string` feature. #### rc.nocontext - `question`: a `string` feature. - `question_id`: a `string` feature. - `question_source`: a `string` feature. - `entity_pages`: a dictionary feature containing: - `doc_source`: a `string` feature. - `filename`: a `string` feature. - `title`: a `string` feature. - `wiki_context`: a `string` feature. - `search_results`: a dictionary feature containing: - `description`: a `string` feature. - `filename`: a `string` feature. - `rank`: a `int32` feature. - `title`: a `string` feature. - `url`: a `string` feature. - `search_context`: a `string` feature. - `aliases`: a `list` of `string` features. - `normalized_aliases`: a `list` of `string` features. - `matched_wiki_entity_name`: a `string` feature. - `normalized_matched_wiki_entity_name`: a `string` feature. - `normalized_value`: a `string` feature. - `type`: a `string` feature. - `value`: a `string` feature. #### unfiltered - `question`: a `string` feature. - `question_id`: a `string` feature. - `question_source`: a `string` feature. - `entity_pages`: a dictionary feature containing: - `doc_source`: a `string` feature. - `filename`: a `string` feature. - `title`: a `string` feature. - `wiki_context`: a `string` feature. - `search_results`: a dictionary feature containing: - `description`: a `string` feature. - `filename`: a `string` feature. - `rank`: a `int32` feature. - `title`: a `string` feature. - `url`: a `string` feature. - `search_context`: a `string` feature. - `aliases`: a `list` of `string` features. - `normalized_aliases`: a `list` of `string` features. - `matched_wiki_entity_name`: a `string` feature. - `normalized_matched_wiki_entity_name`: a `string` feature. - `normalized_value`: a `string` feature. - `type`: a `string` feature. - `value`: a `string` feature. #### unfiltered.nocontext - `question`: a `string` feature. - `question_id`: a `string` feature. - `question_source`: a `string` feature. - `entity_pages`: a dictionary feature containing: - `doc_source`: a `string` feature. - `filename`: a `string` feature. - `title`: a `string` feature. - `wiki_context`: a `string` feature. - `search_results`: a dictionary feature containing: - `description`: a `string` feature. - `filename`: a `string` feature. - `rank`: a `int32` feature. - `title`: a `string` feature. - `url`: a `string` feature. - `search_context`: a `string` feature. - `aliases`: a `list` of `string` features. - `normalized_aliases`: a `list` of `string` features. - `matched_wiki_entity_name`: a `string` feature. - `normalized_matched_wiki_entity_name`: a `string` feature. - `normalized_value`: a `string` feature. - `type`: a `string` feature. - `value`: a `string` feature. ### Data Splits | name |train |validation|test | |--------------------|-----:|---------:|----:| |rc |138384| 18669|17210| |rc.nocontext |138384| 18669|17210| |unfiltered | 87622| 11313|10832| |unfiltered.nocontext| 87622| 11313|10832| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information The University of Washington does not own the copyright of the questions and documents included in TriviaQA. ### Citation Information ``` @article{2017arXivtriviaqa, author = {{Joshi}, Mandar and {Choi}, Eunsol and {Weld}, Daniel and {Zettlemoyer}, Luke}, title = "{triviaqa: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension}", journal = {arXiv e-prints}, year = 2017, eid = {arXiv:1705.03551}, pages = {arXiv:1705.03551}, archivePrefix = {arXiv}, eprint = {1705.03551}, } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@patrickvonplaten](https://github.com/patrickvonplaten), [@lewtun](https://github.com/lewtun) for adding this dataset.
trivia_qa
[ "task_categories:question-answering", "task_categories:text2text-generation", "task_ids:open-domain-qa", "task_ids:open-domain-abstractive-qa", "task_ids:extractive-qa", "task_ids:abstractive-qa", "annotations_creators:crowdsourced", "language_creators:machine-generated", "multilinguality:monolingual", "size_categories:10K<n<100K", "size_categories:100K<n<1M", "source_datasets:original", "language:en", "license:unknown", "arxiv:1705.03551", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["machine-generated"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K", "100K<n<1M"], "source_datasets": ["original"], "task_categories": ["question-answering", "text2text-generation"], "task_ids": ["open-domain-qa", "open-domain-abstractive-qa", "extractive-qa", "abstractive-qa"], "paperswithcode_id": "triviaqa", "pretty_name": "TriviaQA", "dataset_info": [{"config_name": "rc", "features": [{"name": "question", "dtype": "string"}, {"name": "question_id", "dtype": "string"}, {"name": "question_source", "dtype": "string"}, {"name": "entity_pages", "sequence": [{"name": "doc_source", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "wiki_context", "dtype": "string"}]}, {"name": "search_results", "sequence": [{"name": "description", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "rank", "dtype": "int32"}, {"name": "title", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "search_context", "dtype": "string"}]}, {"name": "answer", "struct": [{"name": "aliases", "sequence": "string"}, {"name": "normalized_aliases", "sequence": "string"}, {"name": "matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_value", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "value", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 12749651131, "num_examples": 138384}, {"name": "validation", "num_bytes": 1662321188, "num_examples": 17944}, {"name": "test", "num_bytes": 1577710503, "num_examples": 17210}], "download_size": 8998808983, "dataset_size": 15989682822}, {"config_name": "rc.nocontext", "features": [{"name": "question", "dtype": "string"}, {"name": "question_id", "dtype": "string"}, {"name": "question_source", "dtype": "string"}, {"name": "entity_pages", "sequence": [{"name": "doc_source", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "wiki_context", "dtype": "string"}]}, {"name": "search_results", "sequence": [{"name": "description", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "rank", "dtype": "int32"}, {"name": "title", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "search_context", "dtype": "string"}]}, {"name": "answer", "struct": [{"name": "aliases", "sequence": "string"}, {"name": "normalized_aliases", "sequence": "string"}, {"name": "matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_value", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "value", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 106882730, "num_examples": 138384}, {"name": "validation", "num_bytes": 14059830, "num_examples": 17944}, {"name": "test", "num_bytes": 3667903, "num_examples": 17210}], "download_size": 63926518, "dataset_size": 124610463}, {"config_name": "rc.web", "features": [{"name": "question", "dtype": "string"}, {"name": "question_id", "dtype": "string"}, {"name": "question_source", "dtype": "string"}, {"name": "entity_pages", "sequence": [{"name": "doc_source", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "wiki_context", "dtype": "string"}]}, {"name": "search_results", "sequence": [{"name": "description", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "rank", "dtype": "int32"}, {"name": "title", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "search_context", "dtype": "string"}]}, {"name": "answer", "struct": [{"name": "aliases", "sequence": "string"}, {"name": "normalized_aliases", "sequence": "string"}, {"name": "matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_value", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "value", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 9408851139, "num_examples": 76496}, {"name": "validation", "num_bytes": 1232155138, "num_examples": 9951}, {"name": "test", "num_bytes": 1171663999, "num_examples": 9509}], "download_size": 6626625832, "dataset_size": 11812670276}, {"config_name": "rc.web.nocontext", "features": [{"name": "question", "dtype": "string"}, {"name": "question_id", "dtype": "string"}, {"name": "question_source", "dtype": "string"}, {"name": "entity_pages", "sequence": [{"name": "doc_source", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "wiki_context", "dtype": "string"}]}, {"name": "search_results", "sequence": [{"name": "description", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "rank", "dtype": "int32"}, {"name": "title", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "search_context", "dtype": "string"}]}, {"name": "answer", "struct": [{"name": "aliases", "sequence": "string"}, {"name": "normalized_aliases", "sequence": "string"}, {"name": "matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_value", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "value", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 58523085, "num_examples": 76496}, {"name": "validation", "num_bytes": 7694557, "num_examples": 9951}, {"name": "test", "num_bytes": 2024747, "num_examples": 9509}], "download_size": 35123473, "dataset_size": 68242389}, {"config_name": "rc.wikipedia", "features": [{"name": "question", "dtype": "string"}, {"name": "question_id", "dtype": "string"}, {"name": "question_source", "dtype": "string"}, {"name": "entity_pages", "sequence": [{"name": "doc_source", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "wiki_context", "dtype": "string"}]}, {"name": "search_results", "sequence": [{"name": "description", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "rank", "dtype": "int32"}, {"name": "title", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "search_context", "dtype": "string"}]}, {"name": "answer", "struct": [{"name": "aliases", "sequence": "string"}, {"name": "normalized_aliases", "sequence": "string"}, {"name": "matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_value", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "value", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 3340799992, "num_examples": 61888}, {"name": "validation", "num_bytes": 430166050, "num_examples": 7993}, {"name": "test", "num_bytes": 406046504, "num_examples": 7701}], "download_size": 2293374081, "dataset_size": 4177012546}, {"config_name": "rc.wikipedia.nocontext", "features": [{"name": "question", "dtype": "string"}, {"name": "question_id", "dtype": "string"}, {"name": "question_source", "dtype": "string"}, {"name": "entity_pages", "sequence": [{"name": "doc_source", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "wiki_context", "dtype": "string"}]}, {"name": "search_results", "sequence": [{"name": "description", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "rank", "dtype": "int32"}, {"name": "title", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "search_context", "dtype": "string"}]}, {"name": "answer", "struct": [{"name": "aliases", "sequence": "string"}, {"name": "normalized_aliases", "sequence": "string"}, {"name": "matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_value", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "value", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 48359645, "num_examples": 61888}, {"name": "validation", "num_bytes": 6365273, "num_examples": 7993}, {"name": "test", "num_bytes": 1643156, "num_examples": 7701}], "download_size": 28803950, "dataset_size": 56368074}, {"config_name": "unfiltered", "features": [{"name": "question", "dtype": "string"}, {"name": "question_id", "dtype": "string"}, {"name": "question_source", "dtype": "string"}, {"name": "entity_pages", "sequence": [{"name": "doc_source", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "wiki_context", "dtype": "string"}]}, {"name": "search_results", "sequence": [{"name": "description", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "rank", "dtype": "int32"}, {"name": "title", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "search_context", "dtype": "string"}]}, {"name": "answer", "struct": [{"name": "aliases", "sequence": "string"}, {"name": "normalized_aliases", "sequence": "string"}, {"name": "matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_value", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "value", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 23292199425, "num_examples": 87622}, {"name": "validation", "num_bytes": 3038803743, "num_examples": 11313}, {"name": "test", "num_bytes": 2906455311, "num_examples": 10832}], "download_size": 16695552268, "dataset_size": 29237458479}, {"config_name": "unfiltered.nocontext", "features": [{"name": "question", "dtype": "string"}, {"name": "question_id", "dtype": "string"}, {"name": "question_source", "dtype": "string"}, {"name": "entity_pages", "sequence": [{"name": "doc_source", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "wiki_context", "dtype": "string"}]}, {"name": "search_results", "sequence": [{"name": "description", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "rank", "dtype": "int32"}, {"name": "title", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "search_context", "dtype": "string"}]}, {"name": "answer", "struct": [{"name": "aliases", "sequence": "string"}, {"name": "normalized_aliases", "sequence": "string"}, {"name": "matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_value", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "value", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 63300226, "num_examples": 87622}, {"name": "validation", "num_bytes": 8296870, "num_examples": 11313}, {"name": "test", "num_bytes": 2320660, "num_examples": 10832}], "download_size": 38364033, "dataset_size": 73917756}, {"config_name": "unfiltered.web", "features": [{"name": "question", "dtype": "string"}, {"name": "question_id", "dtype": "string"}, {"name": "question_source", "dtype": "string"}, {"name": "entity_pages", "sequence": [{"name": "doc_source", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "wiki_context", "dtype": "string"}]}, {"name": "search_results", "sequence": [{"name": "description", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "rank", "dtype": "int32"}, {"name": "title", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "search_context", "dtype": "string"}]}, {"name": "answer", "struct": [{"name": "aliases", "sequence": "string"}, {"name": "normalized_aliases", "sequence": "string"}, {"name": "matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_value", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "value", "dtype": "string"}]}], "splits": [{"name": "train"}, {"name": "validation"}, {"name": "test"}], "download_size": 3298328560, "dataset_size": 0}, {"config_name": "unfiltered.web.nocontext", "features": [{"name": "question", "dtype": "string"}, {"name": "question_id", "dtype": "string"}, {"name": "question_source", "dtype": "string"}, {"name": "entity_pages", "sequence": [{"name": "doc_source", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "wiki_context", "dtype": "string"}]}, {"name": "search_results", "sequence": [{"name": "description", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "rank", "dtype": "int32"}, {"name": "title", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "search_context", "dtype": "string"}]}, {"name": "answer", "struct": [{"name": "aliases", "sequence": "string"}, {"name": "normalized_aliases", "sequence": "string"}, {"name": "matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_value", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "value", "dtype": "string"}]}], "splits": [{"name": "train"}, {"name": "validation"}, {"name": "test"}], "download_size": 632549060, "dataset_size": 0}, {"config_name": "unfiltered.wikipedia", "features": [{"name": "question", "dtype": "string"}, {"name": "question_id", "dtype": "string"}, {"name": "question_source", "dtype": "string"}, {"name": "entity_pages", "sequence": [{"name": "doc_source", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "wiki_context", "dtype": "string"}]}, {"name": "search_results", "sequence": [{"name": "description", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "rank", "dtype": "int32"}, {"name": "title", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "search_context", "dtype": "string"}]}, {"name": "answer", "struct": [{"name": "aliases", "sequence": "string"}, {"name": "normalized_aliases", "sequence": "string"}, {"name": "matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_value", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "value", "dtype": "string"}]}], "splits": [{"name": "train"}, {"name": "validation"}, {"name": "test"}], "download_size": 3298328560, "dataset_size": 0}, {"config_name": "unfiltered.wikipedia.nocontext", "features": [{"name": "question", "dtype": "string"}, {"name": "question_id", "dtype": "string"}, {"name": "question_source", "dtype": "string"}, {"name": "entity_pages", "sequence": [{"name": "doc_source", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "wiki_context", "dtype": "string"}]}, {"name": "search_results", "sequence": [{"name": "description", "dtype": "string"}, {"name": "filename", "dtype": "string"}, {"name": "rank", "dtype": "int32"}, {"name": "title", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "search_context", "dtype": "string"}]}, {"name": "answer", "struct": [{"name": "aliases", "sequence": "string"}, {"name": "normalized_aliases", "sequence": "string"}, {"name": "matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_matched_wiki_entity_name", "dtype": "string"}, {"name": "normalized_value", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "value", "dtype": "string"}]}], "splits": [{"name": "train"}, {"name": "validation"}, {"name": "test"}], "download_size": 632549060, "dataset_size": 0}], "configs": [{"config_name": "rc", "data_files": [{"split": "train", "path": "rc/train-*"}, {"split": "validation", "path": "rc/validation-*"}, {"split": "test", "path": "rc/test-*"}]}, {"config_name": "rc.nocontext", "data_files": [{"split": "train", "path": "rc.nocontext/train-*"}, {"split": "validation", "path": "rc.nocontext/validation-*"}, {"split": "test", "path": "rc.nocontext/test-*"}]}, {"config_name": "rc.web", "data_files": [{"split": "train", "path": "rc.web/train-*"}, {"split": "validation", "path": "rc.web/validation-*"}, {"split": "test", "path": "rc.web/test-*"}]}, {"config_name": "rc.web.nocontext", "data_files": [{"split": "train", "path": "rc.web.nocontext/train-*"}, {"split": "validation", "path": "rc.web.nocontext/validation-*"}, {"split": "test", "path": "rc.web.nocontext/test-*"}]}, {"config_name": "rc.wikipedia", "data_files": [{"split": "train", "path": "rc.wikipedia/train-*"}, {"split": "validation", "path": "rc.wikipedia/validation-*"}, {"split": "test", "path": "rc.wikipedia/test-*"}]}, {"config_name": "rc.wikipedia.nocontext", "data_files": [{"split": "train", "path": "rc.wikipedia.nocontext/train-*"}, {"split": "validation", "path": "rc.wikipedia.nocontext/validation-*"}, {"split": "test", "path": "rc.wikipedia.nocontext/test-*"}]}, {"config_name": "unfiltered", "data_files": [{"split": "train", "path": "unfiltered/train-*"}, {"split": "validation", "path": "unfiltered/validation-*"}, {"split": "test", "path": "unfiltered/test-*"}]}, {"config_name": "unfiltered.nocontext", "data_files": [{"split": "train", "path": "unfiltered.nocontext/train-*"}, {"split": "validation", "path": "unfiltered.nocontext/validation-*"}, {"split": "test", "path": "unfiltered.nocontext/test-*"}]}]}
2024-01-05T13:24:37+00:00
[ "1705.03551" ]
[ "en" ]
TAGS #task_categories-question-answering #task_categories-text2text-generation #task_ids-open-domain-qa #task_ids-open-domain-abstractive-qa #task_ids-extractive-qa #task_ids-abstractive-qa #annotations_creators-crowdsourced #language_creators-machine-generated #multilinguality-monolingual #size_categories-10K<n<100K #size_categories-100K<n<1M #source_datasets-original #language-English #license-unknown #arxiv-1705.03551 #region-us
Dataset Card for "trivia\_qa" ============================= Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: URL * Paper: TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension * Leaderboard: CodaLab Leaderboard * Point of Contact: * Size of downloaded dataset files: 9.26 GB * Size of the generated dataset: 45.46 GB * Total amount of disk used: 54.72 GB ### Dataset Summary TriviaqQA is a reading comprehension dataset containing over 650K question-answer-evidence triples. TriviaqQA includes 95K question-answer pairs authored by trivia enthusiasts and independently gathered evidence documents, six per question on average, that provide high quality distant supervision for answering the questions. ### Supported Tasks and Leaderboards ### Languages English. Dataset Structure ----------------- ### Data Instances #### rc * Size of downloaded dataset files: 2.67 GB * Size of the generated dataset: 16.02 GB * Total amount of disk used: 18.68 GB An example of 'train' looks as follows. #### rc.nocontext * Size of downloaded dataset files: 2.67 GB * Size of the generated dataset: 126.27 MB * Total amount of disk used: 2.79 GB An example of 'train' looks as follows. #### unfiltered * Size of downloaded dataset files: 3.30 GB * Size of the generated dataset: 29.24 GB * Total amount of disk used: 32.54 GB An example of 'validation' looks as follows. #### unfiltered.nocontext * Size of downloaded dataset files: 632.55 MB * Size of the generated dataset: 74.56 MB * Total amount of disk used: 707.11 MB An example of 'train' looks as follows. ### Data Fields The data fields are the same among all splits. #### rc * 'question': a 'string' feature. * 'question\_id': a 'string' feature. * 'question\_source': a 'string' feature. * 'entity\_pages': a dictionary feature containing: + 'doc\_source': a 'string' feature. + 'filename': a 'string' feature. + 'title': a 'string' feature. + 'wiki\_context': a 'string' feature. * 'search\_results': a dictionary feature containing: + 'description': a 'string' feature. + 'filename': a 'string' feature. + 'rank': a 'int32' feature. + 'title': a 'string' feature. + 'url': a 'string' feature. + 'search\_context': a 'string' feature. * 'aliases': a 'list' of 'string' features. * 'normalized\_aliases': a 'list' of 'string' features. * 'matched\_wiki\_entity\_name': a 'string' feature. * 'normalized\_matched\_wiki\_entity\_name': a 'string' feature. * 'normalized\_value': a 'string' feature. * 'type': a 'string' feature. * 'value': a 'string' feature. #### rc.nocontext * 'question': a 'string' feature. * 'question\_id': a 'string' feature. * 'question\_source': a 'string' feature. * 'entity\_pages': a dictionary feature containing: + 'doc\_source': a 'string' feature. + 'filename': a 'string' feature. + 'title': a 'string' feature. + 'wiki\_context': a 'string' feature. * 'search\_results': a dictionary feature containing: + 'description': a 'string' feature. + 'filename': a 'string' feature. + 'rank': a 'int32' feature. + 'title': a 'string' feature. + 'url': a 'string' feature. + 'search\_context': a 'string' feature. * 'aliases': a 'list' of 'string' features. * 'normalized\_aliases': a 'list' of 'string' features. * 'matched\_wiki\_entity\_name': a 'string' feature. * 'normalized\_matched\_wiki\_entity\_name': a 'string' feature. * 'normalized\_value': a 'string' feature. * 'type': a 'string' feature. * 'value': a 'string' feature. #### unfiltered * 'question': a 'string' feature. * 'question\_id': a 'string' feature. * 'question\_source': a 'string' feature. * 'entity\_pages': a dictionary feature containing: + 'doc\_source': a 'string' feature. + 'filename': a 'string' feature. + 'title': a 'string' feature. + 'wiki\_context': a 'string' feature. * 'search\_results': a dictionary feature containing: + 'description': a 'string' feature. + 'filename': a 'string' feature. + 'rank': a 'int32' feature. + 'title': a 'string' feature. + 'url': a 'string' feature. + 'search\_context': a 'string' feature. * 'aliases': a 'list' of 'string' features. * 'normalized\_aliases': a 'list' of 'string' features. * 'matched\_wiki\_entity\_name': a 'string' feature. * 'normalized\_matched\_wiki\_entity\_name': a 'string' feature. * 'normalized\_value': a 'string' feature. * 'type': a 'string' feature. * 'value': a 'string' feature. #### unfiltered.nocontext * 'question': a 'string' feature. * 'question\_id': a 'string' feature. * 'question\_source': a 'string' feature. * 'entity\_pages': a dictionary feature containing: + 'doc\_source': a 'string' feature. + 'filename': a 'string' feature. + 'title': a 'string' feature. + 'wiki\_context': a 'string' feature. * 'search\_results': a dictionary feature containing: + 'description': a 'string' feature. + 'filename': a 'string' feature. + 'rank': a 'int32' feature. + 'title': a 'string' feature. + 'url': a 'string' feature. + 'search\_context': a 'string' feature. * 'aliases': a 'list' of 'string' features. * 'normalized\_aliases': a 'list' of 'string' features. * 'matched\_wiki\_entity\_name': a 'string' feature. * 'normalized\_matched\_wiki\_entity\_name': a 'string' feature. * 'normalized\_value': a 'string' feature. * 'type': a 'string' feature. * 'value': a 'string' feature. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information The University of Washington does not own the copyright of the questions and documents included in TriviaQA. ### Contributions Thanks to @thomwolf, @patrickvonplaten, @lewtun for adding this dataset.
[ "### Dataset Summary\n\n\nTriviaqQA is a reading comprehension dataset containing over 650K\nquestion-answer-evidence triples. TriviaqQA includes 95K question-answer\npairs authored by trivia enthusiasts and independently gathered evidence\ndocuments, six per question on average, that provide high quality distant\nsupervision for answering the questions.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nEnglish.\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### rc\n\n\n* Size of downloaded dataset files: 2.67 GB\n* Size of the generated dataset: 16.02 GB\n* Total amount of disk used: 18.68 GB\n\n\nAn example of 'train' looks as follows.", "#### rc.nocontext\n\n\n* Size of downloaded dataset files: 2.67 GB\n* Size of the generated dataset: 126.27 MB\n* Total amount of disk used: 2.79 GB\n\n\nAn example of 'train' looks as follows.", "#### unfiltered\n\n\n* Size of downloaded dataset files: 3.30 GB\n* Size of the generated dataset: 29.24 GB\n* Total amount of disk used: 32.54 GB\n\n\nAn example of 'validation' looks as follows.", "#### unfiltered.nocontext\n\n\n* Size of downloaded dataset files: 632.55 MB\n* Size of the generated dataset: 74.56 MB\n* Total amount of disk used: 707.11 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### rc\n\n\n* 'question': a 'string' feature.\n* 'question\\_id': a 'string' feature.\n* 'question\\_source': a 'string' feature.\n* 'entity\\_pages': a dictionary feature containing:\n\t+ 'doc\\_source': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'wiki\\_context': a 'string' feature.\n* 'search\\_results': a dictionary feature containing:\n\t+ 'description': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'rank': a 'int32' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'url': a 'string' feature.\n\t+ 'search\\_context': a 'string' feature.\n* 'aliases': a 'list' of 'string' features.\n* 'normalized\\_aliases': a 'list' of 'string' features.\n* 'matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_value': a 'string' feature.\n* 'type': a 'string' feature.\n* 'value': a 'string' feature.", "#### rc.nocontext\n\n\n* 'question': a 'string' feature.\n* 'question\\_id': a 'string' feature.\n* 'question\\_source': a 'string' feature.\n* 'entity\\_pages': a dictionary feature containing:\n\t+ 'doc\\_source': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'wiki\\_context': a 'string' feature.\n* 'search\\_results': a dictionary feature containing:\n\t+ 'description': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'rank': a 'int32' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'url': a 'string' feature.\n\t+ 'search\\_context': a 'string' feature.\n* 'aliases': a 'list' of 'string' features.\n* 'normalized\\_aliases': a 'list' of 'string' features.\n* 'matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_value': a 'string' feature.\n* 'type': a 'string' feature.\n* 'value': a 'string' feature.", "#### unfiltered\n\n\n* 'question': a 'string' feature.\n* 'question\\_id': a 'string' feature.\n* 'question\\_source': a 'string' feature.\n* 'entity\\_pages': a dictionary feature containing:\n\t+ 'doc\\_source': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'wiki\\_context': a 'string' feature.\n* 'search\\_results': a dictionary feature containing:\n\t+ 'description': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'rank': a 'int32' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'url': a 'string' feature.\n\t+ 'search\\_context': a 'string' feature.\n* 'aliases': a 'list' of 'string' features.\n* 'normalized\\_aliases': a 'list' of 'string' features.\n* 'matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_value': a 'string' feature.\n* 'type': a 'string' feature.\n* 'value': a 'string' feature.", "#### unfiltered.nocontext\n\n\n* 'question': a 'string' feature.\n* 'question\\_id': a 'string' feature.\n* 'question\\_source': a 'string' feature.\n* 'entity\\_pages': a dictionary feature containing:\n\t+ 'doc\\_source': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'wiki\\_context': a 'string' feature.\n* 'search\\_results': a dictionary feature containing:\n\t+ 'description': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'rank': a 'int32' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'url': a 'string' feature.\n\t+ 'search\\_context': a 'string' feature.\n* 'aliases': a 'list' of 'string' features.\n* 'normalized\\_aliases': a 'list' of 'string' features.\n* 'matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_value': a 'string' feature.\n* 'type': a 'string' feature.\n* 'value': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nThe University of Washington does not own the copyright of the questions and documents included in TriviaQA.", "### Contributions\n\n\nThanks to @thomwolf, @patrickvonplaten, @lewtun for adding this dataset." ]
[ "TAGS\n#task_categories-question-answering #task_categories-text2text-generation #task_ids-open-domain-qa #task_ids-open-domain-abstractive-qa #task_ids-extractive-qa #task_ids-abstractive-qa #annotations_creators-crowdsourced #language_creators-machine-generated #multilinguality-monolingual #size_categories-10K<n<100K #size_categories-100K<n<1M #source_datasets-original #language-English #license-unknown #arxiv-1705.03551 #region-us \n", "### Dataset Summary\n\n\nTriviaqQA is a reading comprehension dataset containing over 650K\nquestion-answer-evidence triples. TriviaqQA includes 95K question-answer\npairs authored by trivia enthusiasts and independently gathered evidence\ndocuments, six per question on average, that provide high quality distant\nsupervision for answering the questions.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nEnglish.\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### rc\n\n\n* Size of downloaded dataset files: 2.67 GB\n* Size of the generated dataset: 16.02 GB\n* Total amount of disk used: 18.68 GB\n\n\nAn example of 'train' looks as follows.", "#### rc.nocontext\n\n\n* Size of downloaded dataset files: 2.67 GB\n* Size of the generated dataset: 126.27 MB\n* Total amount of disk used: 2.79 GB\n\n\nAn example of 'train' looks as follows.", "#### unfiltered\n\n\n* Size of downloaded dataset files: 3.30 GB\n* Size of the generated dataset: 29.24 GB\n* Total amount of disk used: 32.54 GB\n\n\nAn example of 'validation' looks as follows.", "#### unfiltered.nocontext\n\n\n* Size of downloaded dataset files: 632.55 MB\n* Size of the generated dataset: 74.56 MB\n* Total amount of disk used: 707.11 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### rc\n\n\n* 'question': a 'string' feature.\n* 'question\\_id': a 'string' feature.\n* 'question\\_source': a 'string' feature.\n* 'entity\\_pages': a dictionary feature containing:\n\t+ 'doc\\_source': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'wiki\\_context': a 'string' feature.\n* 'search\\_results': a dictionary feature containing:\n\t+ 'description': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'rank': a 'int32' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'url': a 'string' feature.\n\t+ 'search\\_context': a 'string' feature.\n* 'aliases': a 'list' of 'string' features.\n* 'normalized\\_aliases': a 'list' of 'string' features.\n* 'matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_value': a 'string' feature.\n* 'type': a 'string' feature.\n* 'value': a 'string' feature.", "#### rc.nocontext\n\n\n* 'question': a 'string' feature.\n* 'question\\_id': a 'string' feature.\n* 'question\\_source': a 'string' feature.\n* 'entity\\_pages': a dictionary feature containing:\n\t+ 'doc\\_source': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'wiki\\_context': a 'string' feature.\n* 'search\\_results': a dictionary feature containing:\n\t+ 'description': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'rank': a 'int32' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'url': a 'string' feature.\n\t+ 'search\\_context': a 'string' feature.\n* 'aliases': a 'list' of 'string' features.\n* 'normalized\\_aliases': a 'list' of 'string' features.\n* 'matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_value': a 'string' feature.\n* 'type': a 'string' feature.\n* 'value': a 'string' feature.", "#### unfiltered\n\n\n* 'question': a 'string' feature.\n* 'question\\_id': a 'string' feature.\n* 'question\\_source': a 'string' feature.\n* 'entity\\_pages': a dictionary feature containing:\n\t+ 'doc\\_source': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'wiki\\_context': a 'string' feature.\n* 'search\\_results': a dictionary feature containing:\n\t+ 'description': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'rank': a 'int32' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'url': a 'string' feature.\n\t+ 'search\\_context': a 'string' feature.\n* 'aliases': a 'list' of 'string' features.\n* 'normalized\\_aliases': a 'list' of 'string' features.\n* 'matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_value': a 'string' feature.\n* 'type': a 'string' feature.\n* 'value': a 'string' feature.", "#### unfiltered.nocontext\n\n\n* 'question': a 'string' feature.\n* 'question\\_id': a 'string' feature.\n* 'question\\_source': a 'string' feature.\n* 'entity\\_pages': a dictionary feature containing:\n\t+ 'doc\\_source': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'wiki\\_context': a 'string' feature.\n* 'search\\_results': a dictionary feature containing:\n\t+ 'description': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'rank': a 'int32' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'url': a 'string' feature.\n\t+ 'search\\_context': a 'string' feature.\n* 'aliases': a 'list' of 'string' features.\n* 'normalized\\_aliases': a 'list' of 'string' features.\n* 'matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_value': a 'string' feature.\n* 'type': a 'string' feature.\n* 'value': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nThe University of Washington does not own the copyright of the questions and documents included in TriviaQA.", "### Contributions\n\n\nThanks to @thomwolf, @patrickvonplaten, @lewtun for adding this dataset." ]
[ 165, 84, 10, 13, 6, 50, 55, 52, 59, 17, 330, 334, 331, 335, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 26, 28 ]
[ "passage: TAGS\n#task_categories-question-answering #task_categories-text2text-generation #task_ids-open-domain-qa #task_ids-open-domain-abstractive-qa #task_ids-extractive-qa #task_ids-abstractive-qa #annotations_creators-crowdsourced #language_creators-machine-generated #multilinguality-monolingual #size_categories-10K<n<100K #size_categories-100K<n<1M #source_datasets-original #language-English #license-unknown #arxiv-1705.03551 #region-us \n### Dataset Summary\n\n\nTriviaqQA is a reading comprehension dataset containing over 650K\nquestion-answer-evidence triples. TriviaqQA includes 95K question-answer\npairs authored by trivia enthusiasts and independently gathered evidence\ndocuments, six per question on average, that provide high quality distant\nsupervision for answering the questions.### Supported Tasks and Leaderboards### Languages\n\n\nEnglish.\n\n\nDataset Structure\n-----------------### Data Instances#### rc\n\n\n* Size of downloaded dataset files: 2.67 GB\n* Size of the generated dataset: 16.02 GB\n* Total amount of disk used: 18.68 GB\n\n\nAn example of 'train' looks as follows.#### rc.nocontext\n\n\n* Size of downloaded dataset files: 2.67 GB\n* Size of the generated dataset: 126.27 MB\n* Total amount of disk used: 2.79 GB\n\n\nAn example of 'train' looks as follows.#### unfiltered\n\n\n* Size of downloaded dataset files: 3.30 GB\n* Size of the generated dataset: 29.24 GB\n* Total amount of disk used: 32.54 GB\n\n\nAn example of 'validation' looks as follows.#### unfiltered.nocontext\n\n\n* Size of downloaded dataset files: 632.55 MB\n* Size of the generated dataset: 74.56 MB\n* Total amount of disk used: 707.11 MB\n\n\nAn example of 'train' looks as follows.", "passage: ### Data Fields\n\n\nThe data fields are the same among all splits.#### rc\n\n\n* 'question': a 'string' feature.\n* 'question\\_id': a 'string' feature.\n* 'question\\_source': a 'string' feature.\n* 'entity\\_pages': a dictionary feature containing:\n\t+ 'doc\\_source': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'wiki\\_context': a 'string' feature.\n* 'search\\_results': a dictionary feature containing:\n\t+ 'description': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'rank': a 'int32' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'url': a 'string' feature.\n\t+ 'search\\_context': a 'string' feature.\n* 'aliases': a 'list' of 'string' features.\n* 'normalized\\_aliases': a 'list' of 'string' features.\n* 'matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_value': a 'string' feature.\n* 'type': a 'string' feature.\n* 'value': a 'string' feature.", "passage: #### rc.nocontext\n\n\n* 'question': a 'string' feature.\n* 'question\\_id': a 'string' feature.\n* 'question\\_source': a 'string' feature.\n* 'entity\\_pages': a dictionary feature containing:\n\t+ 'doc\\_source': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'wiki\\_context': a 'string' feature.\n* 'search\\_results': a dictionary feature containing:\n\t+ 'description': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'rank': a 'int32' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'url': a 'string' feature.\n\t+ 'search\\_context': a 'string' feature.\n* 'aliases': a 'list' of 'string' features.\n* 'normalized\\_aliases': a 'list' of 'string' features.\n* 'matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_value': a 'string' feature.\n* 'type': a 'string' feature.\n* 'value': a 'string' feature.#### unfiltered\n\n\n* 'question': a 'string' feature.\n* 'question\\_id': a 'string' feature.\n* 'question\\_source': a 'string' feature.\n* 'entity\\_pages': a dictionary feature containing:\n\t+ 'doc\\_source': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'wiki\\_context': a 'string' feature.\n* 'search\\_results': a dictionary feature containing:\n\t+ 'description': a 'string' feature.\n\t+ 'filename': a 'string' feature.\n\t+ 'rank': a 'int32' feature.\n\t+ 'title': a 'string' feature.\n\t+ 'url': a 'string' feature.\n\t+ 'search\\_context': a 'string' feature.\n* 'aliases': a 'list' of 'string' features.\n* 'normalized\\_aliases': a 'list' of 'string' features.\n* 'matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_matched\\_wiki\\_entity\\_name': a 'string' feature.\n* 'normalized\\_value': a 'string' feature.\n* 'type': a 'string' feature.\n* 'value': a 'string' feature." ]
80e5de38cb98c00fa7d95f2057954d7045f1a484
# Dataset Card for Tunisian Sentiment Analysis Corpus ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** None - **Repository:** https://github.com/fbougares/TSAC - **Paper:** https://www.aclweb.org/anthology/W17-1307 - **Leaderboard:** [If the dataset supports an active leaderboard, add link here]() - **Point of Contact:** Salima Mdhaffar (firstname.lastname@univ-lemans.fr) ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@abhishekkrthakur](https://github.com/abhishekkrthakur) for adding this dataset.
tsac
[ "task_categories:text-classification", "task_ids:sentiment-classification", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:aeb", "license:lgpl-3.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found"], "language": ["aeb"], "license": ["lgpl-3.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["sentiment-classification"], "paperswithcode_id": "tsac", "pretty_name": "Tunisian Sentiment Analysis Corpus", "dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "sentence", "dtype": "string"}, {"name": "target", "dtype": {"class_label": {"names": {"0": "1", "1": "-1"}}}}], "splits": [{"name": "train", "num_bytes": 1020146, "num_examples": 13669}, {"name": "test", "num_bytes": 268504, "num_examples": 3400}], "download_size": 963015, "dataset_size": 1288650}}
2024-01-18T11:17:21+00:00
[]
[ "aeb" ]
TAGS #task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Tunisian Arabic #license-lgpl-3.0 #region-us
# Dataset Card for Tunisian Sentiment Analysis Corpus ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: None - Repository: URL - Paper: URL - Leaderboard: [If the dataset supports an active leaderboard, add link here]() - Point of Contact: Salima Mdhaffar (firstname.lastname@URL) ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @abhishekkrthakur for adding this dataset.
[ "# Dataset Card for Tunisian Sentiment Analysis Corpus", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: None\n- Repository: URL\n- Paper: URL\n- Leaderboard: [If the dataset supports an active leaderboard, add link here]()\n- Point of Contact: Salima Mdhaffar (firstname.lastname@URL)", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Tunisian Arabic #license-lgpl-3.0 #region-us \n", "# Dataset Card for Tunisian Sentiment Analysis Corpus", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: None\n- Repository: URL\n- Paper: URL\n- Leaderboard: [If the dataset supports an active leaderboard, add link here]()\n- Point of Contact: Salima Mdhaffar (firstname.lastname@URL)", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ 92, 12, 120, 61, 6, 10, 4, 6, 6, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 20 ]
[ "passage: TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Tunisian Arabic #license-lgpl-3.0 #region-us \n# Dataset Card for Tunisian Sentiment Analysis Corpus## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: None\n- Repository: URL\n- Paper: URL\n- Leaderboard: [If the dataset supports an active leaderboard, add link here]()\n- Point of Contact: Salima Mdhaffar (firstname.lastname@URL)### Dataset Summary### Supported Tasks and Leaderboards### Languages## Dataset Structure### Data Instances### Data Fields### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators### Licensing Information### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
65dc04cb744cb0424fe101c44861d34073fbd5ae
# Dataset Card for TTC4900: A Benchmark Data for Turkish Text Categorization ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [TTC4900 Homepage](https://www.kaggle.com/savasy/ttc4900) - **Repository:** [TTC4900 Repository](https://github.com/savasy/TurkishTextClassification) - **Paper:** [A Comparison of Different Approaches to Document Representation in Turkish Language](https://dergipark.org.tr/en/pub/sdufenbed/issue/38975/456349) - **Point of Contact:** [Savaş Yıldırım](mailto:savasy@gmail.com) ### Dataset Summary The data set is taken from [kemik group](http://www.kemik.yildiz.edu.tr/) The data are pre-processed for the text categorization, collocations are found, character set is corrected, and so forth. We named TTC4900 by mimicking the name convention of TTC 3600 dataset shared by the study ["A Knowledge-poor Approach to Turkish Text Categorization with a Comparative Analysis, Proceedings of CICLING 2014, Springer LNCS, Nepal, 2014"](https://link.springer.com/chapter/10.1007/978-3-642-54903-8_36) If you use the dataset in a paper, please refer https://www.kaggle.com/savasy/ttc4900 as footnote and cite one of the papers as follows: - A Comparison of Different Approaches to Document Representation in Turkish Language, SDU Journal of Natural and Applied Science, Vol 22, Issue 2, 2018 - A comparative analysis of text classification for Turkish language, Pamukkale University Journal of Engineering Science Volume 25 Issue 5, 2018 - A Knowledge-poor Approach to Turkish Text Categorization with a Comparative Analysis, Proceedings of CICLING 2014, Springer LNCS, Nepal, 2014. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The dataset is based on Turkish. ## Dataset Structure ### Data Instances A text classification dataset with 7 different news category. Here is an example from the dataset: ``` { "category": 0, # politics/siyaset "text": "paris teki infaz imralı ile başlayan sürece bir darbe mi elif_çakır ın sunduğu söz_bitmeden in bugünkü konuğu gazeteci melih altınok oldu programdan satıbaşları imralı ile görüşmeler hangi aşamada bundan sonra ne olacak hangi kesimler sürece engel oluyor psikolojik mayınlar neler türk solu bu dönemde evrensel sorumluluğunu yerine getirebiliyor mu elif_çakır sordu melih altınok söz_bitmeden de yanıtladı elif_çakır pkk nın silahsızlandırılmasına yönelik olarak öcalan ile görüşme sonrası 3 kadının infazı enteresan çünkü kurucu isimlerden birisi sen nasıl okudun bu infazı melih altınok herkesin ciddi anlamda şüpheleri var şu an yürüttüğümüz herşey bir delile dayanmadığı için komple teorisinden ibaret kalacak ama şöyle bir durum var imralı görüşmelerin ilk defa bir siyasi iktidar tarafından açıkça söylendiği bir dönem ardından geliyor bu sürecin gerçekleşmemesini isteyen kesimler yaptırmıştır dedi" } ``` ### Data Fields - **category** : Indicates to which category the news text belongs. (Such as "politics", "world", "economy", "culture", "health", "sports", "technology".) - **text** : Contains the text of the news. ### Data Splits It is not divided into Train set and Test set. ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization The data are pre-processed for the text categorization, collocations are found, character set is corrected, and so forth. #### Who are the source language producers? Turkish online news sites. ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators The dataset was created by [Savaş Yıldırım](https://github.com/savasy) ### Licensing Information [More Information Needed] ### Citation Information ``` @article{doi:10.5505/pajes.2018.15931, author = {Yıldırım, Savaş and Yıldız, Tuğba}, title = {A comparative analysis of text classification for Turkish language}, journal = {Pamukkale Univ Muh Bilim Derg}, volume = {24}, number = {5}, pages = {879-886}, year = {2018}, doi = {10.5505/pajes.2018.15931}, note ={doi: 10.5505/pajes.2018.15931}, URL = {https://dx.doi.org/10.5505/pajes.2018.15931}, eprint = {https://dx.doi.org/10.5505/pajes.2018.15931} } ``` ### Contributions Thanks to [@yavuzKomecoglu](https://github.com/yavuzKomecoglu) for adding this dataset.
ttc4900
[ "task_categories:text-classification", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:tr", "license:unknown", "news-category-classification", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["tr"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": [], "pretty_name": "TTC4900 - A Benchmark Data for Turkish Text Categorization", "tags": ["news-category-classification"], "dataset_info": {"features": [{"name": "category", "dtype": {"class_label": {"names": {"0": "siyaset", "1": "dunya", "2": "ekonomi", "3": "kultur", "4": "saglik", "5": "spor", "6": "teknoloji"}}}}, {"name": "text", "dtype": "string"}], "config_name": "ttc4900", "splits": [{"name": "train", "num_bytes": 10640831, "num_examples": 4900}], "download_size": 10627541, "dataset_size": 10640831}}
2024-01-18T11:17:22+00:00
[]
[ "tr" ]
TAGS #task_categories-text-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Turkish #license-unknown #news-category-classification #region-us
# Dataset Card for TTC4900: A Benchmark Data for Turkish Text Categorization ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: TTC4900 Homepage - Repository: TTC4900 Repository - Paper: A Comparison of Different Approaches to Document Representation in Turkish Language - Point of Contact: Savaş Yıldırım ### Dataset Summary The data set is taken from kemik group The data are pre-processed for the text categorization, collocations are found, character set is corrected, and so forth. We named TTC4900 by mimicking the name convention of TTC 3600 dataset shared by the study "A Knowledge-poor Approach to Turkish Text Categorization with a Comparative Analysis, Proceedings of CICLING 2014, Springer LNCS, Nepal, 2014" If you use the dataset in a paper, please refer URL as footnote and cite one of the papers as follows: - A Comparison of Different Approaches to Document Representation in Turkish Language, SDU Journal of Natural and Applied Science, Vol 22, Issue 2, 2018 - A comparative analysis of text classification for Turkish language, Pamukkale University Journal of Engineering Science Volume 25 Issue 5, 2018 - A Knowledge-poor Approach to Turkish Text Categorization with a Comparative Analysis, Proceedings of CICLING 2014, Springer LNCS, Nepal, 2014. ### Supported Tasks and Leaderboards ### Languages The dataset is based on Turkish. ## Dataset Structure ### Data Instances A text classification dataset with 7 different news category. Here is an example from the dataset: ### Data Fields - category : Indicates to which category the news text belongs. (Such as "politics", "world", "economy", "culture", "health", "sports", "technology".) - text : Contains the text of the news. ### Data Splits It is not divided into Train set and Test set. ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization The data are pre-processed for the text categorization, collocations are found, character set is corrected, and so forth. #### Who are the source language producers? Turkish online news sites. ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators The dataset was created by Savaş Yıldırım ### Licensing Information ### Contributions Thanks to @yavuzKomecoglu for adding this dataset.
[ "# Dataset Card for TTC4900: A Benchmark Data for Turkish Text Categorization", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: TTC4900 Homepage\n- Repository: TTC4900 Repository\n- Paper: A Comparison of Different Approaches to Document Representation in Turkish Language\n- Point of Contact: Savaş Yıldırım", "### Dataset Summary\n\nThe data set is taken from kemik group\nThe data are pre-processed for the text categorization, collocations are found, character set is corrected, and so forth.\nWe named TTC4900 by mimicking the name convention of TTC 3600 dataset shared by the study \"A Knowledge-poor Approach to Turkish Text Categorization with a Comparative Analysis, Proceedings of CICLING 2014, Springer LNCS, Nepal, 2014\"\n\nIf you use the dataset in a paper, please refer URL as footnote and cite one of the papers as follows:\n\n- A Comparison of Different Approaches to Document Representation in Turkish Language, SDU Journal of Natural and Applied Science, Vol 22, Issue 2, 2018\n- A comparative analysis of text classification for Turkish language, Pamukkale University Journal of Engineering Science Volume 25 Issue 5, 2018\n- A Knowledge-poor Approach to Turkish Text Categorization with a Comparative Analysis, Proceedings of CICLING 2014, Springer LNCS, Nepal, 2014.", "### Supported Tasks and Leaderboards", "### Languages\n\nThe dataset is based on Turkish.", "## Dataset Structure", "### Data Instances\n\nA text classification dataset with 7 different news category. \n\nHere is an example from the dataset:", "### Data Fields\n\n- category : Indicates to which category the news text belongs.\n(Such as \"politics\", \"world\", \"economy\", \"culture\", \"health\", \"sports\", \"technology\".)\n- text : Contains the text of the news.", "### Data Splits\n\nIt is not divided into Train set and Test set.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe data are pre-processed for the text categorization, collocations are found, character set is corrected, and so forth.", "#### Who are the source language producers?\n\nTurkish online news sites.", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThe dataset was created by Savaş Yıldırım", "### Licensing Information", "### Contributions\n\nThanks to @yavuzKomecoglu for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Turkish #license-unknown #news-category-classification #region-us \n", "# Dataset Card for TTC4900: A Benchmark Data for Turkish Text Categorization", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: TTC4900 Homepage\n- Repository: TTC4900 Repository\n- Paper: A Comparison of Different Approaches to Document Representation in Turkish Language\n- Point of Contact: Savaş Yıldırım", "### Dataset Summary\n\nThe data set is taken from kemik group\nThe data are pre-processed for the text categorization, collocations are found, character set is corrected, and so forth.\nWe named TTC4900 by mimicking the name convention of TTC 3600 dataset shared by the study \"A Knowledge-poor Approach to Turkish Text Categorization with a Comparative Analysis, Proceedings of CICLING 2014, Springer LNCS, Nepal, 2014\"\n\nIf you use the dataset in a paper, please refer URL as footnote and cite one of the papers as follows:\n\n- A Comparison of Different Approaches to Document Representation in Turkish Language, SDU Journal of Natural and Applied Science, Vol 22, Issue 2, 2018\n- A comparative analysis of text classification for Turkish language, Pamukkale University Journal of Engineering Science Volume 25 Issue 5, 2018\n- A Knowledge-poor Approach to Turkish Text Categorization with a Comparative Analysis, Proceedings of CICLING 2014, Springer LNCS, Nepal, 2014.", "### Supported Tasks and Leaderboards", "### Languages\n\nThe dataset is based on Turkish.", "## Dataset Structure", "### Data Instances\n\nA text classification dataset with 7 different news category. \n\nHere is an example from the dataset:", "### Data Fields\n\n- category : Indicates to which category the news text belongs.\n(Such as \"politics\", \"world\", \"economy\", \"culture\", \"health\", \"sports\", \"technology\".)\n- text : Contains the text of the news.", "### Data Splits\n\nIt is not divided into Train set and Test set.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe data are pre-processed for the text categorization, collocations are found, character set is corrected, and so forth.", "#### Who are the source language producers?\n\nTurkish online news sites.", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThe dataset was created by Savaş Yıldırım", "### Licensing Information", "### Contributions\n\nThanks to @yavuzKomecoglu for adding this dataset." ]
[ 85, 23, 125, 51, 242, 10, 13, 6, 27, 61, 17, 5, 7, 4, 40, 16, 5, 5, 9, 8, 8, 7, 8, 7, 5, 14, 6, 20 ]
[ "passage: TAGS\n#task_categories-text-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Turkish #license-unknown #news-category-classification #region-us \n# Dataset Card for TTC4900: A Benchmark Data for Turkish Text Categorization## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: TTC4900 Homepage\n- Repository: TTC4900 Repository\n- Paper: A Comparison of Different Approaches to Document Representation in Turkish Language\n- Point of Contact: Savaş Yıldırım" ]
f2ebdb6975da3ff17468a4414a1fcb95fc6c2998
# Dataset Card for TUNIZI ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://github.com/chaymafourati/TUNIZI-Sentiment-Analysis-Tunisian-Arabizi-Dataset - **Repository:** https://github.com/chaymafourati/TUNIZI-Sentiment-Analysis-Tunisian-Arabizi-Dataset - **Paper:** https://arxiv.org/abs/2004.14303 - **Point of Contact:** Chayma Fourati (chayma@icompass.digital) ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages This dataset uses Tunisian Arabic written with latin script (BCP-47: aeb-Latn) ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@abhishekkrthakur](https://github.com/abhishekkrthakur) for adding this dataset.
tunizi
[ "task_categories:text-classification", "task_ids:sentiment-classification", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:aeb", "license:unknown", "arxiv:2004.14303", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found"], "language": ["aeb"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["sentiment-classification"], "paperswithcode_id": "tunizi", "pretty_name": "TUNIZI", "dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "sentence", "dtype": "string"}, {"name": "target", "dtype": {"class_label": {"names": {"0": "1", "1": "-1"}}}}], "splits": [{"name": "train", "num_bytes": 211166, "num_examples": 3000}], "download_size": 162781, "dataset_size": 211166}}
2024-01-18T11:17:23+00:00
[ "2004.14303" ]
[ "aeb" ]
TAGS #task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Tunisian Arabic #license-unknown #arxiv-2004.14303 #region-us
# Dataset Card for TUNIZI ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: URL - Paper: URL - Point of Contact: Chayma Fourati (chayma@icompass.digital) ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages This dataset uses Tunisian Arabic written with latin script (BCP-47: aeb-Latn) ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @abhishekkrthakur for adding this dataset.
[ "# Dataset Card for TUNIZI", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: URL\n- Point of Contact: Chayma Fourati (chayma@icompass.digital)", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages\n\nThis dataset uses Tunisian Arabic written with latin script (BCP-47: aeb-Latn)", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Tunisian Arabic #license-unknown #arxiv-2004.14303 #region-us \n", "# Dataset Card for TUNIZI", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: URL\n- Point of Contact: Chayma Fourati (chayma@icompass.digital)", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages\n\nThis dataset uses Tunisian Arabic written with latin script (BCP-47: aeb-Latn)", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ 100, 8, 120, 39, 6, 10, 28, 6, 6, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 20 ]
[ "passage: TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Tunisian Arabic #license-unknown #arxiv-2004.14303 #region-us \n# Dataset Card for TUNIZI## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: URL\n- Point of Contact: Chayma Fourati (chayma@icompass.digital)### Dataset Summary### Supported Tasks and Leaderboards### Languages\n\nThis dataset uses Tunisian Arabic written with latin script (BCP-47: aeb-Latn)## Dataset Structure### Data Instances### Data Fields### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators### Licensing Information### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
0fded949b75bf87ac3dd7b4db22d4a56233922a2
# Dataset Card for TupleInf Open IE ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Tuple IE Homepage](https://allenai.org/data/tuple-ie) - **Repository:** - **Paper:** [Answering Complex Questions Using Open Information Extraction](https://www.semanticscholar.org/paper/Answering-Complex-Questions-Using-Open-Information-Khot-Sabharwal/0ff595f0645a3e25a2f37145768985b10ead0509) - **Leaderboard:** - **Point of Contact:** ### Dataset Summary The TupleInf Open IE dataset contains Open IE tuples extracted from 263K sentences that were used by the solver in “Answering Complex Questions Using Open Information Extraction” (referred as Tuple KB, T). These sentences were collected from a large Web corpus using training questions from 4th and 8th grade as queries. This dataset contains 156K sentences collected for 4th grade questions and 107K sentences for 8th grade questions. Each sentence is followed by the Open IE v4 tuples using their simple format. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The text in the dataset is in English, collected from a large Web corpus using training questions from 4th and 8th grade as queries. ## Dataset Structure ### Data Instances This dataset contains setences with corresponding relation tuples extracted from each sentence. Each instance should contain a sentence and followed by the [Open IE v4](https://github.com/allenai/openie-standalone) tuples using their *simple format*. An example of an instance: ```JSON { "sentence": "0.04593 kg Used a triple beam balance to mass a golf ball.", "tuples": { "score": 0.8999999761581421, "tuple_text": "(0.04593 kg; Used; a triple beam balance; to mass a golf ball)", "context": "", "arg1": "0.04593 kg", "rel": "Used", "arg2s": ["a triple beam balance", "to mass a golf ball"], } } ``` ### Data Fields - `sentence`: the input text/sentence. - `tuples`: the extracted relation tuples from the sentence. - `score`: the confident score for each tuple. - `tuple_text`: the relationship representation text of the extraction, in the *simple format* of [Open IE v4](https://github.com/allenai/openie-standalone). - `context`: an optional representation of the context for this extraction. Defaults to `""` if there's no context. - `arg1`: the first argument in the relationship. - `rel`: the relation. - `arg2s`: a sequence of the 2nd arguments in the realtionship. ### Data Splits | name | train| |-----------|-----:| | all |267719| | 4th_grade |158910| | 8th_grade |108809| ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ```bibtex @article{Khot2017AnsweringCQ, title={Answering Complex Questions Using Open Information Extraction}, author={Tushar Khot and A. Sabharwal and Peter Clark}, journal={ArXiv}, year={2017}, volume={abs/1704.05572} } ``` ### Contributions Thanks to [@mattbui](https://github.com/mattbui) for adding this dataset.
tuple_ie
[ "task_categories:other", "annotations_creators:found", "language_creators:machine-generated", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:en", "license:unknown", "open-information-extraction", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["machine-generated"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["other"], "task_ids": [], "paperswithcode_id": "tupleinf-open-ie-dataset", "pretty_name": "TupleInf Open IE", "tags": ["open-information-extraction"], "dataset_info": [{"config_name": "all", "features": [{"name": "sentence", "dtype": "string"}, {"name": "tuples", "sequence": [{"name": "score", "dtype": "float32"}, {"name": "tuple_text", "dtype": "string"}, {"name": "context", "dtype": "string"}, {"name": "arg1", "dtype": "string"}, {"name": "rel", "dtype": "string"}, {"name": "arg2s", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 115621096, "num_examples": 267719}], "download_size": 18026102, "dataset_size": 115621096}, {"config_name": "4th_grade", "features": [{"name": "sentence", "dtype": "string"}, {"name": "tuples", "sequence": [{"name": "score", "dtype": "float32"}, {"name": "tuple_text", "dtype": "string"}, {"name": "context", "dtype": "string"}, {"name": "arg1", "dtype": "string"}, {"name": "rel", "dtype": "string"}, {"name": "arg2s", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 65363445, "num_examples": 158910}], "download_size": 18026102, "dataset_size": 65363445}, {"config_name": "8th_grade", "features": [{"name": "sentence", "dtype": "string"}, {"name": "tuples", "sequence": [{"name": "score", "dtype": "float32"}, {"name": "tuple_text", "dtype": "string"}, {"name": "context", "dtype": "string"}, {"name": "arg1", "dtype": "string"}, {"name": "rel", "dtype": "string"}, {"name": "arg2s", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 50257651, "num_examples": 108809}], "download_size": 18026102, "dataset_size": 50257651}]}
2024-01-18T11:17:24+00:00
[]
[ "en" ]
TAGS #task_categories-other #annotations_creators-found #language_creators-machine-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-unknown #open-information-extraction #region-us
Dataset Card for TupleInf Open IE ================================= Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: Tuple IE Homepage * Repository: * Paper: Answering Complex Questions Using Open Information Extraction * Leaderboard: * Point of Contact: ### Dataset Summary The TupleInf Open IE dataset contains Open IE tuples extracted from 263K sentences that were used by the solver in “Answering Complex Questions Using Open Information Extraction” (referred as Tuple KB, T). These sentences were collected from a large Web corpus using training questions from 4th and 8th grade as queries. This dataset contains 156K sentences collected for 4th grade questions and 107K sentences for 8th grade questions. Each sentence is followed by the Open IE v4 tuples using their simple format. ### Supported Tasks and Leaderboards ### Languages The text in the dataset is in English, collected from a large Web corpus using training questions from 4th and 8th grade as queries. Dataset Structure ----------------- ### Data Instances This dataset contains setences with corresponding relation tuples extracted from each sentence. Each instance should contain a sentence and followed by the Open IE v4 tuples using their *simple format*. An example of an instance: ### Data Fields * 'sentence': the input text/sentence. * 'tuples': the extracted relation tuples from the sentence. + 'score': the confident score for each tuple. + 'tuple\_text': the relationship representation text of the extraction, in the *simple format* of Open IE v4. + 'context': an optional representation of the context for this extraction. Defaults to '""' if there's no context. + 'arg1': the first argument in the relationship. + 'rel': the relation. + 'arg2s': a sequence of the 2nd arguments in the realtionship. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @mattbui for adding this dataset.
[ "### Dataset Summary\n\n\nThe TupleInf Open IE dataset contains Open IE tuples extracted from 263K sentences that were used by the solver in “Answering Complex Questions Using Open Information Extraction” (referred as Tuple KB, T). These sentences were collected from a large Web corpus using training questions from 4th and 8th grade as queries. This dataset contains 156K sentences collected for 4th grade questions and 107K sentences for 8th grade questions. Each sentence is followed by the Open IE v4 tuples using their simple format.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe text in the dataset is in English, collected from a large Web corpus using training questions from 4th and 8th grade as queries.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nThis dataset contains setences with corresponding relation tuples extracted from each sentence. Each instance should contain a sentence and followed by the Open IE v4 tuples using their *simple format*.\nAn example of an instance:", "### Data Fields\n\n\n* 'sentence': the input text/sentence.\n* 'tuples': the extracted relation tuples from the sentence.\n\t+ 'score': the confident score for each tuple.\n\t+ 'tuple\\_text': the relationship representation text of the extraction, in the *simple format* of Open IE v4.\n\t+ 'context': an optional representation of the context for this extraction. Defaults to '\"\"' if there's no context.\n\t+ 'arg1': the first argument in the relationship.\n\t+ 'rel': the relation.\n\t+ 'arg2s': a sequence of the 2nd arguments in the realtionship.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @mattbui for adding this dataset." ]
[ "TAGS\n#task_categories-other #annotations_creators-found #language_creators-machine-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-unknown #open-information-extraction #region-us \n", "### Dataset Summary\n\n\nThe TupleInf Open IE dataset contains Open IE tuples extracted from 263K sentences that were used by the solver in “Answering Complex Questions Using Open Information Extraction” (referred as Tuple KB, T). These sentences were collected from a large Web corpus using training questions from 4th and 8th grade as queries. This dataset contains 156K sentences collected for 4th grade questions and 107K sentences for 8th grade questions. Each sentence is followed by the Open IE v4 tuples using their simple format.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe text in the dataset is in English, collected from a large Web corpus using training questions from 4th and 8th grade as queries.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nThis dataset contains setences with corresponding relation tuples extracted from each sentence. Each instance should contain a sentence and followed by the Open IE v4 tuples using their *simple format*.\nAn example of an instance:", "### Data Fields\n\n\n* 'sentence': the input text/sentence.\n* 'tuples': the extracted relation tuples from the sentence.\n\t+ 'score': the confident score for each tuple.\n\t+ 'tuple\\_text': the relationship representation text of the extraction, in the *simple format* of Open IE v4.\n\t+ 'context': an optional representation of the context for this extraction. Defaults to '\"\"' if there's no context.\n\t+ 'arg1': the first argument in the relationship.\n\t+ 'rel': the relation.\n\t+ 'arg2s': a sequence of the 2nd arguments in the realtionship.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @mattbui for adding this dataset." ]
[ 81, 131, 10, 42, 54, 158, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 6, 18 ]
[ "passage: TAGS\n#task_categories-other #annotations_creators-found #language_creators-machine-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-unknown #open-information-extraction #region-us \n### Dataset Summary\n\n\nThe TupleInf Open IE dataset contains Open IE tuples extracted from 263K sentences that were used by the solver in “Answering Complex Questions Using Open Information Extraction” (referred as Tuple KB, T). These sentences were collected from a large Web corpus using training questions from 4th and 8th grade as queries. This dataset contains 156K sentences collected for 4th grade questions and 107K sentences for 8th grade questions. Each sentence is followed by the Open IE v4 tuples using their simple format.### Supported Tasks and Leaderboards### Languages\n\n\nThe text in the dataset is in English, collected from a large Web corpus using training questions from 4th and 8th grade as queries.\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nThis dataset contains setences with corresponding relation tuples extracted from each sentence. Each instance should contain a sentence and followed by the Open IE v4 tuples using their *simple format*.\nAn example of an instance:### Data Fields\n\n\n* 'sentence': the input text/sentence.\n* 'tuples': the extracted relation tuples from the sentence.\n\t+ 'score': the confident score for each tuple.\n\t+ 'tuple\\_text': the relationship representation text of the extraction, in the *simple format* of Open IE v4.\n\t+ 'context': an optional representation of the context for this extraction. Defaults to '\"\"' if there's no context.\n\t+ 'arg1': the first argument in the relationship.\n\t+ 'rel': the relation.\n\t+ 'arg2s': a sequence of the 2nd arguments in the realtionship.### Data Splits\n\n\n\nDataset Creation\n----------------### Curation Rationale### Source Data" ]
ddf932fe53473749a77845772c9c2b0cce8d6389
# Dataset Card for TURK ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** None - **Repository:** [TURK](https://github.com/cocoxu/simplification) - **Paper:** [Optimizing Statistical Machine Translation for Text Simplification](https://www.aclweb.org/anthology/Q16-1029/) - **Leaderboard:** N/A - **Point of Contact:** [Wei Xu](mailto:wei.xu@cc.gatech.edu) ### Dataset Summary TURK is a multi-reference dataset for the evaluation of sentence simplification in English. The dataset consists of 2,359 sentences from the [Parallel Wikipedia Simplification (PWKP) corpus](https://www.aclweb.org/anthology/C10-1152/). Each sentence is associated with 8 crowdsourced simplifications that focus on only lexical paraphrasing (no sentence splitting or deletion). ### Supported Tasks and Leaderboards No Leaderboard for the task. ### Languages TURK contains English text only (BCP-47: `en`). ## Dataset Structure ### Data Instances An instance consists of an original sentence and 8 possible reference simplifications that focus on lexical paraphrasing. ``` {'original': 'one side of the armed conflicts is composed mainly of the sudanese military and the janjaweed , a sudanese militia group recruited mostly from the afro-arab abbala tribes of the northern rizeigat region in sudan .', 'simplifications': ['one side of the armed conflicts is made of sudanese military and the janjaweed , a sudanese militia recruited from the afro-arab abbala tribes of the northern rizeigat region in sudan .', 'one side of the armed conflicts consist of the sudanese military and the sudanese militia group janjaweed .', 'one side of the armed conflicts is mainly sudanese military and the janjaweed , which recruited from the afro-arab abbala tribes .', 'one side of the armed conflicts is composed mainly of the sudanese military and the janjaweed , a sudanese militia group recruited mostly from the afro-arab abbala tribes in sudan .', 'one side of the armed conflicts is made up mostly of the sudanese military and the janjaweed , a sudanese militia group whose recruits mostly come from the afro-arab abbala tribes from the northern rizeigat region in sudan .', 'the sudanese military and the janjaweed make up one of the armed conflicts , mostly from the afro-arab abbal tribes in sudan .', 'one side of the armed conflicts is composed mainly of the sudanese military and the janjaweed , a sudanese militia group recruited mostly from the afro-arab abbala tribes of the northern rizeigat regime in sudan .', 'one side of the armed conflicts is composed mainly of the sudanese military and the janjaweed , a sudanese militia group recruited mostly from the afro-arab abbala tribes of the northern rizeigat region in sudan .']} ``` ### Data Fields - `original`: an original sentence from the source datasets - `simplifications`: a set of reference simplifications produced by crowd workers. ### Data Splits TURK does not contain a training set; many models use [WikiLarge](https://github.com/XingxingZhang/dress) (Zhang and Lapata, 2017) or [Wiki-Auto](https://github.com/chaojiang06/wiki-auto) (Jiang et. al 2020) for training. Each input sentence has 8 associated reference simplified sentences. 2,359 input sentences are randomly split into 2,000 validation and 359 test sentences. | | Dev | Test | Total | | ----- | ------ | ---- | ----- | | Input Sentences | 2000 | 359 | 2359 | | Reference Simplifications | 16000 | 2872 | 18872 | ## Dataset Creation ### Curation Rationale The TURK dataset was constructed to evaluate the task of text simplification. It contains multiple human-written references that focus on only lexical simplification. ### Source Data #### Initial Data Collection and Normalization The input sentences in the dataset are extracted from the [Parallel Wikipedia Simplification (PWKP) corpus](https://www.aclweb.org/anthology/C10-1152/). #### Who are the source language producers? The references are crowdsourced from Amazon Mechanical Turk. The annotators were asked to provide simplifications without losing any information or splitting the input sentence. No other demographic or compensation information is provided in the paper. ### Annotations #### Annotation process The instructions given to the annotators are available in the paper. #### Who are the annotators? The annotators are Amazon Mechanical Turk workers. ### Personal and Sensitive Information Since the dataset is created from English Wikipedia (August 22, 2009 version), all the information contained in the dataset is already in the public domain. ## Considerations for Using the Data ### Social Impact of Dataset The dataset helps move forward the research towards text simplification by creating a higher quality validation and test dataset. Progress in text simplification in turn has the potential to increase the accessibility of written documents to wider audiences. ### Discussion of Biases The dataset may contain some social biases, as the input sentences are based on Wikipedia. Studies have shown that the English Wikipedia contains both gender biases [(Schmahl et al., 2020)](https://research.tudelft.nl/en/publications/is-wikipedia-succeeding-in-reducing-gender-bias-assessing-changes) and racial biases [(Adams et al., 2019)](https://journals.sagepub.com/doi/pdf/10.1177/2378023118823946). ### Other Known Limitations Since the dataset contains only 2,359 sentences that are derived from Wikipedia, it is limited to a small subset of topics present on Wikipedia. ## Additional Information ### Dataset Curators TURK was developed by researchers at the University of Pennsylvania. The work was supported by the NSF under grant IIS-1430651 and the NSF GRFP under grant 1232825. ### Licensing Information [GNU General Public License v3.0](https://github.com/cocoxu/simplification/blob/master/LICENSE) ### Citation Information ``` @article{Xu-EtAl:2016:TACL, author = {Wei Xu and Courtney Napoles and Ellie Pavlick and Quanze Chen and Chris Callison-Burch}, title = {Optimizing Statistical Machine Translation for Text Simplification}, journal = {Transactions of the Association for Computational Linguistics}, volume = {4}, year = {2016}, url = {https://cocoxu.github.io/publications/tacl2016-smt-simplification.pdf}, pages = {401--415} } ``` ### Contributions Thanks to [@mounicam](https://github.com/mounicam) for adding this dataset.
turk
[ "task_categories:text2text-generation", "task_ids:text-simplification", "annotations_creators:machine-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:gpl-3.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["machine-generated"], "language_creators": ["found"], "language": ["en"], "license": ["gpl-3.0"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text2text-generation"], "task_ids": ["text-simplification"], "pretty_name": "TURK", "dataset_info": {"features": [{"name": "original", "dtype": "string"}, {"name": "simplifications", "sequence": "string"}], "config_name": "simplification", "splits": [{"name": "validation", "num_bytes": 2120187, "num_examples": 2000}, {"name": "test", "num_bytes": 396378, "num_examples": 359}], "download_size": 2443394, "dataset_size": 2516565}}
2024-01-18T11:17:26+00:00
[]
[ "en" ]
TAGS #task_categories-text2text-generation #task_ids-text-simplification #annotations_creators-machine-generated #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-gpl-3.0 #region-us
Dataset Card for TURK ===================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: None * Repository: TURK * Paper: Optimizing Statistical Machine Translation for Text Simplification * Leaderboard: N/A * Point of Contact: Wei Xu ### Dataset Summary TURK is a multi-reference dataset for the evaluation of sentence simplification in English. The dataset consists of 2,359 sentences from the Parallel Wikipedia Simplification (PWKP) corpus. Each sentence is associated with 8 crowdsourced simplifications that focus on only lexical paraphrasing (no sentence splitting or deletion). ### Supported Tasks and Leaderboards No Leaderboard for the task. ### Languages TURK contains English text only (BCP-47: 'en'). Dataset Structure ----------------- ### Data Instances An instance consists of an original sentence and 8 possible reference simplifications that focus on lexical paraphrasing. ### Data Fields * 'original': an original sentence from the source datasets * 'simplifications': a set of reference simplifications produced by crowd workers. ### Data Splits TURK does not contain a training set; many models use WikiLarge (Zhang and Lapata, 2017) or Wiki-Auto (Jiang et. al 2020) for training. Each input sentence has 8 associated reference simplified sentences. 2,359 input sentences are randomly split into 2,000 validation and 359 test sentences. Dataset Creation ---------------- ### Curation Rationale The TURK dataset was constructed to evaluate the task of text simplification. It contains multiple human-written references that focus on only lexical simplification. ### Source Data #### Initial Data Collection and Normalization The input sentences in the dataset are extracted from the Parallel Wikipedia Simplification (PWKP) corpus. #### Who are the source language producers? The references are crowdsourced from Amazon Mechanical Turk. The annotators were asked to provide simplifications without losing any information or splitting the input sentence. No other demographic or compensation information is provided in the paper. ### Annotations #### Annotation process The instructions given to the annotators are available in the paper. #### Who are the annotators? The annotators are Amazon Mechanical Turk workers. ### Personal and Sensitive Information Since the dataset is created from English Wikipedia (August 22, 2009 version), all the information contained in the dataset is already in the public domain. Considerations for Using the Data --------------------------------- ### Social Impact of Dataset The dataset helps move forward the research towards text simplification by creating a higher quality validation and test dataset. Progress in text simplification in turn has the potential to increase the accessibility of written documents to wider audiences. ### Discussion of Biases The dataset may contain some social biases, as the input sentences are based on Wikipedia. Studies have shown that the English Wikipedia contains both gender biases (Schmahl et al., 2020) and racial biases (Adams et al., 2019). ### Other Known Limitations Since the dataset contains only 2,359 sentences that are derived from Wikipedia, it is limited to a small subset of topics present on Wikipedia. Additional Information ---------------------- ### Dataset Curators TURK was developed by researchers at the University of Pennsylvania. The work was supported by the NSF under grant IIS-1430651 and the NSF GRFP under grant 1232825. ### Licensing Information GNU General Public License v3.0 ### Contributions Thanks to @mounicam for adding this dataset.
[ "### Dataset Summary\n\n\nTURK is a multi-reference dataset for the evaluation of sentence simplification in English. The dataset consists of 2,359 sentences from the Parallel Wikipedia Simplification (PWKP) corpus. Each sentence is associated with 8 crowdsourced simplifications that focus on only lexical paraphrasing (no sentence splitting or deletion).", "### Supported Tasks and Leaderboards\n\n\nNo Leaderboard for the task.", "### Languages\n\n\nTURK contains English text only (BCP-47: 'en').\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nAn instance consists of an original sentence and 8 possible reference simplifications that focus on lexical paraphrasing.", "### Data Fields\n\n\n* 'original': an original sentence from the source datasets\n* 'simplifications': a set of reference simplifications produced by crowd workers.", "### Data Splits\n\n\nTURK does not contain a training set; many models use WikiLarge (Zhang and Lapata, 2017) or Wiki-Auto (Jiang et. al 2020) for training.\n\n\nEach input sentence has 8 associated reference simplified sentences. 2,359 input sentences are randomly split into 2,000 validation and 359 test sentences.\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nThe TURK dataset was constructed to evaluate the task of text simplification. It contains multiple human-written references that focus on only lexical simplification.", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nThe input sentences in the dataset are extracted from the Parallel Wikipedia Simplification (PWKP) corpus.", "#### Who are the source language producers?\n\n\nThe references are crowdsourced from Amazon Mechanical Turk. The annotators were asked to provide simplifications without losing any information or splitting the input sentence. No other demographic or compensation information is provided in the paper.", "### Annotations", "#### Annotation process\n\n\nThe instructions given to the annotators are available in the paper.", "#### Who are the annotators?\n\n\nThe annotators are Amazon Mechanical Turk workers.", "### Personal and Sensitive Information\n\n\nSince the dataset is created from English Wikipedia (August 22, 2009 version), all the information contained in the dataset is already in the public domain.\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset\n\n\nThe dataset helps move forward the research towards text simplification by creating a higher quality validation and test dataset. Progress in text simplification in turn has the potential to increase the accessibility of written documents to wider audiences.", "### Discussion of Biases\n\n\nThe dataset may contain some social biases, as the input sentences are based on Wikipedia. Studies have shown that the English Wikipedia contains both gender biases (Schmahl et al., 2020) and racial biases (Adams et al., 2019).", "### Other Known Limitations\n\n\nSince the dataset contains only 2,359 sentences that are derived from Wikipedia, it is limited to a small subset of topics present on Wikipedia.\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nTURK was developed by researchers at the University of Pennsylvania. The work was supported by the NSF under grant IIS-1430651 and the NSF GRFP under grant 1232825.", "### Licensing Information\n\n\nGNU General Public License v3.0", "### Contributions\n\n\nThanks to @mounicam for adding this dataset." ]
[ "TAGS\n#task_categories-text2text-generation #task_ids-text-simplification #annotations_creators-machine-generated #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-gpl-3.0 #region-us \n", "### Dataset Summary\n\n\nTURK is a multi-reference dataset for the evaluation of sentence simplification in English. The dataset consists of 2,359 sentences from the Parallel Wikipedia Simplification (PWKP) corpus. Each sentence is associated with 8 crowdsourced simplifications that focus on only lexical paraphrasing (no sentence splitting or deletion).", "### Supported Tasks and Leaderboards\n\n\nNo Leaderboard for the task.", "### Languages\n\n\nTURK contains English text only (BCP-47: 'en').\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nAn instance consists of an original sentence and 8 possible reference simplifications that focus on lexical paraphrasing.", "### Data Fields\n\n\n* 'original': an original sentence from the source datasets\n* 'simplifications': a set of reference simplifications produced by crowd workers.", "### Data Splits\n\n\nTURK does not contain a training set; many models use WikiLarge (Zhang and Lapata, 2017) or Wiki-Auto (Jiang et. al 2020) for training.\n\n\nEach input sentence has 8 associated reference simplified sentences. 2,359 input sentences are randomly split into 2,000 validation and 359 test sentences.\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nThe TURK dataset was constructed to evaluate the task of text simplification. It contains multiple human-written references that focus on only lexical simplification.", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nThe input sentences in the dataset are extracted from the Parallel Wikipedia Simplification (PWKP) corpus.", "#### Who are the source language producers?\n\n\nThe references are crowdsourced from Amazon Mechanical Turk. The annotators were asked to provide simplifications without losing any information or splitting the input sentence. No other demographic or compensation information is provided in the paper.", "### Annotations", "#### Annotation process\n\n\nThe instructions given to the annotators are available in the paper.", "#### Who are the annotators?\n\n\nThe annotators are Amazon Mechanical Turk workers.", "### Personal and Sensitive Information\n\n\nSince the dataset is created from English Wikipedia (August 22, 2009 version), all the information contained in the dataset is already in the public domain.\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset\n\n\nThe dataset helps move forward the research towards text simplification by creating a higher quality validation and test dataset. Progress in text simplification in turn has the potential to increase the accessibility of written documents to wider audiences.", "### Discussion of Biases\n\n\nThe dataset may contain some social biases, as the input sentences are based on Wikipedia. Studies have shown that the English Wikipedia contains both gender biases (Schmahl et al., 2020) and racial biases (Adams et al., 2019).", "### Other Known Limitations\n\n\nSince the dataset contains only 2,359 sentences that are derived from Wikipedia, it is limited to a small subset of topics present on Wikipedia.\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nTURK was developed by researchers at the University of Pennsylvania. The work was supported by the NSF under grant IIS-1430651 and the NSF GRFP under grant 1232825.", "### Licensing Information\n\n\nGNU General Public License v3.0", "### Contributions\n\n\nThanks to @mounicam for adding this dataset." ]
[ 91, 79, 17, 27, 29, 38, 83, 44, 4, 35, 59, 5, 19, 20, 49, 56, 68, 48, 47, 12, 17 ]
[ "passage: TAGS\n#task_categories-text2text-generation #task_ids-text-simplification #annotations_creators-machine-generated #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-gpl-3.0 #region-us \n### Dataset Summary\n\n\nTURK is a multi-reference dataset for the evaluation of sentence simplification in English. The dataset consists of 2,359 sentences from the Parallel Wikipedia Simplification (PWKP) corpus. Each sentence is associated with 8 crowdsourced simplifications that focus on only lexical paraphrasing (no sentence splitting or deletion).### Supported Tasks and Leaderboards\n\n\nNo Leaderboard for the task.### Languages\n\n\nTURK contains English text only (BCP-47: 'en').\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nAn instance consists of an original sentence and 8 possible reference simplifications that focus on lexical paraphrasing.### Data Fields\n\n\n* 'original': an original sentence from the source datasets\n* 'simplifications': a set of reference simplifications produced by crowd workers.### Data Splits\n\n\nTURK does not contain a training set; many models use WikiLarge (Zhang and Lapata, 2017) or Wiki-Auto (Jiang et. al 2020) for training.\n\n\nEach input sentence has 8 associated reference simplified sentences. 2,359 input sentences are randomly split into 2,000 validation and 359 test sentences.\n\n\n\nDataset Creation\n----------------### Curation Rationale\n\n\nThe TURK dataset was constructed to evaluate the task of text simplification. It contains multiple human-written references that focus on only lexical simplification.### Source Data#### Initial Data Collection and Normalization\n\n\nThe input sentences in the dataset are extracted from the Parallel Wikipedia Simplification (PWKP) corpus.#### Who are the source language producers?\n\n\nThe references are crowdsourced from Amazon Mechanical Turk. The annotators were asked to provide simplifications without losing any information or splitting the input sentence. No other demographic or compensation information is provided in the paper." ]
381e1ba0db868e8cc39cd207433f48fa511d005b
# Dataset Card for turkic_xwmt ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** - **Repository:**[Github](https://github.com/turkic-interlingua/til-mt/tree/master/xwmt) - **Paper:** [https://arxiv.org/abs/2109.04593](https://arxiv.org/abs/2109.04593) - **Leaderboard:** [More Information Needed] - **Point of Contact:** [turkicinterlingua@gmail.com](mailto:turkicinterlingua@gmail.com) ### Dataset Summary To establish a comprehensive and challenging evaluation benchmark for Machine Translation in Turkic languages, we translate a test set originally introduced in WMT 2020 News Translation Task for English-Russian. The original dataset is profesionally translated and consists of sentences from news articles that are both English and Russian-centric. We adopt this evaluation set (X-WMT) and begin efforts to translate it into several Turkic languages. The current version of X-WMT includes covers 8 Turkic languages and 88 language directions with a minimum of 300 sentences per language direction. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages Currently covered languages are (besides English and Russian): - Azerbaijani (az) - Bashkir (ba) - Karakalpak (kaa) - Kazakh (kk) - Kirghiz (ky) - Turkish (tr) - Sakha (sah) - Uzbek (uz) ## Dataset Structure ### Data Instances A random example from the Russian-Uzbek set: ``` {"translation": {'ru': 'Моника Мутсвангва , министр информации Зимбабве , утверждает , что полиция вмешалась в отъезд Магомбейи из соображений безопасности и вследствие состояния его здоровья .', 'uz': 'Zimbabvening Axborot vaziri , Monika Mutsvanva Magombeyining xavfsizligi va sog'ligi tufayli bo'lgan jo'nab ketishinida politsiya aralashuvini ushlab turadi .'}} ``` ### Data Fields Each example has one field "translation" that contains two subfields: one per language, e.g. for the Russian-Uzbek set: - **translation**: a dictionary with two subfields: - **ru**: the russian text - **uz**: the uzbek text ### Data Splits <details> <summary>Click here to show the number of examples per configuration:</summary> | | test | |:--------|-------:| | az-ba | 600 | | az-en | 600 | | az-kaa | 300 | | az-kk | 500 | | az-ky | 500 | | az-ru | 600 | | az-sah | 300 | | az-tr | 500 | | az-uz | 600 | | ba-az | 600 | | ba-en | 1000 | | ba-kaa | 300 | | ba-kk | 700 | | ba-ky | 500 | | ba-ru | 1000 | | ba-sah | 300 | | ba-tr | 700 | | ba-uz | 900 | | en-az | 600 | | en-ba | 1000 | | en-kaa | 300 | | en-kk | 700 | | en-ky | 500 | | en-ru | 1000 | | en-sah | 300 | | en-tr | 700 | | en-uz | 900 | | kaa-az | 300 | | kaa-ba | 300 | | kaa-en | 300 | | kaa-kk | 300 | | kaa-ky | 300 | | kaa-ru | 300 | | kaa-sah | 300 | | kaa-tr | 300 | | kaa-uz | 300 | | kk-az | 500 | | kk-ba | 700 | | kk-en | 700 | | kk-kaa | 300 | | kk-ky | 500 | | kk-ru | 700 | | kk-sah | 300 | | kk-tr | 500 | | kk-uz | 700 | | ky-az | 500 | | ky-ba | 500 | | ky-en | 500 | | ky-kaa | 300 | | ky-kk | 500 | | ky-ru | 500 | | ky-sah | 300 | | ky-tr | 400 | | ky-uz | 500 | | ru-az | 600 | | ru-ba | 1000 | | ru-en | 1000 | | ru-kaa | 300 | | ru-kk | 700 | | ru-ky | 500 | | ru-sah | 300 | | ru-tr | 700 | | ru-uz | 900 | | sah-az | 300 | | sah-ba | 300 | | sah-en | 300 | | sah-kaa | 300 | | sah-kk | 300 | | sah-ky | 300 | | sah-ru | 300 | | sah-tr | 300 | | sah-uz | 300 | | tr-az | 500 | | tr-ba | 700 | | tr-en | 700 | | tr-kaa | 300 | | tr-kk | 500 | | tr-ky | 400 | | tr-ru | 700 | | tr-sah | 300 | | tr-uz | 600 | | uz-az | 600 | | uz-ba | 900 | | uz-en | 900 | | uz-kaa | 300 | | uz-kk | 700 | | uz-ky | 500 | | uz-ru | 900 | | uz-sah | 300 | | uz-tr | 600 | </details> ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? **Translators, annotators and dataset contributors** (in alphabetical order) Abilxayr Zholdybai Aigiz Kunafin Akylbek Khamitov Alperen Cantez Aydos Muxammadiyarov Doniyorbek Rafikjonov Erkinbek Vokhabov Ipek Baris Iskander Shakirov Madina Zokirjonova Mohiyaxon Uzoqova Mukhammadbektosh Khaydarov Nurlan Maharramli Petr Popov Rasul Karimov Sariya Kagarmanova Ziyodabonu Qobiljon qizi ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [MIT License](https://github.com/turkic-interlingua/til-mt/blob/master/xwmt/LICENSE) ### Citation Information ``` @inproceedings{mirzakhalov2021large, title={A Large-Scale Study of Machine Translation in Turkic Languages}, author={Mirzakhalov, Jamshidbek and Babu, Anoop and Ataman, Duygu and Kariev, Sherzod and Tyers, Francis and Abduraufov, Otabek and Hajili, Mammad and Ivanova, Sardana and Khaytbaev, Abror and Laverghetta Jr, Antonio and others}, booktitle={Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing}, pages={5876--5890}, year={2021} } ``` ### Contributions This project was carried out with the help and contributions from dozens of individuals and organizations. We acknowledge and greatly appreciate each and every one of them: **Authors on the publications** (in alphabetical order) Abror Khaytbaev Ahsan Wahab Aigiz Kunafin Anoop Babu Antonio Laverghetta Jr. Behzodbek Moydinboyev Dr. Duygu Ataman Esra Onal Dr. Francis Tyers Jamshidbek Mirzakhalov Dr. John Licato Dr. Julia Kreutzer Mammad Hajili Mokhiyakhon Uzokova Dr. Orhan Firat Otabek Abduraufov Sardana Ivanova Shaxnoza Pulatova Sherzod Kariev Dr. Sriram Chellappan **Translators, annotators and dataset contributors** (in alphabetical order) Abilxayr Zholdybai Aigiz Kunafin Akylbek Khamitov Alperen Cantez Aydos Muxammadiyarov Doniyorbek Rafikjonov Erkinbek Vokhabov Ipek Baris Iskander Shakirov Madina Zokirjonova Mohiyaxon Uzoqova Mukhammadbektosh Khaydarov Nurlan Maharramli Petr Popov Rasul Karimov Sariya Kagarmanova Ziyodabonu Qobiljon qizi **Industry supporters** [Google Cloud](https://cloud.google.com/solutions/education) [Khan Academy Oʻzbek](https://uz.khanacademy.org/) [The Foundation for the Preservation and Development of the Bashkir Language](https://bsfond.ru/) Thanks to [@mirzakhalov](https://github.com/mirzakhalov) for adding this dataset.
turkic_xwmt
[ "task_categories:translation", "annotations_creators:crowdsourced", "language_creators:found", "multilinguality:translation", "size_categories:n<1K", "source_datasets:extended|WMT 2020 News Translation Task", "language:az", "language:ba", "language:en", "language:kaa", "language:kk", "language:ky", "language:ru", "language:sah", "language:tr", "language:uz", "license:mit", "arxiv:2109.04593", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["found"], "language": ["az", "ba", "en", "kaa", "kk", "ky", "ru", "sah", "tr", "uz"], "license": ["mit"], "multilinguality": ["translation"], "size_categories": ["n<1K"], "source_datasets": ["extended|WMT 2020 News Translation Task"], "task_categories": ["translation"], "task_ids": [], "pretty_name": "turkic_xwmt", "config_names": ["az-ba", "az-en", "az-kaa", "az-kk", "az-ky", "az-ru", "az-sah", "az-tr", "az-uz", "ba-az", "ba-en", "ba-kaa", "ba-kk", "ba-ky", "ba-ru", "ba-sah", "ba-tr", "ba-uz", "en-az", "en-ba", "en-kaa", "en-kk", "en-ky", "en-ru", "en-sah", "en-tr", "en-uz", "kaa-az", "kaa-ba", "kaa-en", "kaa-kk", "kaa-ky", "kaa-ru", "kaa-sah", "kaa-tr", "kaa-uz", "kk-az", "kk-ba", "kk-en", "kk-kaa", "kk-ky", "kk-ru", "kk-sah", "kk-tr", "kk-uz", "ky-az", "ky-ba", "ky-en", "ky-kaa", "ky-kk", "ky-ru", "ky-sah", "ky-tr", "ky-uz", "ru-az", "ru-ba", "ru-en", "ru-kaa", "ru-kk", "ru-ky", "ru-sah", "ru-tr", "ru-uz", "sah-az", "sah-ba", "sah-en", "sah-kaa", "sah-kk", "sah-ky", "sah-ru", "sah-tr", "sah-uz", "tr-az", "tr-ba", "tr-en", "tr-kaa", "tr-kk", "tr-ky", "tr-ru", "tr-sah", "tr-uz", "uz-az", "uz-ba", "uz-en", "uz-kaa", "uz-kk", "uz-ky", "uz-ru", "uz-sah", "uz-tr"], "dataset_info": [{"config_name": "az-ba", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["az", "ba"]}}}], "splits": [{"name": "test", "num_bytes": 266801, "num_examples": 600}], "download_size": 12862396, "dataset_size": 266801}, {"config_name": "az-en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["az", "en"]}}}], "splits": [{"name": "test", "num_bytes": 181156, "num_examples": 600}], "download_size": 12862396, "dataset_size": 181156}, {"config_name": "az-kaa", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["az", "kaa"]}}}], "splits": [{"name": "test", "num_bytes": 134071, "num_examples": 300}], "download_size": 12862396, "dataset_size": 134071}, {"config_name": "az-kk", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["az", "kk"]}}}], "splits": [{"name": "test", "num_bytes": 203798, "num_examples": 500}], "download_size": 12862396, "dataset_size": 203798}, {"config_name": "az-ky", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["az", "ky"]}}}], "splits": [{"name": "test", "num_bytes": 210549, "num_examples": 500}], "download_size": 12862396, "dataset_size": 210549}, {"config_name": "az-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["az", "ru"]}}}], "splits": [{"name": "test", "num_bytes": 262739, "num_examples": 600}], "download_size": 12862396, "dataset_size": 262739}, {"config_name": "az-sah", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["az", "sah"]}}}], "splits": [{"name": "test", "num_bytes": 144198, "num_examples": 300}], "download_size": 12862396, "dataset_size": 144198}, {"config_name": "az-tr", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["az", "tr"]}}}], "splits": [{"name": "test", "num_bytes": 162447, "num_examples": 500}], "download_size": 12862396, "dataset_size": 162447}, {"config_name": "az-uz", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["az", "uz"]}}}], "splits": [{"name": "test", "num_bytes": 194231, "num_examples": 600}], "download_size": 12862396, "dataset_size": 194231}, {"config_name": "ba-az", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ba", "az"]}}}], "splits": [{"name": "test", "num_bytes": 266801, "num_examples": 600}], "download_size": 12862396, "dataset_size": 266801}, {"config_name": "ba-en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ba", "en"]}}}], "splits": [{"name": "test", "num_bytes": 431223, "num_examples": 1000}], "download_size": 12862396, "dataset_size": 431223}, {"config_name": "ba-kaa", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ba", "kaa"]}}}], "splits": [{"name": "test", "num_bytes": 168895, "num_examples": 300}], "download_size": 12862396, "dataset_size": 168895}, {"config_name": "ba-kk", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ba", "kk"]}}}], "splits": [{"name": "test", "num_bytes": 374756, "num_examples": 700}], "download_size": 12862396, "dataset_size": 374756}, {"config_name": "ba-ky", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ba", "ky"]}}}], "splits": [{"name": "test", "num_bytes": 268986, "num_examples": 500}], "download_size": 12862396, "dataset_size": 268986}, {"config_name": "ba-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ba", "ru"]}}}], "splits": [{"name": "test", "num_bytes": 568101, "num_examples": 1000}], "download_size": 12862396, "dataset_size": 568101}, {"config_name": "ba-sah", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ba", "sah"]}}}], "splits": [{"name": "test", "num_bytes": 179022, "num_examples": 300}], "download_size": 12862396, "dataset_size": 179022}, {"config_name": "ba-tr", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ba", "tr"]}}}], "splits": [{"name": "test", "num_bytes": 309455, "num_examples": 700}], "download_size": 12862396, "dataset_size": 309455}, {"config_name": "ba-uz", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ba", "uz"]}}}], "splits": [{"name": "test", "num_bytes": 410874, "num_examples": 900}], "download_size": 12862396, "dataset_size": 410874}, {"config_name": "en-az", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "az"]}}}], "splits": [{"name": "test", "num_bytes": 181156, "num_examples": 600}], "download_size": 12862396, "dataset_size": 181156}, {"config_name": "en-ba", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "ba"]}}}], "splits": [{"name": "test", "num_bytes": 431223, "num_examples": 1000}], "download_size": 12862396, "dataset_size": 431223}, {"config_name": "en-kaa", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "kaa"]}}}], "splits": [{"name": "test", "num_bytes": 126304, "num_examples": 300}], "download_size": 12862396, "dataset_size": 126304}, {"config_name": "en-kk", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "kk"]}}}], "splits": [{"name": "test", "num_bytes": 274728, "num_examples": 700}], "download_size": 12862396, "dataset_size": 274728}, {"config_name": "en-ky", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "ky"]}}}], "splits": [{"name": "test", "num_bytes": 198854, "num_examples": 500}], "download_size": 12862396, "dataset_size": 198854}, {"config_name": "en-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "ru"]}}}], "splits": [{"name": "test", "num_bytes": 422718, "num_examples": 1000}], "download_size": 12862396, "dataset_size": 422718}, {"config_name": "en-sah", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "sah"]}}}], "splits": [{"name": "test", "num_bytes": 136431, "num_examples": 300}], "download_size": 12862396, "dataset_size": 136431}, {"config_name": "en-tr", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "tr"]}}}], "splits": [{"name": "test", "num_bytes": 210144, "num_examples": 700}], "download_size": 12862396, "dataset_size": 210144}, {"config_name": "en-uz", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "uz"]}}}], "splits": [{"name": "test", "num_bytes": 278971, "num_examples": 900}], "download_size": 12862396, "dataset_size": 278971}, {"config_name": "kaa-az", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kaa", "az"]}}}], "splits": [{"name": "test", "num_bytes": 134071, "num_examples": 300}], "download_size": 12862396, "dataset_size": 134071}, {"config_name": "kaa-ba", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kaa", "ba"]}}}], "splits": [{"name": "test", "num_bytes": 168895, "num_examples": 300}], "download_size": 12862396, "dataset_size": 168895}, {"config_name": "kaa-en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kaa", "en"]}}}], "splits": [{"name": "test", "num_bytes": 126304, "num_examples": 300}], "download_size": 12862396, "dataset_size": 126304}, {"config_name": "kaa-kk", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kaa", "kk"]}}}], "splits": [{"name": "test", "num_bytes": 160022, "num_examples": 300}], "download_size": 12862396, "dataset_size": 160022}, {"config_name": "kaa-ky", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kaa", "ky"]}}}], "splits": [{"name": "test", "num_bytes": 163763, "num_examples": 300}], "download_size": 12862396, "dataset_size": 163763}, {"config_name": "kaa-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kaa", "ru"]}}}], "splits": [{"name": "test", "num_bytes": 168349, "num_examples": 300}], "download_size": 12862396, "dataset_size": 168349}, {"config_name": "kaa-sah", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kaa", "sah"]}}}], "splits": [{"name": "test", "num_bytes": 177151, "num_examples": 300}], "download_size": 12862396, "dataset_size": 177151}, {"config_name": "kaa-tr", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kaa", "tr"]}}}], "splits": [{"name": "test", "num_bytes": 132055, "num_examples": 300}], "download_size": 12862396, "dataset_size": 132055}, {"config_name": "kaa-uz", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kaa", "uz"]}}}], "splits": [{"name": "test", "num_bytes": 132789, "num_examples": 300}], "download_size": 12862396, "dataset_size": 132789}, {"config_name": "kk-az", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kk", "az"]}}}], "splits": [{"name": "test", "num_bytes": 203798, "num_examples": 500}], "download_size": 12862396, "dataset_size": 203798}, {"config_name": "kk-ba", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kk", "ba"]}}}], "splits": [{"name": "test", "num_bytes": 374756, "num_examples": 700}], "download_size": 12862396, "dataset_size": 374756}, {"config_name": "kk-en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kk", "en"]}}}], "splits": [{"name": "test", "num_bytes": 274728, "num_examples": 700}], "download_size": 12862396, "dataset_size": 274728}, {"config_name": "kk-kaa", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kk", "kaa"]}}}], "splits": [{"name": "test", "num_bytes": 160022, "num_examples": 300}], "download_size": 12862396, "dataset_size": 160022}, {"config_name": "kk-ky", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kk", "ky"]}}}], "splits": [{"name": "test", "num_bytes": 253421, "num_examples": 500}], "download_size": 12862396, "dataset_size": 253421}, {"config_name": "kk-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kk", "ru"]}}}], "splits": [{"name": "test", "num_bytes": 369633, "num_examples": 700}], "download_size": 12862396, "dataset_size": 369633}, {"config_name": "kk-sah", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kk", "sah"]}}}], "splits": [{"name": "test", "num_bytes": 170149, "num_examples": 300}], "download_size": 12862396, "dataset_size": 170149}, {"config_name": "kk-tr", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kk", "tr"]}}}], "splits": [{"name": "test", "num_bytes": 204442, "num_examples": 500}], "download_size": 12862396, "dataset_size": 204442}, {"config_name": "kk-uz", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["kk", "uz"]}}}], "splits": [{"name": "test", "num_bytes": 290325, "num_examples": 700}], "download_size": 12862396, "dataset_size": 290325}, {"config_name": "ky-az", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ky", "az"]}}}], "splits": [{"name": "test", "num_bytes": 210549, "num_examples": 500}], "download_size": 12862396, "dataset_size": 210549}, {"config_name": "ky-ba", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ky", "ba"]}}}], "splits": [{"name": "test", "num_bytes": 268986, "num_examples": 500}], "download_size": 12862396, "dataset_size": 268986}, {"config_name": "ky-en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ky", "en"]}}}], "splits": [{"name": "test", "num_bytes": 198854, "num_examples": 500}], "download_size": 12862396, "dataset_size": 198854}, {"config_name": "ky-kaa", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ky", "kaa"]}}}], "splits": [{"name": "test", "num_bytes": 163763, "num_examples": 300}], "download_size": 12862396, "dataset_size": 163763}, {"config_name": "ky-kk", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ky", "kk"]}}}], "splits": [{"name": "test", "num_bytes": 253421, "num_examples": 500}], "download_size": 12862396, "dataset_size": 253421}, {"config_name": "ky-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ky", "ru"]}}}], "splits": [{"name": "test", "num_bytes": 265803, "num_examples": 500}], "download_size": 12862396, "dataset_size": 265803}, {"config_name": "ky-sah", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ky", "sah"]}}}], "splits": [{"name": "test", "num_bytes": 173890, "num_examples": 300}], "download_size": 12862396, "dataset_size": 173890}, {"config_name": "ky-tr", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ky", "tr"]}}}], "splits": [{"name": "test", "num_bytes": 168026, "num_examples": 400}], "download_size": 12862396, "dataset_size": 168026}, {"config_name": "ky-uz", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ky", "uz"]}}}], "splits": [{"name": "test", "num_bytes": 209619, "num_examples": 500}], "download_size": 12862396, "dataset_size": 209619}, {"config_name": "ru-az", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ru", "az"]}}}], "splits": [{"name": "test", "num_bytes": 262739, "num_examples": 600}], "download_size": 12862396, "dataset_size": 262739}, {"config_name": "ru-ba", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ru", "ba"]}}}], "splits": [{"name": "test", "num_bytes": 568101, "num_examples": 1000}], "download_size": 12862396, "dataset_size": 568101}, {"config_name": "ru-en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ru", "en"]}}}], "splits": [{"name": "test", "num_bytes": 422718, "num_examples": 1000}], "download_size": 12862396, "dataset_size": 422718}, {"config_name": "ru-kaa", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ru", "kaa"]}}}], "splits": [{"name": "test", "num_bytes": 168349, "num_examples": 300}], "download_size": 12862396, "dataset_size": 168349}, {"config_name": "ru-kk", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ru", "kk"]}}}], "splits": [{"name": "test", "num_bytes": 369633, "num_examples": 700}], "download_size": 12862396, "dataset_size": 369633}, {"config_name": "ru-ky", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ru", "ky"]}}}], "splits": [{"name": "test", "num_bytes": 265803, "num_examples": 500}], "download_size": 12862396, "dataset_size": 265803}, {"config_name": "ru-sah", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ru", "sah"]}}}], "splits": [{"name": "test", "num_bytes": 178476, "num_examples": 300}], "download_size": 12862396, "dataset_size": 178476}, {"config_name": "ru-tr", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ru", "tr"]}}}], "splits": [{"name": "test", "num_bytes": 304586, "num_examples": 700}], "download_size": 12862396, "dataset_size": 304586}, {"config_name": "ru-uz", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ru", "uz"]}}}], "splits": [{"name": "test", "num_bytes": 403551, "num_examples": 900}], "download_size": 12862396, "dataset_size": 403551}, {"config_name": "sah-az", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["sah", "az"]}}}], "splits": [{"name": "test", "num_bytes": 144198, "num_examples": 300}], "download_size": 12862396, "dataset_size": 144198}, {"config_name": "sah-ba", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["sah", "ba"]}}}], "splits": [{"name": "test", "num_bytes": 179022, "num_examples": 300}], "download_size": 12862396, "dataset_size": 179022}, {"config_name": "sah-en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["sah", "en"]}}}], "splits": [{"name": "test", "num_bytes": 136431, "num_examples": 300}], "download_size": 12862396, "dataset_size": 136431}, {"config_name": "sah-kaa", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["sah", "kaa"]}}}], "splits": [{"name": "test", "num_bytes": 177151, "num_examples": 300}], "download_size": 12862396, "dataset_size": 177151}, {"config_name": "sah-kk", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["sah", "kk"]}}}], "splits": [{"name": "test", "num_bytes": 170149, "num_examples": 300}], "download_size": 12862396, "dataset_size": 170149}, {"config_name": "sah-ky", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["sah", "ky"]}}}], "splits": [{"name": "test", "num_bytes": 173890, "num_examples": 300}], "download_size": 12862396, "dataset_size": 173890}, {"config_name": "sah-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["sah", "ru"]}}}], "splits": [{"name": "test", "num_bytes": 178476, "num_examples": 300}], "download_size": 12862396, "dataset_size": 178476}, {"config_name": "sah-tr", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["sah", "tr"]}}}], "splits": [{"name": "test", "num_bytes": 142182, "num_examples": 300}], "download_size": 12862396, "dataset_size": 142182}, {"config_name": "sah-uz", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["sah", "uz"]}}}], "splits": [{"name": "test", "num_bytes": 142916, "num_examples": 300}], "download_size": 12862396, "dataset_size": 142916}, {"config_name": "tr-az", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["tr", "az"]}}}], "splits": [{"name": "test", "num_bytes": 162447, "num_examples": 500}], "download_size": 12862396, "dataset_size": 162447}, {"config_name": "tr-ba", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["tr", "ba"]}}}], "splits": [{"name": "test", "num_bytes": 309455, "num_examples": 700}], "download_size": 12862396, "dataset_size": 309455}, {"config_name": "tr-en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["tr", "en"]}}}], "splits": [{"name": "test", "num_bytes": 210144, "num_examples": 700}], "download_size": 12862396, "dataset_size": 210144}, {"config_name": "tr-kaa", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["tr", "kaa"]}}}], "splits": [{"name": "test", "num_bytes": 132055, "num_examples": 300}], "download_size": 12862396, "dataset_size": 132055}, {"config_name": "tr-kk", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["tr", "kk"]}}}], "splits": [{"name": "test", "num_bytes": 204442, "num_examples": 500}], "download_size": 12862396, "dataset_size": 204442}, {"config_name": "tr-ky", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["tr", "ky"]}}}], "splits": [{"name": "test", "num_bytes": 168026, "num_examples": 400}], "download_size": 12862396, "dataset_size": 168026}, {"config_name": "tr-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["tr", "ru"]}}}], "splits": [{"name": "test", "num_bytes": 304586, "num_examples": 700}], "download_size": 12862396, "dataset_size": 304586}, {"config_name": "tr-sah", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["tr", "sah"]}}}], "splits": [{"name": "test", "num_bytes": 142182, "num_examples": 300}], "download_size": 12862396, "dataset_size": 142182}, {"config_name": "tr-uz", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["tr", "uz"]}}}], "splits": [{"name": "test", "num_bytes": 194761, "num_examples": 600}], "download_size": 12862396, "dataset_size": 194761}, {"config_name": "uz-az", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["uz", "az"]}}}], "splits": [{"name": "test", "num_bytes": 194231, "num_examples": 600}], "download_size": 12862396, "dataset_size": 194231}, {"config_name": "uz-ba", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["uz", "ba"]}}}], "splits": [{"name": "test", "num_bytes": 410874, "num_examples": 900}], "download_size": 12862396, "dataset_size": 410874}, {"config_name": "uz-en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["uz", "en"]}}}], "splits": [{"name": "test", "num_bytes": 278971, "num_examples": 900}], "download_size": 12862396, "dataset_size": 278971}, {"config_name": "uz-kaa", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["uz", "kaa"]}}}], "splits": [{"name": "test", "num_bytes": 132789, "num_examples": 300}], "download_size": 12862396, "dataset_size": 132789}, {"config_name": "uz-kk", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["uz", "kk"]}}}], "splits": [{"name": "test", "num_bytes": 290325, "num_examples": 700}], "download_size": 12862396, "dataset_size": 290325}, {"config_name": "uz-ky", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["uz", "ky"]}}}], "splits": [{"name": "test", "num_bytes": 209619, "num_examples": 500}], "download_size": 12862396, "dataset_size": 209619}, {"config_name": "uz-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["uz", "ru"]}}}], "splits": [{"name": "test", "num_bytes": 403551, "num_examples": 900}], "download_size": 12862396, "dataset_size": 403551}, {"config_name": "uz-sah", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["uz", "sah"]}}}], "splits": [{"name": "test", "num_bytes": 142916, "num_examples": 300}], "download_size": 12862396, "dataset_size": 142916}, {"config_name": "uz-tr", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["uz", "tr"]}}}], "splits": [{"name": "test", "num_bytes": 194761, "num_examples": 600}], "download_size": 12862396, "dataset_size": 194761}]}
2024-01-18T11:17:27+00:00
[ "2109.04593" ]
[ "az", "ba", "en", "kaa", "kk", "ky", "ru", "sah", "tr", "uz" ]
TAGS #task_categories-translation #annotations_creators-crowdsourced #language_creators-found #multilinguality-translation #size_categories-n<1K #source_datasets-extended|WMT 2020 News Translation Task #language-Azerbaijani #language-Bashkir #language-English #language-Kara-Kalpak #language-Kazakh #language-Kirghiz #language-Russian #language-Yakut #language-Turkish #language-Uzbek #license-mit #arxiv-2109.04593 #region-us
# Dataset Card for turkic_xwmt ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: - Repository:Github - Paper: URL - Leaderboard: - Point of Contact: turkicinterlingua@URL ### Dataset Summary To establish a comprehensive and challenging evaluation benchmark for Machine Translation in Turkic languages, we translate a test set originally introduced in WMT 2020 News Translation Task for English-Russian. The original dataset is profesionally translated and consists of sentences from news articles that are both English and Russian-centric. We adopt this evaluation set (X-WMT) and begin efforts to translate it into several Turkic languages. The current version of X-WMT includes covers 8 Turkic languages and 88 language directions with a minimum of 300 sentences per language direction. ### Supported Tasks and Leaderboards ### Languages Currently covered languages are (besides English and Russian): - Azerbaijani (az) - Bashkir (ba) - Karakalpak (kaa) - Kazakh (kk) - Kirghiz (ky) - Turkish (tr) - Sakha (sah) - Uzbek (uz) ## Dataset Structure ### Data Instances A random example from the Russian-Uzbek set: ### Data Fields Each example has one field "translation" that contains two subfields: one per language, e.g. for the Russian-Uzbek set: - translation: a dictionary with two subfields: - ru: the russian text - uz: the uzbek text ### Data Splits <details> <summary>Click here to show the number of examples per configuration:</summary> | | test | |:--------|-------:| | az-ba | 600 | | az-en | 600 | | az-kaa | 300 | | az-kk | 500 | | az-ky | 500 | | az-ru | 600 | | az-sah | 300 | | az-tr | 500 | | az-uz | 600 | | ba-az | 600 | | ba-en | 1000 | | ba-kaa | 300 | | ba-kk | 700 | | ba-ky | 500 | | ba-ru | 1000 | | ba-sah | 300 | | ba-tr | 700 | | ba-uz | 900 | | en-az | 600 | | en-ba | 1000 | | en-kaa | 300 | | en-kk | 700 | | en-ky | 500 | | en-ru | 1000 | | en-sah | 300 | | en-tr | 700 | | en-uz | 900 | | kaa-az | 300 | | kaa-ba | 300 | | kaa-en | 300 | | kaa-kk | 300 | | kaa-ky | 300 | | kaa-ru | 300 | | kaa-sah | 300 | | kaa-tr | 300 | | kaa-uz | 300 | | kk-az | 500 | | kk-ba | 700 | | kk-en | 700 | | kk-kaa | 300 | | kk-ky | 500 | | kk-ru | 700 | | kk-sah | 300 | | kk-tr | 500 | | kk-uz | 700 | | ky-az | 500 | | ky-ba | 500 | | ky-en | 500 | | ky-kaa | 300 | | ky-kk | 500 | | ky-ru | 500 | | ky-sah | 300 | | ky-tr | 400 | | ky-uz | 500 | | ru-az | 600 | | ru-ba | 1000 | | ru-en | 1000 | | ru-kaa | 300 | | ru-kk | 700 | | ru-ky | 500 | | ru-sah | 300 | | ru-tr | 700 | | ru-uz | 900 | | sah-az | 300 | | sah-ba | 300 | | sah-en | 300 | | sah-kaa | 300 | | sah-kk | 300 | | sah-ky | 300 | | sah-ru | 300 | | sah-tr | 300 | | sah-uz | 300 | | tr-az | 500 | | tr-ba | 700 | | tr-en | 700 | | tr-kaa | 300 | | tr-kk | 500 | | tr-ky | 400 | | tr-ru | 700 | | tr-sah | 300 | | tr-uz | 600 | | uz-az | 600 | | uz-ba | 900 | | uz-en | 900 | | uz-kaa | 300 | | uz-kk | 700 | | uz-ky | 500 | | uz-ru | 900 | | uz-sah | 300 | | uz-tr | 600 | </details> ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? Translators, annotators and dataset contributors (in alphabetical order) Abilxayr Zholdybai Aigiz Kunafin Akylbek Khamitov Alperen Cantez Aydos Muxammadiyarov Doniyorbek Rafikjonov Erkinbek Vokhabov Ipek Baris Iskander Shakirov Madina Zokirjonova Mohiyaxon Uzoqova Mukhammadbektosh Khaydarov Nurlan Maharramli Petr Popov Rasul Karimov Sariya Kagarmanova Ziyodabonu Qobiljon qizi ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information MIT License ### Contributions This project was carried out with the help and contributions from dozens of individuals and organizations. We acknowledge and greatly appreciate each and every one of them: Authors on the publications (in alphabetical order) Abror Khaytbaev Ahsan Wahab Aigiz Kunafin Anoop Babu Antonio Laverghetta Jr. Behzodbek Moydinboyev Dr. Duygu Ataman Esra Onal Dr. Francis Tyers Jamshidbek Mirzakhalov Dr. John Licato Dr. Julia Kreutzer Mammad Hajili Mokhiyakhon Uzokova Dr. Orhan Firat Otabek Abduraufov Sardana Ivanova Shaxnoza Pulatova Sherzod Kariev Dr. Sriram Chellappan Translators, annotators and dataset contributors (in alphabetical order) Abilxayr Zholdybai Aigiz Kunafin Akylbek Khamitov Alperen Cantez Aydos Muxammadiyarov Doniyorbek Rafikjonov Erkinbek Vokhabov Ipek Baris Iskander Shakirov Madina Zokirjonova Mohiyaxon Uzoqova Mukhammadbektosh Khaydarov Nurlan Maharramli Petr Popov Rasul Karimov Sariya Kagarmanova Ziyodabonu Qobiljon qizi Industry supporters Google Cloud Khan Academy Oʻzbek The Foundation for the Preservation and Development of the Bashkir Language Thanks to @mirzakhalov for adding this dataset.
[ "# Dataset Card for turkic_xwmt", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: \n- Repository:Github\n- Paper: URL\n- Leaderboard: \n- Point of Contact: turkicinterlingua@URL", "### Dataset Summary\n\nTo establish a comprehensive and challenging evaluation benchmark for Machine Translation in Turkic languages, we translate a test set originally introduced in WMT 2020 News Translation Task for English-Russian. The original dataset is profesionally translated and consists of sentences from news articles that are both English and Russian-centric. We adopt this evaluation set (X-WMT) and begin efforts to translate it into several Turkic languages. The current version of X-WMT includes covers 8 Turkic languages and 88 language directions with a minimum of 300 sentences per language direction.", "### Supported Tasks and Leaderboards", "### Languages\n\nCurrently covered languages are (besides English and Russian):\n- Azerbaijani (az)\n- Bashkir (ba)\n- Karakalpak (kaa)\n- Kazakh (kk)\n- Kirghiz (ky)\n- Turkish (tr)\n- Sakha (sah)\n- Uzbek (uz)", "## Dataset Structure", "### Data Instances\n\nA random example from the Russian-Uzbek set:", "### Data Fields\n\nEach example has one field \"translation\" that contains two subfields: one per language, e.g. for the Russian-Uzbek set:\n- translation: a dictionary with two subfields:\n - ru: the russian text\n - uz: the uzbek text", "### Data Splits\n\n<details>\n <summary>Click here to show the number of examples per configuration:</summary>\n| | test |\n|:--------|-------:|\n| az-ba | 600 |\n| az-en | 600 |\n| az-kaa | 300 |\n| az-kk | 500 |\n| az-ky | 500 |\n| az-ru | 600 |\n| az-sah | 300 |\n| az-tr | 500 |\n| az-uz | 600 |\n| ba-az | 600 |\n| ba-en | 1000 |\n| ba-kaa | 300 |\n| ba-kk | 700 |\n| ba-ky | 500 |\n| ba-ru | 1000 |\n| ba-sah | 300 |\n| ba-tr | 700 |\n| ba-uz | 900 |\n| en-az | 600 |\n| en-ba | 1000 |\n| en-kaa | 300 |\n| en-kk | 700 |\n| en-ky | 500 |\n| en-ru | 1000 |\n| en-sah | 300 |\n| en-tr | 700 |\n| en-uz | 900 |\n| kaa-az | 300 |\n| kaa-ba | 300 |\n| kaa-en | 300 |\n| kaa-kk | 300 |\n| kaa-ky | 300 |\n| kaa-ru | 300 |\n| kaa-sah | 300 |\n| kaa-tr | 300 |\n| kaa-uz | 300 |\n| kk-az | 500 |\n| kk-ba | 700 |\n| kk-en | 700 |\n| kk-kaa | 300 |\n| kk-ky | 500 |\n| kk-ru | 700 |\n| kk-sah | 300 |\n| kk-tr | 500 |\n| kk-uz | 700 |\n| ky-az | 500 |\n| ky-ba | 500 |\n| ky-en | 500 |\n| ky-kaa | 300 |\n| ky-kk | 500 |\n| ky-ru | 500 |\n| ky-sah | 300 |\n| ky-tr | 400 |\n| ky-uz | 500 |\n| ru-az | 600 |\n| ru-ba | 1000 |\n| ru-en | 1000 |\n| ru-kaa | 300 |\n| ru-kk | 700 |\n| ru-ky | 500 |\n| ru-sah | 300 |\n| ru-tr | 700 |\n| ru-uz | 900 |\n| sah-az | 300 |\n| sah-ba | 300 |\n| sah-en | 300 |\n| sah-kaa | 300 |\n| sah-kk | 300 |\n| sah-ky | 300 |\n| sah-ru | 300 |\n| sah-tr | 300 |\n| sah-uz | 300 |\n| tr-az | 500 |\n| tr-ba | 700 |\n| tr-en | 700 |\n| tr-kaa | 300 |\n| tr-kk | 500 |\n| tr-ky | 400 |\n| tr-ru | 700 |\n| tr-sah | 300 |\n| tr-uz | 600 |\n| uz-az | 600 |\n| uz-ba | 900 |\n| uz-en | 900 |\n| uz-kaa | 300 |\n| uz-kk | 700 |\n| uz-ky | 500 |\n| uz-ru | 900 |\n| uz-sah | 300 |\n| uz-tr | 600 |\n</details>", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?\n\nTranslators, annotators and dataset contributors (in alphabetical order)\n\nAbilxayr Zholdybai \nAigiz Kunafin \nAkylbek Khamitov \nAlperen Cantez \nAydos Muxammadiyarov \nDoniyorbek Rafikjonov \nErkinbek Vokhabov \nIpek Baris \nIskander Shakirov \nMadina Zokirjonova \nMohiyaxon Uzoqova \nMukhammadbektosh Khaydarov \nNurlan Maharramli \nPetr Popov \nRasul Karimov \nSariya Kagarmanova \nZiyodabonu Qobiljon qizi", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nMIT License", "### Contributions\n\nThis project was carried out with the help and contributions from dozens of individuals and organizations. We acknowledge and greatly appreciate each and every one of them:\n\nAuthors on the publications (in alphabetical order)\n\nAbror Khaytbaev \nAhsan Wahab \nAigiz Kunafin \nAnoop Babu \nAntonio Laverghetta Jr. \nBehzodbek Moydinboyev \nDr. Duygu Ataman \nEsra Onal \nDr. Francis Tyers \nJamshidbek Mirzakhalov \nDr. John Licato \nDr. Julia Kreutzer \nMammad Hajili \nMokhiyakhon Uzokova \nDr. Orhan Firat \nOtabek Abduraufov \nSardana Ivanova \nShaxnoza Pulatova \nSherzod Kariev \nDr. Sriram Chellappan \n\nTranslators, annotators and dataset contributors (in alphabetical order)\n\nAbilxayr Zholdybai \nAigiz Kunafin \nAkylbek Khamitov \nAlperen Cantez \nAydos Muxammadiyarov \nDoniyorbek Rafikjonov \nErkinbek Vokhabov \nIpek Baris \nIskander Shakirov \nMadina Zokirjonova \nMohiyaxon Uzoqova \nMukhammadbektosh Khaydarov \nNurlan Maharramli \nPetr Popov \nRasul Karimov \nSariya Kagarmanova \nZiyodabonu Qobiljon qizi \n\nIndustry supporters\n\nGoogle Cloud \nKhan Academy Oʻzbek \nThe Foundation for the Preservation and Development of the Bashkir Language \n\nThanks to @mirzakhalov for adding this dataset." ]
[ "TAGS\n#task_categories-translation #annotations_creators-crowdsourced #language_creators-found #multilinguality-translation #size_categories-n<1K #source_datasets-extended|WMT 2020 News Translation Task #language-Azerbaijani #language-Bashkir #language-English #language-Kara-Kalpak #language-Kazakh #language-Kirghiz #language-Russian #language-Yakut #language-Turkish #language-Uzbek #license-mit #arxiv-2109.04593 #region-us \n", "# Dataset Card for turkic_xwmt", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: \n- Repository:Github\n- Paper: URL\n- Leaderboard: \n- Point of Contact: turkicinterlingua@URL", "### Dataset Summary\n\nTo establish a comprehensive and challenging evaluation benchmark for Machine Translation in Turkic languages, we translate a test set originally introduced in WMT 2020 News Translation Task for English-Russian. The original dataset is profesionally translated and consists of sentences from news articles that are both English and Russian-centric. We adopt this evaluation set (X-WMT) and begin efforts to translate it into several Turkic languages. The current version of X-WMT includes covers 8 Turkic languages and 88 language directions with a minimum of 300 sentences per language direction.", "### Supported Tasks and Leaderboards", "### Languages\n\nCurrently covered languages are (besides English and Russian):\n- Azerbaijani (az)\n- Bashkir (ba)\n- Karakalpak (kaa)\n- Kazakh (kk)\n- Kirghiz (ky)\n- Turkish (tr)\n- Sakha (sah)\n- Uzbek (uz)", "## Dataset Structure", "### Data Instances\n\nA random example from the Russian-Uzbek set:", "### Data Fields\n\nEach example has one field \"translation\" that contains two subfields: one per language, e.g. for the Russian-Uzbek set:\n- translation: a dictionary with two subfields:\n - ru: the russian text\n - uz: the uzbek text", "### Data Splits\n\n<details>\n <summary>Click here to show the number of examples per configuration:</summary>\n| | test |\n|:--------|-------:|\n| az-ba | 600 |\n| az-en | 600 |\n| az-kaa | 300 |\n| az-kk | 500 |\n| az-ky | 500 |\n| az-ru | 600 |\n| az-sah | 300 |\n| az-tr | 500 |\n| az-uz | 600 |\n| ba-az | 600 |\n| ba-en | 1000 |\n| ba-kaa | 300 |\n| ba-kk | 700 |\n| ba-ky | 500 |\n| ba-ru | 1000 |\n| ba-sah | 300 |\n| ba-tr | 700 |\n| ba-uz | 900 |\n| en-az | 600 |\n| en-ba | 1000 |\n| en-kaa | 300 |\n| en-kk | 700 |\n| en-ky | 500 |\n| en-ru | 1000 |\n| en-sah | 300 |\n| en-tr | 700 |\n| en-uz | 900 |\n| kaa-az | 300 |\n| kaa-ba | 300 |\n| kaa-en | 300 |\n| kaa-kk | 300 |\n| kaa-ky | 300 |\n| kaa-ru | 300 |\n| kaa-sah | 300 |\n| kaa-tr | 300 |\n| kaa-uz | 300 |\n| kk-az | 500 |\n| kk-ba | 700 |\n| kk-en | 700 |\n| kk-kaa | 300 |\n| kk-ky | 500 |\n| kk-ru | 700 |\n| kk-sah | 300 |\n| kk-tr | 500 |\n| kk-uz | 700 |\n| ky-az | 500 |\n| ky-ba | 500 |\n| ky-en | 500 |\n| ky-kaa | 300 |\n| ky-kk | 500 |\n| ky-ru | 500 |\n| ky-sah | 300 |\n| ky-tr | 400 |\n| ky-uz | 500 |\n| ru-az | 600 |\n| ru-ba | 1000 |\n| ru-en | 1000 |\n| ru-kaa | 300 |\n| ru-kk | 700 |\n| ru-ky | 500 |\n| ru-sah | 300 |\n| ru-tr | 700 |\n| ru-uz | 900 |\n| sah-az | 300 |\n| sah-ba | 300 |\n| sah-en | 300 |\n| sah-kaa | 300 |\n| sah-kk | 300 |\n| sah-ky | 300 |\n| sah-ru | 300 |\n| sah-tr | 300 |\n| sah-uz | 300 |\n| tr-az | 500 |\n| tr-ba | 700 |\n| tr-en | 700 |\n| tr-kaa | 300 |\n| tr-kk | 500 |\n| tr-ky | 400 |\n| tr-ru | 700 |\n| tr-sah | 300 |\n| tr-uz | 600 |\n| uz-az | 600 |\n| uz-ba | 900 |\n| uz-en | 900 |\n| uz-kaa | 300 |\n| uz-kk | 700 |\n| uz-ky | 500 |\n| uz-ru | 900 |\n| uz-sah | 300 |\n| uz-tr | 600 |\n</details>", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?\n\nTranslators, annotators and dataset contributors (in alphabetical order)\n\nAbilxayr Zholdybai \nAigiz Kunafin \nAkylbek Khamitov \nAlperen Cantez \nAydos Muxammadiyarov \nDoniyorbek Rafikjonov \nErkinbek Vokhabov \nIpek Baris \nIskander Shakirov \nMadina Zokirjonova \nMohiyaxon Uzoqova \nMukhammadbektosh Khaydarov \nNurlan Maharramli \nPetr Popov \nRasul Karimov \nSariya Kagarmanova \nZiyodabonu Qobiljon qizi", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nMIT License", "### Contributions\n\nThis project was carried out with the help and contributions from dozens of individuals and organizations. We acknowledge and greatly appreciate each and every one of them:\n\nAuthors on the publications (in alphabetical order)\n\nAbror Khaytbaev \nAhsan Wahab \nAigiz Kunafin \nAnoop Babu \nAntonio Laverghetta Jr. \nBehzodbek Moydinboyev \nDr. Duygu Ataman \nEsra Onal \nDr. Francis Tyers \nJamshidbek Mirzakhalov \nDr. John Licato \nDr. Julia Kreutzer \nMammad Hajili \nMokhiyakhon Uzokova \nDr. Orhan Firat \nOtabek Abduraufov \nSardana Ivanova \nShaxnoza Pulatova \nSherzod Kariev \nDr. Sriram Chellappan \n\nTranslators, annotators and dataset contributors (in alphabetical order)\n\nAbilxayr Zholdybai \nAigiz Kunafin \nAkylbek Khamitov \nAlperen Cantez \nAydos Muxammadiyarov \nDoniyorbek Rafikjonov \nErkinbek Vokhabov \nIpek Baris \nIskander Shakirov \nMadina Zokirjonova \nMohiyaxon Uzoqova \nMukhammadbektosh Khaydarov \nNurlan Maharramli \nPetr Popov \nRasul Karimov \nSariya Kagarmanova \nZiyodabonu Qobiljon qizi \n\nIndustry supporters\n\nGoogle Cloud \nKhan Academy Oʻzbek \nThe Foundation for the Preservation and Development of the Bashkir Language \n\nThanks to @mirzakhalov for adding this dataset." ]
[ 142, 11, 125, 35, 133, 10, 67, 6, 17, 65, 961, 5, 7, 4, 10, 10, 5, 5, 127, 8, 8, 7, 8, 7, 5, 6, 8, 313 ]
[ "passage: TAGS\n#task_categories-translation #annotations_creators-crowdsourced #language_creators-found #multilinguality-translation #size_categories-n<1K #source_datasets-extended|WMT 2020 News Translation Task #language-Azerbaijani #language-Bashkir #language-English #language-Kara-Kalpak #language-Kazakh #language-Kirghiz #language-Russian #language-Yakut #language-Turkish #language-Uzbek #license-mit #arxiv-2109.04593 #region-us \n# Dataset Card for turkic_xwmt## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: \n- Repository:Github\n- Paper: URL\n- Leaderboard: \n- Point of Contact: turkicinterlingua@URL### Dataset Summary\n\nTo establish a comprehensive and challenging evaluation benchmark for Machine Translation in Turkic languages, we translate a test set originally introduced in WMT 2020 News Translation Task for English-Russian. The original dataset is profesionally translated and consists of sentences from news articles that are both English and Russian-centric. We adopt this evaluation set (X-WMT) and begin efforts to translate it into several Turkic languages. The current version of X-WMT includes covers 8 Turkic languages and 88 language directions with a minimum of 300 sentences per language direction.### Supported Tasks and Leaderboards", "passage: ### Languages\n\nCurrently covered languages are (besides English and Russian):\n- Azerbaijani (az)\n- Bashkir (ba)\n- Karakalpak (kaa)\n- Kazakh (kk)\n- Kirghiz (ky)\n- Turkish (tr)\n- Sakha (sah)\n- Uzbek (uz)## Dataset Structure### Data Instances\n\nA random example from the Russian-Uzbek set:### Data Fields\n\nEach example has one field \"translation\" that contains two subfields: one per language, e.g. for the Russian-Uzbek set:\n- translation: a dictionary with two subfields:\n - ru: the russian text\n - uz: the uzbek text", "passage: ### Data Splits\n\n<details>\n <summary>Click here to show the number of examples per configuration:</summary>\n| | test |\n|:--------|-------:|\n| az-ba | 600 |\n| az-en | 600 |\n| az-kaa | 300 |\n| az-kk | 500 |\n| az-ky | 500 |\n| az-ru | 600 |\n| az-sah | 300 |\n| az-tr | 500 |\n| az-uz | 600 |\n| ba-az | 600 |\n| ba-en | 1000 |\n| ba-kaa | 300 |\n| ba-kk | 700 |\n| ba-ky | 500 |\n| ba-ru | 1000 |\n| ba-sah | 300 |\n| ba-tr | 700 |\n| ba-uz | 900 |\n| en-az | 600 |\n| en-ba | 1000 |\n| en-kaa | 300 |\n| en-kk | 700 |\n| en-ky | 500 |\n| en-ru | 1000 |\n| en-sah | 300 |\n| en-tr | 700 |\n| en-uz | 900 |\n| kaa-az | 300 |\n| kaa-ba | 300 |\n| kaa-en | 300 |\n| kaa-kk | 300 |\n| kaa-ky | 300 |\n| kaa-ru | 300 |\n| kaa-sah | 300 |\n| kaa-tr | 300 |\n| kaa-uz | 300 |\n| kk-az | 500 |\n| kk-ba | 700 |\n| kk-en | 700 |\n| kk-kaa | 300 |\n| kk-ky | 500 |\n| kk-ru | 700 |\n| kk-sah | 300 |\n| kk-tr | 500 |\n| kk-uz | 700 |\n| ky-az | 500 |\n| ky-ba | 500 |\n| ky-en | 500 |\n| ky-kaa | 300 |\n| ky-kk | 500 |\n| ky-ru | 500 |\n| ky-sah | 300 |\n| ky-tr | 400 |\n| ky-uz | 500 |\n| ru-az | 600 |\n| ru-ba | 1000 |\n| ru-en | 1000 |\n| ru-kaa | 300 |\n| ru-kk | 700 |\n| ru-ky | 500 |\n| ru-sah | 300 |\n| ru-tr | 700 |\n| ru-uz | 900 |\n| sah-az | 300 |\n| sah-ba | 300 |\n| sah-en | 300 |\n| sah-kaa | 300 |\n| sah-kk | 300 |\n| sah-ky | 300 |\n| sah-ru | 300 |\n| sah-tr | 300 |\n| sah-uz | 300 |\n| tr-az | 500 |\n| tr-ba | 700 |\n| tr-en | 700 |\n| tr-kaa | 300 |\n| tr-kk | 500 |\n| tr-ky | 400 |\n| tr-ru | 700 |\n| tr-sah | 300 |\n| tr-uz | 600 |\n| uz-az | 600 |\n| uz-ba | 900 |\n| uz-en | 900 |\n| uz-kaa | 300 |\n| uz-kk | 700 |\n| uz-ky | 500 |\n| uz-ru | 900 |\n| uz-sah | 300 |\n| uz-tr | 600 |\n</details>## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?\n\nTranslators, annotators and dataset contributors (in alphabetical order)\n\nAbilxayr Zholdybai \nAigiz Kunafin \nAkylbek Khamitov \nAlperen Cantez \nAydos Muxammadiyarov \nDoniyorbek Rafikjonov \nErkinbek Vokhabov \nIpek Baris \nIskander Shakirov \nMadina Zokirjonova \nMohiyaxon Uzoqova \nMukhammadbektosh Khaydarov \nNurlan Maharramli \nPetr Popov \nRasul Karimov \nSariya Kagarmanova \nZiyodabonu Qobiljon qizi### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators### Licensing Information\n\nMIT License" ]
d9f02423150691c4615ffb9403f8b3db1129480c
# Dataset Card for TurkishMovieSentiment: This dataset contains turkish movie reviews. ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://www.kaggle.com/mustfkeskin/turkish-movie-sentiment-analysis-dataset/tasks](https://www.kaggle.com/mustfkeskin/turkish-movie-sentiment-analysis-dataset/tasks) - **Point of Contact:** [Mustafa Keskin](https://www.linkedin.com/in/mustfkeskin/) ### Dataset Summary This data set is a dataset from kaggle consisting of Turkish movie reviews and scored between 0-5. ### Languages The dataset is based on Turkish. ## Dataset Structure ### Data Instances **Example 1:** **Comment:** Jean Reno denince zaten leon filmi gelir akla izlemeyen kalmamıştır ama kaldıysada ee ne duruyorsun hemen izle :), **Film_name:** Sevginin Gücü, **Point:** 5,0 **Example 2:** **Comment:** Bence güzel bi film olmush.İzlenmeli.İnsana şükretmek gerektini hatırlatıyor.Ama cok da poh pohlanacak bi sey yapmamıslar, **Film_name:** Cinderella Man, **Point:** 2,5 ### Data Fields - **comment**(string) : Contatins turkish movie review - **film_name**(string) : Film name in Turkish. - **point**(float) : [0-5] floating point ### Data Splits It is not divided into Train set and Test set. ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations The dataset does not contain any additional annotations. #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Discussion of Social Impact and Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators The dataset was created by [Mustafa Keskin](https://www.linkedin.com/in/mustfkeskin/). ### Licensing Information The data is under the [CC0: Public Domain](https://creativecommons.org/publicdomain/zero/1.0/) ### Citation Information [More Information Needed] ### Contributions Thanks to [@yavuzKomecoglu](https://github.com/yavuzKomecoglu) for adding this dataset.
turkish_movie_sentiment
[ "task_categories:text-classification", "task_ids:sentiment-classification", "task_ids:sentiment-scoring", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:tr", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["tr"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["sentiment-classification", "sentiment-scoring"], "pretty_name": "TurkishMovieSentiment: This dataset contains turkish movie reviews.", "dataset_info": {"features": [{"name": "point", "dtype": "float32"}, {"name": "comment", "dtype": "string"}, {"name": "film_name", "dtype": "string"}], "config_name": "turkishmoviesentiment", "splits": [{"name": "train", "num_bytes": 33954560, "num_examples": 83227}], "download_size": 0, "dataset_size": 33954560}}
2024-01-18T11:17:28+00:00
[]
[ "tr" ]
TAGS #task_categories-text-classification #task_ids-sentiment-classification #task_ids-sentiment-scoring #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Turkish #license-unknown #region-us
# Dataset Card for TurkishMovieSentiment: This dataset contains turkish movie reviews. ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Point of Contact: Mustafa Keskin ### Dataset Summary This data set is a dataset from kaggle consisting of Turkish movie reviews and scored between 0-5. ### Languages The dataset is based on Turkish. ## Dataset Structure ### Data Instances Example 1: Comment: Jean Reno denince zaten leon filmi gelir akla izlemeyen kalmamıştır ama kaldıysada ee ne duruyorsun hemen izle :), Film_name: Sevginin Gücü, Point: 5,0 Example 2: Comment: Bence güzel bi film olmush.İzlenmeli.İnsana şükretmek gerektini hatırlatıyor.Ama cok da poh pohlanacak bi sey yapmamıslar, Film_name: Cinderella Man, Point: 2,5 ### Data Fields - comment(string) : Contatins turkish movie review - film_name(string) : Film name in Turkish. - point(float) : [0-5] floating point ### Data Splits It is not divided into Train set and Test set. ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations The dataset does not contain any additional annotations. #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Discussion of Social Impact and Biases ### Other Known Limitations ## Additional Information ### Dataset Curators The dataset was created by Mustafa Keskin. ### Licensing Information The data is under the CC0: Public Domain ### Contributions Thanks to @yavuzKomecoglu for adding this dataset.
[ "# Dataset Card for TurkishMovieSentiment: This dataset contains turkish movie reviews.", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Point of Contact: Mustafa Keskin", "### Dataset Summary\n\nThis data set is a dataset from kaggle consisting of Turkish movie reviews and scored between 0-5.", "### Languages\n\nThe dataset is based on Turkish.", "## Dataset Structure", "### Data Instances\n\nExample 1:\n\nComment: Jean Reno denince zaten leon filmi gelir akla izlemeyen kalmamıştır ama kaldıysada ee ne duruyorsun hemen izle :),\n\nFilm_name: Sevginin Gücü,\n\nPoint: 5,0\n\nExample 2:\n\nComment: Bence güzel bi film olmush.İzlenmeli.İnsana şükretmek gerektini hatırlatıyor.Ama cok da poh pohlanacak bi sey yapmamıslar,\n\nFilm_name: Cinderella Man,\n\nPoint: 2,5", "### Data Fields\n\n- comment(string) : Contatins turkish movie review\n- film_name(string) : Film name in Turkish.\n- point(float) : [0-5] floating point", "### Data Splits\n\nIt is not divided into Train set and Test set.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations\n\nThe dataset does not contain any additional annotations.", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Discussion of Social Impact and Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThe dataset was created by Mustafa Keskin.", "### Licensing Information\n\nThe data is under the CC0: Public Domain", "### Contributions\n\nThanks to @yavuzKomecoglu for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #task_ids-sentiment-scoring #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Turkish #license-unknown #region-us \n", "# Dataset Card for TurkishMovieSentiment: This dataset contains turkish movie reviews.", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Point of Contact: Mustafa Keskin", "### Dataset Summary\n\nThis data set is a dataset from kaggle consisting of Turkish movie reviews and scored between 0-5.", "### Languages\n\nThe dataset is based on Turkish.", "## Dataset Structure", "### Data Instances\n\nExample 1:\n\nComment: Jean Reno denince zaten leon filmi gelir akla izlemeyen kalmamıştır ama kaldıysada ee ne duruyorsun hemen izle :),\n\nFilm_name: Sevginin Gücü,\n\nPoint: 5,0\n\nExample 2:\n\nComment: Bence güzel bi film olmush.İzlenmeli.İnsana şükretmek gerektini hatırlatıyor.Ama cok da poh pohlanacak bi sey yapmamıslar,\n\nFilm_name: Cinderella Man,\n\nPoint: 2,5", "### Data Fields\n\n- comment(string) : Contatins turkish movie review\n- film_name(string) : Film name in Turkish.\n- point(float) : [0-5] floating point", "### Data Splits\n\nIt is not divided into Train set and Test set.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations\n\nThe dataset does not contain any additional annotations.", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Discussion of Social Impact and Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThe dataset was created by Mustafa Keskin.", "### Licensing Information\n\nThe data is under the CC0: Public Domain", "### Contributions\n\nThanks to @yavuzKomecoglu for adding this dataset." ]
[ 98, 22, 120, 16, 30, 13, 6, 110, 46, 17, 5, 7, 4, 10, 10, 17, 5, 9, 8, 8, 11, 7, 5, 16, 16, 20 ]
[ "passage: TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #task_ids-sentiment-scoring #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Turkish #license-unknown #region-us \n# Dataset Card for TurkishMovieSentiment: This dataset contains turkish movie reviews.## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Point of Contact: Mustafa Keskin### Dataset Summary\n\nThis data set is a dataset from kaggle consisting of Turkish movie reviews and scored between 0-5.### Languages\n\nThe dataset is based on Turkish.## Dataset Structure### Data Instances\n\nExample 1:\n\nComment: Jean Reno denince zaten leon filmi gelir akla izlemeyen kalmamıştır ama kaldıysada ee ne duruyorsun hemen izle :),\n\nFilm_name: Sevginin Gücü,\n\nPoint: 5,0\n\nExample 2:\n\nComment: Bence güzel bi film olmush.İzlenmeli.İnsana şükretmek gerektini hatırlatıyor.Ama cok da poh pohlanacak bi sey yapmamıslar,\n\nFilm_name: Cinderella Man,\n\nPoint: 2,5### Data Fields\n\n- comment(string) : Contatins turkish movie review\n- film_name(string) : Film name in Turkish.\n- point(float) : [0-5] floating point### Data Splits\n\nIt is not divided into Train set and Test set.## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization" ]
5b9a630a07d82e178c177469c2a9eeb0602076a3
# Dataset Card for turkish_ner ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** http://arxiv.org/abs/1702.02363 - **Repository:** [Needs More Information] - **Paper:** http://arxiv.org/abs/1702.02363 - **Leaderboard:** [Needs More Information] - **Point of Contact:** erayyildiz@ktu.edu.tr ### Dataset Summary Automatically annotated Turkish corpus for named entity recognition and text categorization using large-scale gazetteers. The constructed gazetteers contains approximately 300K entities with thousands of fine-grained entity types under 25 different domains. ### Supported Tasks and Leaderboards [Needs More Information] ### Languages Turkish ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits There's only the training set. ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators H. Bahadir Sahin, Caglar Tirkaz, Eray Yildiz, Mustafa Tolga Eren and Omer Ozan Sonmez ### Licensing Information Creative Commons Attribution 4.0 International ### Citation Information @InProceedings@article{DBLP:journals/corr/SahinTYES17, author = {H. Bahadir Sahin and Caglar Tirkaz and Eray Yildiz and Mustafa Tolga Eren and Omer Ozan Sonmez}, title = {Automatically Annotated Turkish Corpus for Named Entity Recognition and Text Categorization using Large-Scale Gazetteers}, journal = {CoRR}, volume = {abs/1702.02363}, year = {2017}, url = {http://arxiv.org/abs/1702.02363}, archivePrefix = {arXiv}, eprint = {1702.02363}, timestamp = {Mon, 13 Aug 2018 16:46:36 +0200}, biburl = {https://dblp.org/rec/journals/corr/SahinTYES17.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ### Contributions Thanks to [@merveenoyan](https://github.com/merveenoyan) for adding this dataset.
turkish_ner
[ "task_categories:token-classification", "task_ids:named-entity-recognition", "annotations_creators:machine-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:tr", "license:cc-by-4.0", "arxiv:1702.02363", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["machine-generated"], "language_creators": ["expert-generated"], "language": ["tr"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["token-classification"], "task_ids": ["named-entity-recognition"], "pretty_name": "TurkishNer", "dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "domain", "dtype": {"class_label": {"names": {"0": "architecture", "1": "basketball", "2": "book", "3": "business", "4": "education", "5": "fictional_universe", "6": "film", "7": "food", "8": "geography", "9": "government", "10": "law", "11": "location", "12": "military", "13": "music", "14": "opera", "15": "organization", "16": "people", "17": "religion", "18": "royalty", "19": "soccer", "20": "sports", "21": "theater", "22": "time", "23": "travel", "24": "tv"}}}}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PERSON", "2": "I-PERSON", "3": "B-ORGANIZATION", "4": "I-ORGANIZATION", "5": "B-LOCATION", "6": "I-LOCATION", "7": "B-MISC", "8": "I-MISC"}}}}], "splits": [{"name": "train", "num_bytes": 177658278, "num_examples": 532629}], "download_size": 204393976, "dataset_size": 177658278}}
2024-01-18T11:17:29+00:00
[ "1702.02363" ]
[ "tr" ]
TAGS #task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-machine-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Turkish #license-cc-by-4.0 #arxiv-1702.02363 #region-us
# Dataset Card for turkish_ner ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: URL - Leaderboard: - Point of Contact: erayyildiz@URL ### Dataset Summary Automatically annotated Turkish corpus for named entity recognition and text categorization using large-scale gazetteers. The constructed gazetteers contains approximately 300K entities with thousands of fine-grained entity types under 25 different domains. ### Supported Tasks and Leaderboards ### Languages Turkish ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits There's only the training set. ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators H. Bahadir Sahin, Caglar Tirkaz, Eray Yildiz, Mustafa Tolga Eren and Omer Ozan Sonmez ### Licensing Information Creative Commons Attribution 4.0 International @InProceedings@article{DBLP:journals/corr/SahinTYES17, author = {H. Bahadir Sahin and Caglar Tirkaz and Eray Yildiz and Mustafa Tolga Eren and Omer Ozan Sonmez}, title = {Automatically Annotated Turkish Corpus for Named Entity Recognition and Text Categorization using Large-Scale Gazetteers}, journal = {CoRR}, volume = {abs/1702.02363}, year = {2017}, url = {URL archivePrefix = {arXiv}, eprint = {1702.02363}, timestamp = {Mon, 13 Aug 2018 16:46:36 +0200}, biburl = {URL bibsource = {dblp computer science bibliography, URL} } ### Contributions Thanks to @merveenoyan for adding this dataset.
[ "# Dataset Card for turkish_ner", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: URL\n- Leaderboard: \n- Point of Contact: erayyildiz@URL", "### Dataset Summary\n\nAutomatically annotated Turkish corpus for named entity recognition and text categorization using large-scale gazetteers. The constructed gazetteers contains approximately 300K entities with thousands of fine-grained entity types under 25 different domains.", "### Supported Tasks and Leaderboards", "### Languages\n\nTurkish", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits\n\nThere's only the training set.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nH. Bahadir Sahin, Caglar Tirkaz, Eray Yildiz, Mustafa Tolga Eren and Omer Ozan Sonmez", "### Licensing Information\n\nCreative Commons Attribution 4.0 International\n\n\n\n@InProceedings@article{DBLP:journals/corr/SahinTYES17,\n author = {H. Bahadir Sahin and\n Caglar Tirkaz and\n Eray Yildiz and\n Mustafa Tolga Eren and\n Omer Ozan Sonmez},\n title = {Automatically Annotated Turkish Corpus for Named Entity Recognition\n and Text Categorization using Large-Scale Gazetteers},\n journal = {CoRR},\n volume = {abs/1702.02363},\n year = {2017},\n url = {URL\n archivePrefix = {arXiv},\n eprint = {1702.02363},\n timestamp = {Mon, 13 Aug 2018 16:46:36 +0200},\n biburl = {URL\n bibsource = {dblp computer science bibliography, URL}\n}", "### Contributions\n\nThanks to @merveenoyan for adding this dataset." ]
[ "TAGS\n#task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-machine-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Turkish #license-cc-by-4.0 #arxiv-1702.02363 #region-us \n", "# Dataset Card for turkish_ner", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: URL\n- Leaderboard: \n- Point of Contact: erayyildiz@URL", "### Dataset Summary\n\nAutomatically annotated Turkish corpus for named entity recognition and text categorization using large-scale gazetteers. The constructed gazetteers contains approximately 300K entities with thousands of fine-grained entity types under 25 different domains.", "### Supported Tasks and Leaderboards", "### Languages\n\nTurkish", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits\n\nThere's only the training set.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nH. Bahadir Sahin, Caglar Tirkaz, Eray Yildiz, Mustafa Tolga Eren and Omer Ozan Sonmez", "### Licensing Information\n\nCreative Commons Attribution 4.0 International\n\n\n\n@InProceedings@article{DBLP:journals/corr/SahinTYES17,\n author = {H. Bahadir Sahin and\n Caglar Tirkaz and\n Eray Yildiz and\n Mustafa Tolga Eren and\n Omer Ozan Sonmez},\n title = {Automatically Annotated Turkish Corpus for Named Entity Recognition\n and Text Categorization using Large-Scale Gazetteers},\n journal = {CoRR},\n volume = {abs/1702.02363},\n year = {2017},\n url = {URL\n archivePrefix = {arXiv},\n eprint = {1702.02363},\n timestamp = {Mon, 13 Aug 2018 16:46:36 +0200},\n biburl = {URL\n bibsource = {dblp computer science bibliography, URL}\n}", "### Contributions\n\nThanks to @merveenoyan for adding this dataset." ]
[ 108, 9, 120, 32, 63, 10, 6, 6, 6, 5, 13, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 37, 195, 18 ]
[ "passage: TAGS\n#task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-machine-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Turkish #license-cc-by-4.0 #arxiv-1702.02363 #region-us \n# Dataset Card for turkish_ner## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: URL\n- Leaderboard: \n- Point of Contact: erayyildiz@URL### Dataset Summary\n\nAutomatically annotated Turkish corpus for named entity recognition and text categorization using large-scale gazetteers. The constructed gazetteers contains approximately 300K entities with thousands of fine-grained entity types under 25 different domains.### Supported Tasks and Leaderboards### Languages\n\nTurkish## Dataset Structure### Data Instances### Data Fields### Data Splits\n\nThere's only the training set.## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information" ]
d16d987819a416520ad5a8ca08d8f57017bdc972
# Dataset Card for Turkish Product Reviews ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Repository:** [turkish-text-data](https://github.com/fthbrmnby/turkish-text-data) - **Point of Contact:** [Fatih Barmanbay](https://github.com/fthbrmnby) ### Dataset Summary This Turkish Product Reviews Dataset contains 235.165 product reviews collected online. There are 220.284 positive, 14881 negative reviews. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The dataset is based on Turkish. ## Dataset Structure ### Data Instances **Example 1:** **sentence:** beklentimin altında bir ürün kaliteli değil **sentiment:** 0 (negative) **Example 2:** **sentence:** fiyat ve performans olarak gayet iyi **sentiment:** 1 (positive) ### Data Fields - **sentence**(string) : Contatins turkish product review - **sentiment**(int) : 0 (negative) or 1 (positive) ### Data Splits It is not divided into Train set and Test set. ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations The dataset does not contain any additional annotations. #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators The dataset was created by [Fatih Barmanbay](https://github.com/fthbrmnby). ### Licensing Information The data is under the [CC-BY-SA-4.0 License](https://github.com/fthbrmnby/turkish-text-data/blob/master/LICENCE) ### Citation Information No citation available for this dataset. ### Contributions Thanks to [@basakbuluz](https://github.com/basakbuluz) for adding this dataset.
turkish_product_reviews
[ "task_categories:text-classification", "task_ids:sentiment-classification", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:tr", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["tr"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["sentiment-classification"], "pretty_name": "Turkish Product Reviews", "dataset_info": {"features": [{"name": "sentence", "dtype": "string"}, {"name": "sentiment", "dtype": {"class_label": {"names": {"0": "negative", "1": "positive"}}}}], "splits": [{"name": "train", "num_bytes": 43369710, "num_examples": 235165}], "download_size": 13184332, "dataset_size": 43369710}}
2024-01-18T11:17:30+00:00
[]
[ "tr" ]
TAGS #task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Turkish #license-unknown #region-us
# Dataset Card for Turkish Product Reviews ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Repository: turkish-text-data - Point of Contact: Fatih Barmanbay ### Dataset Summary This Turkish Product Reviews Dataset contains 235.165 product reviews collected online. There are 220.284 positive, 14881 negative reviews. ### Supported Tasks and Leaderboards ### Languages The dataset is based on Turkish. ## Dataset Structure ### Data Instances Example 1: sentence: beklentimin altında bir ürün kaliteli değil sentiment: 0 (negative) Example 2: sentence: fiyat ve performans olarak gayet iyi sentiment: 1 (positive) ### Data Fields - sentence(string) : Contatins turkish product review - sentiment(int) : 0 (negative) or 1 (positive) ### Data Splits It is not divided into Train set and Test set. ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations The dataset does not contain any additional annotations. #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators The dataset was created by Fatih Barmanbay. ### Licensing Information The data is under the CC-BY-SA-4.0 License No citation available for this dataset. ### Contributions Thanks to @basakbuluz for adding this dataset.
[ "# Dataset Card for Turkish Product Reviews", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Repository: turkish-text-data\n- Point of Contact: Fatih Barmanbay", "### Dataset Summary\n\nThis Turkish Product Reviews Dataset contains 235.165 product reviews collected online. There are 220.284 positive, 14881 negative reviews.", "### Supported Tasks and Leaderboards", "### Languages\n\nThe dataset is based on Turkish.", "## Dataset Structure", "### Data Instances\n\nExample 1:\n\nsentence: beklentimin altında bir ürün kaliteli değil\n\nsentiment: 0 (negative)\n\nExample 2:\n\nsentence: fiyat ve performans olarak gayet iyi\n\nsentiment: 1 (positive)", "### Data Fields\n\n- sentence(string) : Contatins turkish product review\n- sentiment(int) : 0 (negative) or 1 (positive)", "### Data Splits\n\nIt is not divided into Train set and Test set.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations\n\nThe dataset does not contain any additional annotations.", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThe dataset was created by Fatih Barmanbay.", "### Licensing Information\n\nThe data is under the CC-BY-SA-4.0 License\n\n\n\nNo citation available for this dataset.", "### Contributions\n\nThanks to @basakbuluz for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Turkish #license-unknown #region-us \n", "# Dataset Card for Turkish Product Reviews", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Repository: turkish-text-data\n- Point of Contact: Fatih Barmanbay", "### Dataset Summary\n\nThis Turkish Product Reviews Dataset contains 235.165 product reviews collected online. There are 220.284 positive, 14881 negative reviews.", "### Supported Tasks and Leaderboards", "### Languages\n\nThe dataset is based on Turkish.", "## Dataset Structure", "### Data Instances\n\nExample 1:\n\nsentence: beklentimin altında bir ürün kaliteli değil\n\nsentiment: 0 (negative)\n\nExample 2:\n\nsentence: fiyat ve performans olarak gayet iyi\n\nsentiment: 1 (positive)", "### Data Fields\n\n- sentence(string) : Contatins turkish product review\n- sentiment(int) : 0 (negative) or 1 (positive)", "### Data Splits\n\nIt is not divided into Train set and Test set.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations\n\nThe dataset does not contain any additional annotations.", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThe dataset was created by Fatih Barmanbay.", "### Licensing Information\n\nThe data is under the CC-BY-SA-4.0 License\n\n\n\nNo citation available for this dataset.", "### Contributions\n\nThanks to @basakbuluz for adding this dataset." ]
[ 87, 10, 120, 24, 37, 10, 13, 6, 44, 35, 17, 5, 7, 4, 10, 10, 17, 5, 9, 8, 8, 7, 8, 7, 5, 17, 28, 18 ]
[ "passage: TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Turkish #license-unknown #region-us \n# Dataset Card for Turkish Product Reviews## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Repository: turkish-text-data\n- Point of Contact: Fatih Barmanbay### Dataset Summary\n\nThis Turkish Product Reviews Dataset contains 235.165 product reviews collected online. There are 220.284 positive, 14881 negative reviews.### Supported Tasks and Leaderboards### Languages\n\nThe dataset is based on Turkish.## Dataset Structure### Data Instances\n\nExample 1:\n\nsentence: beklentimin altında bir ürün kaliteli değil\n\nsentiment: 0 (negative)\n\nExample 2:\n\nsentence: fiyat ve performans olarak gayet iyi\n\nsentiment: 1 (positive)### Data Fields\n\n- sentence(string) : Contatins turkish product review\n- sentiment(int) : 0 (negative) or 1 (positive)### Data Splits\n\nIt is not divided into Train set and Test set.## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations\n\nThe dataset does not contain any additional annotations.#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases" ]
c5929cc0be4a2ab45136414f4c33fb95fa236fd8
# Dataset Card for turkish_shrinked_ner ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.kaggle.com/behcetsenturk/shrinked-twnertc-turkish-ner-data-by-kuzgunlar - **Repository:** [Needs More Information] - **Paper:** [Needs More Information] - **Leaderboard:** [Needs More Information] - **Point of Contact:** https://www.kaggle.com/behcetsenturk ### Dataset Summary Shrinked processed version (48 entity type) of the turkish_ner. Original turkish_ner dataset: Automatically annotated Turkish corpus for named entity recognition and text categorization using large-scale gazetteers. The constructed gazetteers contains approximately 300K entities with thousands of fine-grained entity types under 25 different domains. Shrinked entity types are: academic, academic_person, aircraft, album_person, anatomy, animal, architect_person, capital, chemical, clothes, country, culture, currency, date, food, genre, government, government_person, language, location, material, measure, medical, military, military_person, nation, newspaper, organization, organization_person, person, production_art_music, production_art_music_person, quantity, religion, science, shape, ship, software, space, space_person, sport, sport_name, sport_person, structure, subject, tech, train, vehicle ### Supported Tasks and Leaderboards [Needs More Information] ### Languages Turkish ## Dataset Structure ### Data Instances [Needs More Information] ### Data Fields [Needs More Information] ### Data Splits There's only the training set. ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators Behcet Senturk ### Licensing Information Creative Commons Attribution 4.0 International ### Citation Information [Needs More Information] ### Contributions Thanks to [@bhctsntrk](https://github.com/bhctsntrk) for adding this dataset.
turkish_shrinked_ner
[ "task_categories:token-classification", "task_ids:named-entity-recognition", "annotations_creators:machine-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:extended|other-turkish_ner", "language:tr", "license:cc-by-4.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["machine-generated"], "language_creators": ["expert-generated"], "language": ["tr"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["extended|other-turkish_ner"], "task_categories": ["token-classification"], "task_ids": ["named-entity-recognition"], "pretty_name": "TurkishShrinkedNer", "dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-academic", "2": "I-academic", "3": "B-academic_person", "4": "I-academic_person", "5": "B-aircraft", "6": "I-aircraft", "7": "B-album_person", "8": "I-album_person", "9": "B-anatomy", "10": "I-anatomy", "11": "B-animal", "12": "I-animal", "13": "B-architect_person", "14": "I-architect_person", "15": "B-capital", "16": "I-capital", "17": "B-chemical", "18": "I-chemical", "19": "B-clothes", "20": "I-clothes", "21": "B-country", "22": "I-country", "23": "B-culture", "24": "I-culture", "25": "B-currency", "26": "I-currency", "27": "B-date", "28": "I-date", "29": "B-food", "30": "I-food", "31": "B-genre", "32": "I-genre", "33": "B-government", "34": "I-government", "35": "B-government_person", "36": "I-government_person", "37": "B-language", "38": "I-language", "39": "B-location", "40": "I-location", "41": "B-material", "42": "I-material", "43": "B-measure", "44": "I-measure", "45": "B-medical", "46": "I-medical", "47": "B-military", "48": "I-military", "49": "B-military_person", "50": "I-military_person", "51": "B-nation", "52": "I-nation", "53": "B-newspaper", "54": "I-newspaper", "55": "B-organization", "56": "I-organization", "57": "B-organization_person", "58": "I-organization_person", "59": "B-person", "60": "I-person", "61": "B-production_art_music", "62": "I-production_art_music", "63": "B-production_art_music_person", "64": "I-production_art_music_person", "65": "B-quantity", "66": "I-quantity", "67": "B-religion", "68": "I-religion", "69": "B-science", "70": "I-science", "71": "B-shape", "72": "I-shape", "73": "B-ship", "74": "I-ship", "75": "B-software", "76": "I-software", "77": "B-space", "78": "I-space", "79": "B-space_person", "80": "I-space_person", "81": "B-sport", "82": "I-sport", "83": "B-sport_name", "84": "I-sport_name", "85": "B-sport_person", "86": "I-sport_person", "87": "B-structure", "88": "I-structure", "89": "B-subject", "90": "I-subject", "91": "B-tech", "92": "I-tech", "93": "B-train", "94": "I-train", "95": "B-vehicle", "96": "I-vehicle"}}}}], "splits": [{"name": "train", "num_bytes": 200728389, "num_examples": 614515}], "download_size": 0, "dataset_size": 200728389}}
2024-01-18T11:17:31+00:00
[]
[ "tr" ]
TAGS #task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-machine-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-extended|other-turkish_ner #language-Turkish #license-cc-by-4.0 #region-us
# Dataset Card for turkish_shrinked_ner ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: - Leaderboard: - Point of Contact: URL ### Dataset Summary Shrinked processed version (48 entity type) of the turkish_ner. Original turkish_ner dataset: Automatically annotated Turkish corpus for named entity recognition and text categorization using large-scale gazetteers. The constructed gazetteers contains approximately 300K entities with thousands of fine-grained entity types under 25 different domains. Shrinked entity types are: academic, academic_person, aircraft, album_person, anatomy, animal, architect_person, capital, chemical, clothes, country, culture, currency, date, food, genre, government, government_person, language, location, material, measure, medical, military, military_person, nation, newspaper, organization, organization_person, person, production_art_music, production_art_music_person, quantity, religion, science, shape, ship, software, space, space_person, sport, sport_name, sport_person, structure, subject, tech, train, vehicle ### Supported Tasks and Leaderboards ### Languages Turkish ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits There's only the training set. ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators Behcet Senturk ### Licensing Information Creative Commons Attribution 4.0 International ### Contributions Thanks to @bhctsntrk for adding this dataset.
[ "# Dataset Card for turkish_shrinked_ner", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: \n- Leaderboard: \n- Point of Contact: URL", "### Dataset Summary\n\nShrinked processed version (48 entity type) of the turkish_ner.\n\nOriginal turkish_ner dataset: Automatically annotated Turkish corpus for named entity recognition and text categorization using large-scale gazetteers. The constructed gazetteers contains approximately 300K entities with thousands of fine-grained entity types under 25 different domains.\n\nShrinked entity types are: academic, academic_person, aircraft, album_person, anatomy, animal, architect_person, capital, chemical, clothes, country, culture, currency, date, food, genre, government, government_person, language, location, material, measure, medical, military, military_person, nation, newspaper, organization, organization_person, person, production_art_music, production_art_music_person, quantity, religion, science, shape, ship, software, space, space_person, sport, sport_name, sport_person, structure, subject, tech, train, vehicle", "### Supported Tasks and Leaderboards", "### Languages\n\nTurkish", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits\n\nThere's only the training set.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nBehcet Senturk", "### Licensing Information\n\nCreative Commons Attribution 4.0 International", "### Contributions\n\nThanks to @bhctsntrk for adding this dataset." ]
[ "TAGS\n#task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-machine-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-extended|other-turkish_ner #language-Turkish #license-cc-by-4.0 #region-us \n", "# Dataset Card for turkish_shrinked_ner", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: \n- Leaderboard: \n- Point of Contact: URL", "### Dataset Summary\n\nShrinked processed version (48 entity type) of the turkish_ner.\n\nOriginal turkish_ner dataset: Automatically annotated Turkish corpus for named entity recognition and text categorization using large-scale gazetteers. The constructed gazetteers contains approximately 300K entities with thousands of fine-grained entity types under 25 different domains.\n\nShrinked entity types are: academic, academic_person, aircraft, album_person, anatomy, animal, architect_person, capital, chemical, clothes, country, culture, currency, date, food, genre, government, government_person, language, location, material, measure, medical, military, military_person, nation, newspaper, organization, organization_person, person, production_art_music, production_art_music_person, quantity, religion, science, shape, ship, software, space, space_person, sport, sport_name, sport_person, structure, subject, tech, train, vehicle", "### Supported Tasks and Leaderboards", "### Languages\n\nTurkish", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits\n\nThere's only the training set.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nBehcet Senturk", "### Licensing Information\n\nCreative Commons Attribution 4.0 International", "### Contributions\n\nThanks to @bhctsntrk for adding this dataset." ]
[ 109, 13, 120, 26, 225, 10, 6, 6, 6, 5, 13, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 10, 11, 19 ]
[ "passage: TAGS\n#task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-machine-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-extended|other-turkish_ner #language-Turkish #license-cc-by-4.0 #region-us \n# Dataset Card for turkish_shrinked_ner## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: \n- Leaderboard: \n- Point of Contact: URL### Dataset Summary\n\nShrinked processed version (48 entity type) of the turkish_ner.\n\nOriginal turkish_ner dataset: Automatically annotated Turkish corpus for named entity recognition and text categorization using large-scale gazetteers. The constructed gazetteers contains approximately 300K entities with thousands of fine-grained entity types under 25 different domains.\n\nShrinked entity types are: academic, academic_person, aircraft, album_person, anatomy, animal, architect_person, capital, chemical, clothes, country, culture, currency, date, food, genre, government, government_person, language, location, material, measure, medical, military, military_person, nation, newspaper, organization, organization_person, person, production_art_music, production_art_music_person, quantity, religion, science, shape, ship, software, space, space_person, sport, sport_name, sport_person, structure, subject, tech, train, vehicle### Supported Tasks and Leaderboards" ]
59904695b2039af98825a9576bade1533107db99
# Dataset Card for Turku NER corpus ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://turkunlp.org/fin-ner.html - **Repository:** https://github.com/TurkuNLP/turku-ner-corpus/ - **Paper:** https://www.aclweb.org/anthology/2020.lrec-1.567/ - **Leaderboard:** [If the dataset supports an active leaderboard, add link here]() - **Point of Contact:** {jouni.a.luoma,mhtoin,maria.h.pyykonen,mavela,sampo.pyysalo}@utu.f ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@abhishekkrthakur](https://github.com/abhishekkrthakur) for adding this dataset.
turku_ner_corpus
[ "task_categories:token-classification", "task_ids:named-entity-recognition", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:fi", "license:cc-by-nc-sa-4.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["expert-generated"], "language": ["fi"], "license": ["cc-by-nc-sa-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["token-classification"], "task_ids": ["named-entity-recognition"], "pretty_name": "Turku NER corpus", "dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "B-DATE", "1": "B-EVENT", "2": "B-LOC", "3": "B-ORG", "4": "B-PER", "5": "B-PRO", "6": "I-DATE", "7": "I-EVENT", "8": "I-LOC", "9": "I-ORG", "10": "I-PER", "11": "I-PRO", "12": "O"}}}}], "splits": [{"name": "train", "num_bytes": 3257447, "num_examples": 12217}, {"name": "validation", "num_bytes": 364223, "num_examples": 1364}, {"name": "test", "num_bytes": 416644, "num_examples": 1555}], "download_size": 1659911, "dataset_size": 4038314}}
2024-01-18T11:17:32+00:00
[]
[ "fi" ]
TAGS #task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Finnish #license-cc-by-nc-sa-4.0 #region-us
# Dataset Card for Turku NER corpus ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: URL - Paper: URL - Leaderboard: [If the dataset supports an active leaderboard, add link here]() - Point of Contact: {jouni.a.luoma,mhtoin,maria.h.pyykonen,mavela,sampo.pyysalo}@utu.f ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @abhishekkrthakur for adding this dataset.
[ "# Dataset Card for Turku NER corpus", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: URL\n- Leaderboard: [If the dataset supports an active leaderboard, add link here]()\n- Point of Contact: {jouni.a.luoma,mhtoin,maria.h.pyykonen,mavela,sampo.pyysalo}@utu.f", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ "TAGS\n#task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Finnish #license-cc-by-nc-sa-4.0 #region-us \n", "# Dataset Card for Turku NER corpus", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: URL\n- Leaderboard: [If the dataset supports an active leaderboard, add link here]()\n- Point of Contact: {jouni.a.luoma,mhtoin,maria.h.pyykonen,mavela,sampo.pyysalo}@utu.f", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ 104, 9, 120, 82, 6, 10, 4, 6, 6, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 20 ]
[ "passage: TAGS\n#task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Finnish #license-cc-by-nc-sa-4.0 #region-us \n# Dataset Card for Turku NER corpus## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: URL\n- Leaderboard: [If the dataset supports an active leaderboard, add link here]()\n- Point of Contact: {jouni.a.luoma,mhtoin,maria.h.pyykonen,mavela,sampo.pyysalo}@utu.f### Dataset Summary### Supported Tasks and Leaderboards### Languages## Dataset Structure### Data Instances### Data Fields### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators### Licensing Information### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
b3a375baf0f409c77e6bc7aa35102b7b3534f8be
# Dataset Card for tweet_eval ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Needs More Information] - **Repository:** [GitHub](https://github.com/cardiffnlp/tweeteval) - **Paper:** [EMNLP Paper](https://arxiv.org/pdf/2010.12421.pdf) - **Leaderboard:** [GitHub Leaderboard](https://github.com/cardiffnlp/tweeteval) - **Point of Contact:** [Needs More Information] ### Dataset Summary TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. The tasks include - irony, hate, offensive, stance, emoji, emotion, and sentiment. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits. ### Supported Tasks and Leaderboards - `text_classification`: The dataset can be trained using a SentenceClassification model from HuggingFace transformers. ### Languages The text in the dataset is in English, as spoken by Twitter users. ## Dataset Structure ### Data Instances An instance from `emoji` config: ``` {'label': 12, 'text': 'Sunday afternoon walking through Venice in the sun with @user ️ ️ ️ @ Abbot Kinney, Venice'} ``` An instance from `emotion` config: ``` {'label': 2, 'text': "“Worry is a down payment on a problem you may never have'. \xa0Joyce Meyer. #motivation #leadership #worry"} ``` An instance from `hate` config: ``` {'label': 0, 'text': '@user nice new signage. Are you not concerned by Beatlemania -style hysterical crowds crongregating on you…'} ``` An instance from `irony` config: ``` {'label': 1, 'text': 'seeing ppl walking w/ crutches makes me really excited for the next 3 weeks of my life'} ``` An instance from `offensive` config: ``` {'label': 0, 'text': '@user Bono... who cares. Soon people will understand that they gain nothing from following a phony celebrity. Become a Leader of your people instead or help and support your fellow countrymen.'} ``` An instance from `sentiment` config: ``` {'label': 2, 'text': '"QT @user In the original draft of the 7th book, Remus Lupin survived the Battle of Hogwarts. #HappyBirthdayRemusLupin"'} ``` An instance from `stance_abortion` config: ``` {'label': 1, 'text': 'we remind ourselves that love means to be willing to give until it hurts - Mother Teresa'} ``` An instance from `stance_atheism` config: ``` {'label': 1, 'text': '@user Bless Almighty God, Almighty Holy Spirit and the Messiah. #SemST'} ``` An instance from `stance_climate` config: ``` {'label': 0, 'text': 'Why Is The Pope Upset? via @user #UnzippedTruth #PopeFrancis #SemST'} ``` An instance from `stance_feminist` config: ``` {'label': 1, 'text': "@user @user is the UK's answer to @user and @user #GamerGate #SemST"} ``` An instance from `stance_hillary` config: ``` {'label': 1, 'text': "If a man demanded staff to get him an ice tea he'd be called a sexists elitist pig.. Oink oink #Hillary #SemST"} ``` ### Data Fields For `emoji` config: - `text`: a `string` feature containing the tweet. - `label`: an `int` classification label with the following mapping: `0`: ❤ `1`: 😍 `2`: 😂 `3`: 💕 `4`: 🔥 `5`: 😊 `6`: 😎 `7`: ✨ `8`: 💙 `9`: 😘 `10`: 📷 `11`: 🇺🇸 `12`: ☀ `13`: 💜 `14`: 😉 `15`: 💯 `16`: 😁 `17`: 🎄 `18`: 📸 `19`: 😜 For `emotion` config: - `text`: a `string` feature containing the tweet. - `label`: an `int` classification label with the following mapping: `0`: anger `1`: joy `2`: optimism `3`: sadness For `hate` config: - `text`: a `string` feature containing the tweet. - `label`: an `int` classification label with the following mapping: `0`: non-hate `1`: hate For `irony` config: - `text`: a `string` feature containing the tweet. - `label`: an `int` classification label with the following mapping: `0`: non_irony `1`: irony For `offensive` config: - `text`: a `string` feature containing the tweet. - `label`: an `int` classification label with the following mapping: `0`: non-offensive `1`: offensive For `sentiment` config: - `text`: a `string` feature containing the tweet. - `label`: an `int` classification label with the following mapping: `0`: negative `1`: neutral `2`: positive For `stance_abortion` config: - `text`: a `string` feature containing the tweet. - `label`: an `int` classification label with the following mapping: `0`: none `1`: against `2`: favor For `stance_atheism` config: - `text`: a `string` feature containing the tweet. - `label`: an `int` classification label with the following mapping: `0`: none `1`: against `2`: favor For `stance_climate` config: - `text`: a `string` feature containing the tweet. - `label`: an `int` classification label with the following mapping: `0`: none `1`: against `2`: favor For `stance_feminist` config: - `text`: a `string` feature containing the tweet. - `label`: an `int` classification label with the following mapping: `0`: none `1`: against `2`: favor For `stance_hillary` config: - `text`: a `string` feature containing the tweet. - `label`: an `int` classification label with the following mapping: `0`: none `1`: against `2`: favor ### Data Splits | name | train | validation | test | | --------------- | ----- | ---------- | ----- | | emoji | 45000 | 5000 | 50000 | | emotion | 3257 | 374 | 1421 | | hate | 9000 | 1000 | 2970 | | irony | 2862 | 955 | 784 | | offensive | 11916 | 1324 | 860 | | sentiment | 45615 | 2000 | 12284 | | stance_abortion | 587 | 66 | 280 | | stance_atheism | 461 | 52 | 220 | | stance_climate | 355 | 40 | 169 | | stance_feminist | 597 | 67 | 285 | | stance_hillary | 620 | 69 | 295 | ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators Francesco Barbieri, Jose Camacho-Collados, Luis Espiinosa-Anke and Leonardo Neves through Cardiff NLP. ### Licensing Information This is not a single dataset, therefore each subset has its own license (the collection itself does not have additional restrictions). All of the datasets require complying with Twitter [Terms Of Service](https://twitter.com/tos) and Twitter API [Terms Of Service](https://developer.twitter.com/en/developer-terms/agreement-and-policy) Additionally the license are: - emoji: Undefined - emotion(EmoInt): Undefined - hate (HateEval): Need permission [here](http://hatespeech.di.unito.it/hateval.html) - irony: Undefined - Offensive: Undefined - Sentiment: [Creative Commons Attribution 3.0 Unported License](https://groups.google.com/g/semevaltweet/c/k5DDcvVb_Vo/m/zEOdECFyBQAJ) - Stance: Undefined ### Citation Information ``` @inproceedings{barbieri2020tweeteval, title={{TweetEval:Unified Benchmark and Comparative Evaluation for Tweet Classification}}, author={Barbieri, Francesco and Camacho-Collados, Jose and Espinosa-Anke, Luis and Neves, Leonardo}, booktitle={Proceedings of Findings of EMNLP}, year={2020} } ``` If you use any of the TweetEval datasets, please cite their original publications: #### Emotion Recognition: ``` @inproceedings{mohammad2018semeval, title={Semeval-2018 task 1: Affect in tweets}, author={Mohammad, Saif and Bravo-Marquez, Felipe and Salameh, Mohammad and Kiritchenko, Svetlana}, booktitle={Proceedings of the 12th international workshop on semantic evaluation}, pages={1--17}, year={2018} } ``` #### Emoji Prediction: ``` @inproceedings{barbieri2018semeval, title={Semeval 2018 task 2: Multilingual emoji prediction}, author={Barbieri, Francesco and Camacho-Collados, Jose and Ronzano, Francesco and Espinosa-Anke, Luis and Ballesteros, Miguel and Basile, Valerio and Patti, Viviana and Saggion, Horacio}, booktitle={Proceedings of The 12th International Workshop on Semantic Evaluation}, pages={24--33}, year={2018} } ``` #### Irony Detection: ``` @inproceedings{van2018semeval, title={Semeval-2018 task 3: Irony detection in english tweets}, author={Van Hee, Cynthia and Lefever, Els and Hoste, V{\'e}ronique}, booktitle={Proceedings of The 12th International Workshop on Semantic Evaluation}, pages={39--50}, year={2018} } ``` #### Hate Speech Detection: ``` @inproceedings{basile-etal-2019-semeval, title = "{S}em{E}val-2019 Task 5: Multilingual Detection of Hate Speech Against Immigrants and Women in {T}witter", author = "Basile, Valerio and Bosco, Cristina and Fersini, Elisabetta and Nozza, Debora and Patti, Viviana and Rangel Pardo, Francisco Manuel and Rosso, Paolo and Sanguinetti, Manuela", booktitle = "Proceedings of the 13th International Workshop on Semantic Evaluation", year = "2019", address = "Minneapolis, Minnesota, USA", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/S19-2007", doi = "10.18653/v1/S19-2007", pages = "54--63" } ``` #### Offensive Language Identification: ``` @inproceedings{zampieri2019semeval, title={SemEval-2019 Task 6: Identifying and Categorizing Offensive Language in Social Media (OffensEval)}, author={Zampieri, Marcos and Malmasi, Shervin and Nakov, Preslav and Rosenthal, Sara and Farra, Noura and Kumar, Ritesh}, booktitle={Proceedings of the 13th International Workshop on Semantic Evaluation}, pages={75--86}, year={2019} } ``` #### Sentiment Analysis: ``` @inproceedings{rosenthal2017semeval, title={SemEval-2017 task 4: Sentiment analysis in Twitter}, author={Rosenthal, Sara and Farra, Noura and Nakov, Preslav}, booktitle={Proceedings of the 11th international workshop on semantic evaluation (SemEval-2017)}, pages={502--518}, year={2017} } ``` #### Stance Detection: ``` @inproceedings{mohammad2016semeval, title={Semeval-2016 task 6: Detecting stance in tweets}, author={Mohammad, Saif and Kiritchenko, Svetlana and Sobhani, Parinaz and Zhu, Xiaodan and Cherry, Colin}, booktitle={Proceedings of the 10th International Workshop on Semantic Evaluation (SemEval-2016)}, pages={31--41}, year={2016} } ``` ### Contributions Thanks to [@gchhablani](https://github.com/gchhablani) and [@abhishekkrthakur](https://github.com/abhishekkrthakur) for adding this dataset.
tweet_eval
[ "task_categories:text-classification", "task_ids:intent-classification", "task_ids:multi-class-classification", "task_ids:sentiment-classification", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "size_categories:10K<n<100K", "size_categories:1K<n<10K", "size_categories:n<1K", "source_datasets:extended|other-tweet-datasets", "language:en", "license:unknown", "arxiv:2010.12421", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M", "10K<n<100K", "1K<n<10K", "n<1K"], "source_datasets": ["extended|other-tweet-datasets"], "task_categories": ["text-classification"], "task_ids": ["intent-classification", "multi-class-classification", "sentiment-classification"], "paperswithcode_id": "tweeteval", "pretty_name": "TweetEval", "config_names": ["emoji", "emotion", "hate", "irony", "offensive", "sentiment", "stance_abortion", "stance_atheism", "stance_climate", "stance_feminist", "stance_hillary"], "dataset_info": [{"config_name": "emoji", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "\u2764", "1": "\ud83d\ude0d", "2": "\ud83d\ude02", "3": "\ud83d\udc95", "4": "\ud83d\udd25", "5": "\ud83d\ude0a", "6": "\ud83d\ude0e", "7": "\u2728", "8": "\ud83d\udc99", "9": "\ud83d\ude18", "10": "\ud83d\udcf7", "11": "\ud83c\uddfa\ud83c\uddf8", "12": "\u2600", "13": "\ud83d\udc9c", "14": "\ud83d\ude09", "15": "\ud83d\udcaf", "16": "\ud83d\ude01", "17": "\ud83c\udf84", "18": "\ud83d\udcf8", "19": "\ud83d\ude1c"}}}}], "splits": [{"name": "train", "num_bytes": 3803167, "num_examples": 45000}, {"name": "test", "num_bytes": 4255901, "num_examples": 50000}, {"name": "validation", "num_bytes": 396079, "num_examples": 5000}], "download_size": 5939308, "dataset_size": 8455147}, {"config_name": "emotion", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "anger", "1": "joy", "2": "optimism", "3": "sadness"}}}}], "splits": [{"name": "train", "num_bytes": 338871, "num_examples": 3257}, {"name": "test", "num_bytes": 146645, "num_examples": 1421}, {"name": "validation", "num_bytes": 38273, "num_examples": 374}], "download_size": 367016, "dataset_size": 523789}, {"config_name": "hate", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "non-hate", "1": "hate"}}}}], "splits": [{"name": "train", "num_bytes": 1223650, "num_examples": 9000}, {"name": "test", "num_bytes": 428934, "num_examples": 2970}, {"name": "validation", "num_bytes": 154144, "num_examples": 1000}], "download_size": 1196346, "dataset_size": 1806728}, {"config_name": "irony", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "non_irony", "1": "irony"}}}}], "splits": [{"name": "train", "num_bytes": 259187, "num_examples": 2862}, {"name": "test", "num_bytes": 75897, "num_examples": 784}, {"name": "validation", "num_bytes": 86017, "num_examples": 955}], "download_size": 297647, "dataset_size": 421101}, {"config_name": "offensive", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "non-offensive", "1": "offensive"}}}}], "splits": [{"name": "train", "num_bytes": 1648061, "num_examples": 11916}, {"name": "test", "num_bytes": 135473, "num_examples": 860}, {"name": "validation", "num_bytes": 192417, "num_examples": 1324}], "download_size": 1234528, "dataset_size": 1975951}, {"config_name": "sentiment", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "negative", "1": "neutral", "2": "positive"}}}}], "splits": [{"name": "train", "num_bytes": 5425122, "num_examples": 45615}, {"name": "test", "num_bytes": 1279540, "num_examples": 12284}, {"name": "validation", "num_bytes": 239084, "num_examples": 2000}], "download_size": 4849675, "dataset_size": 6943746}, {"config_name": "stance_abortion", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "none", "1": "against", "2": "favor"}}}}], "splits": [{"name": "train", "num_bytes": 68694, "num_examples": 587}, {"name": "test", "num_bytes": 33171, "num_examples": 280}, {"name": "validation", "num_bytes": 7657, "num_examples": 66}], "download_size": 73517, "dataset_size": 109522}, {"config_name": "stance_atheism", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "none", "1": "against", "2": "favor"}}}}], "splits": [{"name": "train", "num_bytes": 54775, "num_examples": 461}, {"name": "test", "num_bytes": 25716, "num_examples": 220}, {"name": "validation", "num_bytes": 6320, "num_examples": 52}], "download_size": 62265, "dataset_size": 86811}, {"config_name": "stance_climate", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "none", "1": "against", "2": "favor"}}}}], "splits": [{"name": "train", "num_bytes": 40249, "num_examples": 355}, {"name": "test", "num_bytes": 19925, "num_examples": 169}, {"name": "validation", "num_bytes": 4801, "num_examples": 40}], "download_size": 48493, "dataset_size": 64975}, {"config_name": "stance_feminist", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "none", "1": "against", "2": "favor"}}}}], "splits": [{"name": "train", "num_bytes": 70509, "num_examples": 597}, {"name": "test", "num_bytes": 33305, "num_examples": 285}, {"name": "validation", "num_bytes": 8035, "num_examples": 67}], "download_size": 76345, "dataset_size": 111849}, {"config_name": "stance_hillary", "features": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "none", "1": "against", "2": "favor"}}}}], "splits": [{"name": "train", "num_bytes": 69596, "num_examples": 620}, {"name": "test", "num_bytes": 34487, "num_examples": 295}, {"name": "validation", "num_bytes": 7532, "num_examples": 69}], "download_size": 74057, "dataset_size": 111615}], "configs": [{"config_name": "emoji", "data_files": [{"split": "train", "path": "emoji/train-*"}, {"split": "test", "path": "emoji/test-*"}, {"split": "validation", "path": "emoji/validation-*"}]}, {"config_name": "emotion", "data_files": [{"split": "train", "path": "emotion/train-*"}, {"split": "test", "path": "emotion/test-*"}, {"split": "validation", "path": "emotion/validation-*"}]}, {"config_name": "hate", "data_files": [{"split": "train", "path": "hate/train-*"}, {"split": "test", "path": "hate/test-*"}, {"split": "validation", "path": "hate/validation-*"}]}, {"config_name": "irony", "data_files": [{"split": "train", "path": "irony/train-*"}, {"split": "test", "path": "irony/test-*"}, {"split": "validation", "path": "irony/validation-*"}]}, {"config_name": "offensive", "data_files": [{"split": "train", "path": "offensive/train-*"}, {"split": "test", "path": "offensive/test-*"}, {"split": "validation", "path": "offensive/validation-*"}]}, {"config_name": "sentiment", "data_files": [{"split": "train", "path": "sentiment/train-*"}, {"split": "test", "path": "sentiment/test-*"}, {"split": "validation", "path": "sentiment/validation-*"}]}, {"config_name": "stance_abortion", "data_files": [{"split": "train", "path": "stance_abortion/train-*"}, {"split": "test", "path": "stance_abortion/test-*"}, {"split": "validation", "path": "stance_abortion/validation-*"}]}, {"config_name": "stance_atheism", "data_files": [{"split": "train", "path": "stance_atheism/train-*"}, {"split": "test", "path": "stance_atheism/test-*"}, {"split": "validation", "path": "stance_atheism/validation-*"}]}, {"config_name": "stance_climate", "data_files": [{"split": "train", "path": "stance_climate/train-*"}, {"split": "test", "path": "stance_climate/test-*"}, {"split": "validation", "path": "stance_climate/validation-*"}]}, {"config_name": "stance_feminist", "data_files": [{"split": "train", "path": "stance_feminist/train-*"}, {"split": "test", "path": "stance_feminist/test-*"}, {"split": "validation", "path": "stance_feminist/validation-*"}]}, {"config_name": "stance_hillary", "data_files": [{"split": "train", "path": "stance_hillary/train-*"}, {"split": "test", "path": "stance_hillary/test-*"}, {"split": "validation", "path": "stance_hillary/validation-*"}]}], "train-eval-index": [{"config": "emotion", "task": "text-classification", "task_id": "multi_class_classification", "splits": {"train_split": "train", "eval_split": "test"}, "col_mapping": {"text": "text", "label": "target"}, "metrics": [{"type": "accuracy", "name": "Accuracy"}, {"type": "f1", "name": "F1 macro", "args": {"average": "macro"}}, {"type": "f1", "name": "F1 micro", "args": {"average": "micro"}}, {"type": "f1", "name": "F1 weighted", "args": {"average": "weighted"}}, {"type": "precision", "name": "Precision macro", "args": {"average": "macro"}}, {"type": "precision", "name": "Precision micro", "args": {"average": "micro"}}, {"type": "precision", "name": "Precision weighted", "args": {"average": "weighted"}}, {"type": "recall", "name": "Recall macro", "args": {"average": "macro"}}, {"type": "recall", "name": "Recall micro", "args": {"average": "micro"}}, {"type": "recall", "name": "Recall weighted", "args": {"average": "weighted"}}]}, {"config": "hate", "task": "text-classification", "task_id": "binary_classification", "splits": {"train_split": "train", "eval_split": "test"}, "col_mapping": {"text": "text", "label": "target"}, "metrics": [{"type": "accuracy", "name": "Accuracy"}, {"type": "f1", "name": "F1 binary", "args": {"average": "binary"}}, {"type": "precision", "name": "Precision macro", "args": {"average": "macro"}}, {"type": "precision", "name": "Precision micro", "args": {"average": "micro"}}, {"type": "precision", "name": "Precision weighted", "args": {"average": "weighted"}}, {"type": "recall", "name": "Recall macro", "args": {"average": "macro"}}, {"type": "recall", "name": "Recall micro", "args": {"average": "micro"}}, {"type": "recall", "name": "Recall weighted", "args": {"average": "weighted"}}]}, {"config": "irony", "task": "text-classification", "task_id": "binary_classification", "splits": {"train_split": "train", "eval_split": "test"}, "col_mapping": {"text": "text", "label": "target"}, "metrics": [{"type": "accuracy", "name": "Accuracy"}, {"type": "f1", "name": "F1 binary", "args": {"average": "binary"}}, {"type": "precision", "name": "Precision macro", "args": {"average": "macro"}}, {"type": "precision", "name": "Precision micro", "args": {"average": "micro"}}, {"type": "precision", "name": "Precision weighted", "args": {"average": "weighted"}}, {"type": "recall", "name": "Recall macro", "args": {"average": "macro"}}, {"type": "recall", "name": "Recall micro", "args": {"average": "micro"}}, {"type": "recall", "name": "Recall weighted", "args": {"average": "weighted"}}]}, {"config": "offensive", "task": "text-classification", "task_id": "binary_classification", "splits": {"train_split": "train", "eval_split": "test"}, "col_mapping": {"text": "text", "label": "target"}, "metrics": [{"type": "accuracy", "name": "Accuracy"}, {"type": "f1", "name": "F1 binary", "args": {"average": "binary"}}, {"type": "precision", "name": "Precision macro", "args": {"average": "macro"}}, {"type": "precision", "name": "Precision micro", "args": {"average": "micro"}}, {"type": "precision", "name": "Precision weighted", "args": {"average": "weighted"}}, {"type": "recall", "name": "Recall macro", "args": {"average": "macro"}}, {"type": "recall", "name": "Recall micro", "args": {"average": "micro"}}, {"type": "recall", "name": "Recall weighted", "args": {"average": "weighted"}}]}, {"config": "sentiment", "task": "text-classification", "task_id": "multi_class_classification", "splits": {"train_split": "train", "eval_split": "test"}, "col_mapping": {"text": "text", "label": "target"}, "metrics": [{"type": "accuracy", "name": "Accuracy"}, {"type": "f1", "name": "F1 macro", "args": {"average": "macro"}}, {"type": "f1", "name": "F1 micro", "args": {"average": "micro"}}, {"type": "f1", "name": "F1 weighted", "args": {"average": "weighted"}}, {"type": "precision", "name": "Precision macro", "args": {"average": "macro"}}, {"type": "precision", "name": "Precision micro", "args": {"average": "micro"}}, {"type": "precision", "name": "Precision weighted", "args": {"average": "weighted"}}, {"type": "recall", "name": "Recall macro", "args": {"average": "macro"}}, {"type": "recall", "name": "Recall micro", "args": {"average": "micro"}}, {"type": "recall", "name": "Recall weighted", "args": {"average": "weighted"}}]}]}
2024-01-04T16:40:33+00:00
[ "2010.12421" ]
[ "en" ]
TAGS #task_categories-text-classification #task_ids-intent-classification #task_ids-multi-class-classification #task_ids-sentiment-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #size_categories-10K<n<100K #size_categories-1K<n<10K #size_categories-n<1K #source_datasets-extended|other-tweet-datasets #language-English #license-unknown #arxiv-2010.12421 #region-us
Dataset Card for tweet\_eval ============================ Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: * Repository: GitHub * Paper: EMNLP Paper * Leaderboard: GitHub Leaderboard * Point of Contact: ### Dataset Summary TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. The tasks include - irony, hate, offensive, stance, emoji, emotion, and sentiment. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits. ### Supported Tasks and Leaderboards * 'text\_classification': The dataset can be trained using a SentenceClassification model from HuggingFace transformers. ### Languages The text in the dataset is in English, as spoken by Twitter users. Dataset Structure ----------------- ### Data Instances An instance from 'emoji' config: An instance from 'emotion' config: An instance from 'hate' config: An instance from 'irony' config: An instance from 'offensive' config: An instance from 'sentiment' config: An instance from 'stance\_abortion' config: An instance from 'stance\_atheism' config: An instance from 'stance\_climate' config: An instance from 'stance\_feminist' config: An instance from 'stance\_hillary' config: ### Data Fields For 'emoji' config: * 'text': a 'string' feature containing the tweet. * 'label': an 'int' classification label with the following mapping: '0': '1': '2': '3': '4': '5': '6': '7': '8': '9': '10': '11': 🇺🇸 '12': '13': '14': '15': '16': '17': '18': '19': For 'emotion' config: * 'text': a 'string' feature containing the tweet. * 'label': an 'int' classification label with the following mapping: '0': anger '1': joy '2': optimism '3': sadness For 'hate' config: * 'text': a 'string' feature containing the tweet. * 'label': an 'int' classification label with the following mapping: '0': non-hate '1': hate For 'irony' config: * 'text': a 'string' feature containing the tweet. * 'label': an 'int' classification label with the following mapping: '0': non\_irony '1': irony For 'offensive' config: * 'text': a 'string' feature containing the tweet. * 'label': an 'int' classification label with the following mapping: '0': non-offensive '1': offensive For 'sentiment' config: * 'text': a 'string' feature containing the tweet. * 'label': an 'int' classification label with the following mapping: '0': negative '1': neutral '2': positive For 'stance\_abortion' config: * 'text': a 'string' feature containing the tweet. * 'label': an 'int' classification label with the following mapping: '0': none '1': against '2': favor For 'stance\_atheism' config: * 'text': a 'string' feature containing the tweet. * 'label': an 'int' classification label with the following mapping: '0': none '1': against '2': favor For 'stance\_climate' config: * 'text': a 'string' feature containing the tweet. * 'label': an 'int' classification label with the following mapping: '0': none '1': against '2': favor For 'stance\_feminist' config: * 'text': a 'string' feature containing the tweet. * 'label': an 'int' classification label with the following mapping: '0': none '1': against '2': favor For 'stance\_hillary' config: * 'text': a 'string' feature containing the tweet. * 'label': an 'int' classification label with the following mapping: '0': none '1': against '2': favor ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators Francesco Barbieri, Jose Camacho-Collados, Luis Espiinosa-Anke and Leonardo Neves through Cardiff NLP. ### Licensing Information This is not a single dataset, therefore each subset has its own license (the collection itself does not have additional restrictions). All of the datasets require complying with Twitter Terms Of Service and Twitter API Terms Of Service Additionally the license are: * emoji: Undefined * emotion(EmoInt): Undefined * hate (HateEval): Need permission here * irony: Undefined * Offensive: Undefined * Sentiment: Creative Commons Attribution 3.0 Unported License * Stance: Undefined If you use any of the TweetEval datasets, please cite their original publications: #### Emotion Recognition: #### Emoji Prediction: #### Irony Detection: #### Hate Speech Detection: #### Offensive Language Identification: #### Sentiment Analysis: #### Stance Detection: ### Contributions Thanks to @gchhablani and @abhishekkrthakur for adding this dataset.
[ "### Dataset Summary\n\n\nTweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. The tasks include - irony, hate, offensive, stance, emoji, emotion, and sentiment. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.", "### Supported Tasks and Leaderboards\n\n\n* 'text\\_classification': The dataset can be trained using a SentenceClassification model from HuggingFace transformers.", "### Languages\n\n\nThe text in the dataset is in English, as spoken by Twitter users.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nAn instance from 'emoji' config:\n\n\nAn instance from 'emotion' config:\n\n\nAn instance from 'hate' config:\n\n\nAn instance from 'irony' config:\n\n\nAn instance from 'offensive' config:\n\n\nAn instance from 'sentiment' config:\n\n\nAn instance from 'stance\\_abortion' config:\n\n\nAn instance from 'stance\\_atheism' config:\n\n\nAn instance from 'stance\\_climate' config:\n\n\nAn instance from 'stance\\_feminist' config:\n\n\nAn instance from 'stance\\_hillary' config:", "### Data Fields\n\n\nFor 'emoji' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0':\n\n\n'1':\n\n\n'2':\n\n\n'3':\n\n\n'4':\n\n\n'5':\n\n\n'6':\n\n\n'7':\n\n\n'8':\n\n\n'9':\n\n\n'10':\n\n\n'11': 🇺🇸\n\n\n'12':\n\n\n'13':\n\n\n'14':\n\n\n'15':\n\n\n'16':\n\n\n'17':\n\n\n'18':\n\n\n'19':\n\n\nFor 'emotion' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': anger\n\n\n'1': joy\n\n\n'2': optimism\n\n\n'3': sadness\n\n\nFor 'hate' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': non-hate\n\n\n'1': hate\n\n\nFor 'irony' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': non\\_irony\n\n\n'1': irony\n\n\nFor 'offensive' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': non-offensive\n\n\n'1': offensive\n\n\nFor 'sentiment' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': negative\n\n\n'1': neutral\n\n\n'2': positive\n\n\nFor 'stance\\_abortion' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': none\n\n\n'1': against\n\n\n'2': favor\n\n\nFor 'stance\\_atheism' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': none\n\n\n'1': against\n\n\n'2': favor\n\n\nFor 'stance\\_climate' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': none\n\n\n'1': against\n\n\n'2': favor\n\n\nFor 'stance\\_feminist' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': none\n\n\n'1': against\n\n\n'2': favor\n\n\nFor 'stance\\_hillary' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': none\n\n\n'1': against\n\n\n'2': favor", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nFrancesco Barbieri, Jose Camacho-Collados, Luis Espiinosa-Anke and Leonardo Neves through Cardiff NLP.", "### Licensing Information\n\n\nThis is not a single dataset, therefore each subset has its own license (the collection itself does not have additional restrictions).\n\n\nAll of the datasets require complying with Twitter Terms Of Service and Twitter API Terms Of Service\n\n\nAdditionally the license are:\n\n\n* emoji: Undefined\n* emotion(EmoInt): Undefined\n* hate (HateEval): Need permission here\n* irony: Undefined\n* Offensive: Undefined\n* Sentiment: Creative Commons Attribution 3.0 Unported License\n* Stance: Undefined\n\n\nIf you use any of the TweetEval datasets, please cite their original publications:", "#### Emotion Recognition:", "#### Emoji Prediction:", "#### Irony Detection:", "#### Hate Speech Detection:", "#### Offensive Language Identification:", "#### Sentiment Analysis:", "#### Stance Detection:", "### Contributions\n\n\nThanks to @gchhablani and @abhishekkrthakur for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #task_ids-intent-classification #task_ids-multi-class-classification #task_ids-sentiment-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #size_categories-10K<n<100K #size_categories-1K<n<10K #size_categories-n<1K #source_datasets-extended|other-tweet-datasets #language-English #license-unknown #arxiv-2010.12421 #region-us \n", "### Dataset Summary\n\n\nTweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. The tasks include - irony, hate, offensive, stance, emoji, emotion, and sentiment. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.", "### Supported Tasks and Leaderboards\n\n\n* 'text\\_classification': The dataset can be trained using a SentenceClassification model from HuggingFace transformers.", "### Languages\n\n\nThe text in the dataset is in English, as spoken by Twitter users.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nAn instance from 'emoji' config:\n\n\nAn instance from 'emotion' config:\n\n\nAn instance from 'hate' config:\n\n\nAn instance from 'irony' config:\n\n\nAn instance from 'offensive' config:\n\n\nAn instance from 'sentiment' config:\n\n\nAn instance from 'stance\\_abortion' config:\n\n\nAn instance from 'stance\\_atheism' config:\n\n\nAn instance from 'stance\\_climate' config:\n\n\nAn instance from 'stance\\_feminist' config:\n\n\nAn instance from 'stance\\_hillary' config:", "### Data Fields\n\n\nFor 'emoji' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0':\n\n\n'1':\n\n\n'2':\n\n\n'3':\n\n\n'4':\n\n\n'5':\n\n\n'6':\n\n\n'7':\n\n\n'8':\n\n\n'9':\n\n\n'10':\n\n\n'11': 🇺🇸\n\n\n'12':\n\n\n'13':\n\n\n'14':\n\n\n'15':\n\n\n'16':\n\n\n'17':\n\n\n'18':\n\n\n'19':\n\n\nFor 'emotion' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': anger\n\n\n'1': joy\n\n\n'2': optimism\n\n\n'3': sadness\n\n\nFor 'hate' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': non-hate\n\n\n'1': hate\n\n\nFor 'irony' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': non\\_irony\n\n\n'1': irony\n\n\nFor 'offensive' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': non-offensive\n\n\n'1': offensive\n\n\nFor 'sentiment' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': negative\n\n\n'1': neutral\n\n\n'2': positive\n\n\nFor 'stance\\_abortion' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': none\n\n\n'1': against\n\n\n'2': favor\n\n\nFor 'stance\\_atheism' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': none\n\n\n'1': against\n\n\n'2': favor\n\n\nFor 'stance\\_climate' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': none\n\n\n'1': against\n\n\n'2': favor\n\n\nFor 'stance\\_feminist' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': none\n\n\n'1': against\n\n\n'2': favor\n\n\nFor 'stance\\_hillary' config:\n\n\n* 'text': a 'string' feature containing the tweet.\n* 'label': an 'int' classification label with the following mapping:\n\n\n'0': none\n\n\n'1': against\n\n\n'2': favor", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nFrancesco Barbieri, Jose Camacho-Collados, Luis Espiinosa-Anke and Leonardo Neves through Cardiff NLP.", "### Licensing Information\n\n\nThis is not a single dataset, therefore each subset has its own license (the collection itself does not have additional restrictions).\n\n\nAll of the datasets require complying with Twitter Terms Of Service and Twitter API Terms Of Service\n\n\nAdditionally the license are:\n\n\n* emoji: Undefined\n* emotion(EmoInt): Undefined\n* hate (HateEval): Need permission here\n* irony: Undefined\n* Offensive: Undefined\n* Sentiment: Creative Commons Attribution 3.0 Unported License\n* Stance: Undefined\n\n\nIf you use any of the TweetEval datasets, please cite their original publications:", "#### Emotion Recognition:", "#### Emoji Prediction:", "#### Irony Detection:", "#### Hate Speech Detection:", "#### Offensive Language Identification:", "#### Sentiment Analysis:", "#### Stance Detection:", "### Contributions\n\n\nThanks to @gchhablani and @abhishekkrthakur for adding this dataset." ]
[ 160, 90, 41, 28, 135, 717, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 36, 143, 8, 8, 7, 8, 9, 7, 7, 26 ]
[ "passage: TAGS\n#task_categories-text-classification #task_ids-intent-classification #task_ids-multi-class-classification #task_ids-sentiment-classification #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #size_categories-10K<n<100K #size_categories-1K<n<10K #size_categories-n<1K #source_datasets-extended|other-tweet-datasets #language-English #license-unknown #arxiv-2010.12421 #region-us \n### Dataset Summary\n\n\nTweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. The tasks include - irony, hate, offensive, stance, emoji, emotion, and sentiment. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.### Supported Tasks and Leaderboards\n\n\n* 'text\\_classification': The dataset can be trained using a SentenceClassification model from HuggingFace transformers.### Languages\n\n\nThe text in the dataset is in English, as spoken by Twitter users.\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nAn instance from 'emoji' config:\n\n\nAn instance from 'emotion' config:\n\n\nAn instance from 'hate' config:\n\n\nAn instance from 'irony' config:\n\n\nAn instance from 'offensive' config:\n\n\nAn instance from 'sentiment' config:\n\n\nAn instance from 'stance\\_abortion' config:\n\n\nAn instance from 'stance\\_atheism' config:\n\n\nAn instance from 'stance\\_climate' config:\n\n\nAn instance from 'stance\\_feminist' config:\n\n\nAn instance from 'stance\\_hillary' config:" ]
4271bbdba0deb6570abdfe9766785cfd305876de
# Dataset Card for TweetQA ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [TweetQA homepage](https://tweetqa.github.io/) - **Repository:** - **Paper:** [TWEETQA: A Social Media Focused Question Answering Dataset](https://arxiv.org/abs/1907.06292) - **Leaderboard:** [TweetQA Leaderboard](https://tweetqa.github.io/) - **Point of Contact:** [Wenhan Xiong](xwhan@cs.ucsb.edu) ### Dataset Summary With social media becoming increasingly popular on which lots of news and real-time events are reported, developing automated question answering systems is critical to the effectiveness of many applications that rely on real-time knowledge. While previous question answering (QA) datasets have concentrated on formal text like news and Wikipedia, the first large-scale dataset for QA over social media data is presented. To make sure the tweets are meaningful and contain interesting information, tweets used by journalists to write news articles are gathered. Then human annotators are asked to write questions and answers upon these tweets. Unlike other QA datasets like SQuAD in which the answers are extractive, the answer are allowed to be abstractive. The task requires model to read a short tweet and a question and outputs a text phrase (does not need to be in the tweet) as the answer. ### Supported Tasks and Leaderboards - `question-answering`: The dataset can be used to train a model for Open-Domain Question Answering where the task is to answer the given questions for a tweet. The performance is measured by comparing the model answers to the the annoted groundtruth and calculating the BLEU-1/Meteor/ROUGE-L score. This task has an active leaderboard which can be found [here](https://tweetqa.github.io/) and ranks models based on [BLEU-1](https://huggingface.co/metrics/blue), [Meteor](https://huggingface.co/metrics/meteor) and [ROUGLE-L](https://huggingface.co/metrics/rouge). ### Languages English. ## Dataset Structure ### Data Instances Sample data: ``` { "Question": "who is the tallest host?", "Answer": ["sam bee","sam bee"], "Tweet": "Don't believe @ConanOBrien's height lies. Sam Bee is the tallest host in late night. #alternativefacts\u2014 Full Frontal (@FullFrontalSamB) January 22, 2017", "qid": "3554ee17d86b678be34c4dc2c04e334f" } ``` The test split doesn't include answers so the Answer field is an empty list. ### Data Fields - `Question`: a question based on information from a tweet - `Answer`: list of possible answers from the tweet - `Tweet`: source tweet - `qid`: question id ### Data Splits The dataset is split in train, validation and test set. The train set cointains 10692 examples, the validation set 1086 and the test set 1979 examples. ## Dataset Creation ### Curation Rationale With social media becoming increasingly popular on which lots of news and real-time events are reported, developing automated question answering systems is critical to the effectiveness of many applications that rely on real-time knowledge. While previous question answering (QA) datasets have concentrated on formal text like news and Wikipedia, the first large-scale dataset for QA over social media data is presented. To make sure the tweets are meaningful and contain interesting information, tweets used by journalists to write news articles are gathered. Then human annotators are asked to write questions and answers upon these tweets. Unlike other QA datasets like SQuAD in which the answers are extractive, the answer are allowed to be abstractive. The task requires model to read a short tweet and a question and outputs a text phrase (does not need to be in the tweet) as the answer. ### Source Data #### Initial Data Collection and Normalization The authors look into the the archived snapshots of two major news websites (CNN, NBC), and then extract the tweet blocks that are embedded in the news articles. In order to get enough data, they first extract the URLs of all section pages (e.g. World, Politics, Money, Tech) from the snapshot of each home page and then crawl all articles with tweets from these section pages. Then, they filter out the tweets that heavily rely on attached media to convey information, for which they utilize a state-of-the-art semantic role labeling model trained on CoNLL-2005 (He et al., 2017) to analyze the predicate-argument structure of the tweets collected from news articles and keep only the tweets with more than two labeled arguments. This filtering process also automatically filters out most of the short tweets. For the tweets collected from CNN, 22.8% of them were filtered via semantic role labeling. For tweets from NBC, 24.1% of the tweets were filtered. #### Who are the source language producers? Twitter users. ### Annotations #### Annotation process The Amazon Mechanical Turk workers were used to collect question-answer pairs for the filtered tweets. For each Human Intelligence Task (HIT), the authors ask the worker to read three tweets and write two question-answer pairs for each tweet. To ensure the quality, they require the workers to be located in major English speaking countries (i.e. Canada, US, and UK) and have an acceptance rate larger than 95%. Since the authors use tweets as context, lots of important information are contained in hashtags or even emojis. Instead of only showing the text to the workers, they use javascript to directly embed the whole tweet into each HIT. This gives workers the same experience as reading tweets via web browsers and help them to better compose questions. To avoid trivial questions that can be simply answered by superficial text matching methods or too challenging questions that require background knowledge, the authors explicitly state the following items in the HIT instructions for question writing: - No Yes-no questions should be asked. - The question should have at least five words. - Videos, images or inserted links should not be considered. - No background knowledge should be required to answer the question. To help the workers better follow the instructions, they also include a representative example showing both good and bad questions or answers in the instructions. As for the answers, since the context they consider is relatively shorter than the context of previous datasets, they do not restrict the answers to be in the tweet, otherwise, the task may potentially be simplified as a classification problem. The workers are allowed to write their answers in their own words, but the authors require the answers to be brief and can be directly inferred from the tweets. After they retrieve the QA pairs from all HITs, they conduct further post-filtering to filter out the pairs from workers that obviously do not follow instructions. They remove QA pairs with yes/no answers. Questions with less than five words are also filtered out. This process filtered 13% of the QA pairs. The dataset now includes 10,898 articles, 17,794 tweets, and 13,757 crowdsourced question-answer pairs. All QA pairs were written by 492 individual workers. #### Who are the annotators? Amazon Mechanical Turk workers. ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases From the paper: > It is also worth noting that the data collected from social media can not only capture events and developments in real-time but also capture individual opinions and thus requires reasoning related to the authorship of the content as is illustrated in Table 1. > Specifically, a significant amount of questions require certain reasoning skills that are specific to social media data: - Understanding authorship: Since tweets are highly personal, it is critical to understand how questions/tweets related to the authors. - Oral English & Tweet English: Tweets are often oral and informal. QA over tweets requires the understanding of common oral English. Our TWEETQA also requires understanding some tweet-specific English, like conversation-style English. - Understanding of user IDs & hashtags: Tweets often contains user IDs and hashtags, which are single special tokens. Understanding these special tokens is important to answer person- or event-related questions. ### Other Known Limitations [More Information Needed] ## Additional Information The annotated answers are validated by the authors as follows: For the purposes of human performance evaluation and inter-annotator agreement checking, the authors launch a different set of HITs to ask workers to answer questions in the test and development set. The workers are shown with the tweet blocks as well as the questions collected in the previous step. At this step, workers are allowed to label the questions as “NA” if they think the questions are not answerable. They find that 3.1% of the questions are labeled as unanswerable by the workers (for SQuAD, the ratio is 2.6%). Since the answers collected at this step and previous step are written by different workers, the answers can be written in different text forms even they are semantically equal to each other. For example, one answer can be “Hillary Clinton” while the other is “@HillaryClinton”. As it is not straightforward to automatically calculate the overall agreement, they manually check the agreement on a subset of 200 random samples from the development set and ask an independent human moderator to verify the result. It turns out that 90% of the answers pairs are semantically equivalent, 2% of them are partially equivalent (one of them is incomplete) and 8% are totally inconsistent. The answers collected at this step are also used to measure the human performance. 59 individual workers participated in this process. ### Dataset Curators Xiong, Wenhan and Wu, Jiawei and Wang, Hong and Kulkarni, Vivek and Yu, Mo and Guo, Xiaoxiao and Chang, Shiyu and Wang, William Yang. ### Licensing Information CC BY-SA 4.0. ### Citation Information ``` @inproceedings{xiong2019tweetqa, title={TweetQA: A Social Media Focused Question Answering Dataset}, author={Xiong, Wenhan and Wu, Jiawei and Wang, Hong and Kulkarni, Vivek and Yu, Mo and Guo, Xiaoxiao and Chang, Shiyu and Wang, William Yang}, booktitle={Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics}, year={2019} } ``` ### Contributions Thanks to [@anaerobeth](https://github.com/anaerobeth) for adding this dataset.
tweet_qa
[ "task_categories:question-answering", "task_ids:open-domain-qa", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:cc-by-sa-4.0", "arxiv:1907.06292", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["cc-by-sa-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["question-answering"], "task_ids": ["open-domain-qa"], "paperswithcode_id": "tweetqa", "pretty_name": "TweetQA", "dataset_info": {"features": [{"name": "Question", "dtype": "string"}, {"name": "Answer", "sequence": "string"}, {"name": "Tweet", "dtype": "string"}, {"name": "qid", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 2769996, "num_examples": 10692}, {"name": "validation", "num_bytes": 295415, "num_examples": 1086}, {"name": "test", "num_bytes": 473710, "num_examples": 1979}], "download_size": 2434334, "dataset_size": 3539121}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "validation", "path": "data/validation-*"}, {"split": "test", "path": "data/test-*"}]}]}
2024-01-24T08:53:20+00:00
[ "1907.06292" ]
[ "en" ]
TAGS #task_categories-question-answering #task_ids-open-domain-qa #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-sa-4.0 #arxiv-1907.06292 #region-us
# Dataset Card for TweetQA ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: TweetQA homepage - Repository: - Paper: TWEETQA: A Social Media Focused Question Answering Dataset - Leaderboard: TweetQA Leaderboard - Point of Contact: Wenhan Xiong ### Dataset Summary With social media becoming increasingly popular on which lots of news and real-time events are reported, developing automated question answering systems is critical to the effectiveness of many applications that rely on real-time knowledge. While previous question answering (QA) datasets have concentrated on formal text like news and Wikipedia, the first large-scale dataset for QA over social media data is presented. To make sure the tweets are meaningful and contain interesting information, tweets used by journalists to write news articles are gathered. Then human annotators are asked to write questions and answers upon these tweets. Unlike other QA datasets like SQuAD in which the answers are extractive, the answer are allowed to be abstractive. The task requires model to read a short tweet and a question and outputs a text phrase (does not need to be in the tweet) as the answer. ### Supported Tasks and Leaderboards - 'question-answering': The dataset can be used to train a model for Open-Domain Question Answering where the task is to answer the given questions for a tweet. The performance is measured by comparing the model answers to the the annoted groundtruth and calculating the BLEU-1/Meteor/ROUGE-L score. This task has an active leaderboard which can be found here and ranks models based on BLEU-1, Meteor and ROUGLE-L. ### Languages English. ## Dataset Structure ### Data Instances Sample data: The test split doesn't include answers so the Answer field is an empty list. ### Data Fields - 'Question': a question based on information from a tweet - 'Answer': list of possible answers from the tweet - 'Tweet': source tweet - 'qid': question id ### Data Splits The dataset is split in train, validation and test set. The train set cointains 10692 examples, the validation set 1086 and the test set 1979 examples. ## Dataset Creation ### Curation Rationale With social media becoming increasingly popular on which lots of news and real-time events are reported, developing automated question answering systems is critical to the effectiveness of many applications that rely on real-time knowledge. While previous question answering (QA) datasets have concentrated on formal text like news and Wikipedia, the first large-scale dataset for QA over social media data is presented. To make sure the tweets are meaningful and contain interesting information, tweets used by journalists to write news articles are gathered. Then human annotators are asked to write questions and answers upon these tweets. Unlike other QA datasets like SQuAD in which the answers are extractive, the answer are allowed to be abstractive. The task requires model to read a short tweet and a question and outputs a text phrase (does not need to be in the tweet) as the answer. ### Source Data #### Initial Data Collection and Normalization The authors look into the the archived snapshots of two major news websites (CNN, NBC), and then extract the tweet blocks that are embedded in the news articles. In order to get enough data, they first extract the URLs of all section pages (e.g. World, Politics, Money, Tech) from the snapshot of each home page and then crawl all articles with tweets from these section pages. Then, they filter out the tweets that heavily rely on attached media to convey information, for which they utilize a state-of-the-art semantic role labeling model trained on CoNLL-2005 (He et al., 2017) to analyze the predicate-argument structure of the tweets collected from news articles and keep only the tweets with more than two labeled arguments. This filtering process also automatically filters out most of the short tweets. For the tweets collected from CNN, 22.8% of them were filtered via semantic role labeling. For tweets from NBC, 24.1% of the tweets were filtered. #### Who are the source language producers? Twitter users. ### Annotations #### Annotation process The Amazon Mechanical Turk workers were used to collect question-answer pairs for the filtered tweets. For each Human Intelligence Task (HIT), the authors ask the worker to read three tweets and write two question-answer pairs for each tweet. To ensure the quality, they require the workers to be located in major English speaking countries (i.e. Canada, US, and UK) and have an acceptance rate larger than 95%. Since the authors use tweets as context, lots of important information are contained in hashtags or even emojis. Instead of only showing the text to the workers, they use javascript to directly embed the whole tweet into each HIT. This gives workers the same experience as reading tweets via web browsers and help them to better compose questions. To avoid trivial questions that can be simply answered by superficial text matching methods or too challenging questions that require background knowledge, the authors explicitly state the following items in the HIT instructions for question writing: - No Yes-no questions should be asked. - The question should have at least five words. - Videos, images or inserted links should not be considered. - No background knowledge should be required to answer the question. To help the workers better follow the instructions, they also include a representative example showing both good and bad questions or answers in the instructions. As for the answers, since the context they consider is relatively shorter than the context of previous datasets, they do not restrict the answers to be in the tweet, otherwise, the task may potentially be simplified as a classification problem. The workers are allowed to write their answers in their own words, but the authors require the answers to be brief and can be directly inferred from the tweets. After they retrieve the QA pairs from all HITs, they conduct further post-filtering to filter out the pairs from workers that obviously do not follow instructions. They remove QA pairs with yes/no answers. Questions with less than five words are also filtered out. This process filtered 13% of the QA pairs. The dataset now includes 10,898 articles, 17,794 tweets, and 13,757 crowdsourced question-answer pairs. All QA pairs were written by 492 individual workers. #### Who are the annotators? Amazon Mechanical Turk workers. ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases From the paper: > It is also worth noting that the data collected from social media can not only capture events and developments in real-time but also capture individual opinions and thus requires reasoning related to the authorship of the content as is illustrated in Table 1. > Specifically, a significant amount of questions require certain reasoning skills that are specific to social media data: - Understanding authorship: Since tweets are highly personal, it is critical to understand how questions/tweets related to the authors. - Oral English & Tweet English: Tweets are often oral and informal. QA over tweets requires the understanding of common oral English. Our TWEETQA also requires understanding some tweet-specific English, like conversation-style English. - Understanding of user IDs & hashtags: Tweets often contains user IDs and hashtags, which are single special tokens. Understanding these special tokens is important to answer person- or event-related questions. ### Other Known Limitations ## Additional Information The annotated answers are validated by the authors as follows: For the purposes of human performance evaluation and inter-annotator agreement checking, the authors launch a different set of HITs to ask workers to answer questions in the test and development set. The workers are shown with the tweet blocks as well as the questions collected in the previous step. At this step, workers are allowed to label the questions as “NA” if they think the questions are not answerable. They find that 3.1% of the questions are labeled as unanswerable by the workers (for SQuAD, the ratio is 2.6%). Since the answers collected at this step and previous step are written by different workers, the answers can be written in different text forms even they are semantically equal to each other. For example, one answer can be “Hillary Clinton” while the other is “@HillaryClinton”. As it is not straightforward to automatically calculate the overall agreement, they manually check the agreement on a subset of 200 random samples from the development set and ask an independent human moderator to verify the result. It turns out that 90% of the answers pairs are semantically equivalent, 2% of them are partially equivalent (one of them is incomplete) and 8% are totally inconsistent. The answers collected at this step are also used to measure the human performance. 59 individual workers participated in this process. ### Dataset Curators Xiong, Wenhan and Wu, Jiawei and Wang, Hong and Kulkarni, Vivek and Yu, Mo and Guo, Xiaoxiao and Chang, Shiyu and Wang, William Yang. ### Licensing Information CC BY-SA 4.0. ### Contributions Thanks to @anaerobeth for adding this dataset.
[ "# Dataset Card for TweetQA", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: TweetQA homepage\n- Repository:\n- Paper: TWEETQA: A Social Media Focused Question Answering Dataset\n- Leaderboard: TweetQA Leaderboard\n- Point of Contact: Wenhan Xiong", "### Dataset Summary\n\nWith social media becoming increasingly popular on which lots of news and real-time events are reported, developing automated question answering systems is critical to the effectiveness of many applications that rely on real-time knowledge. While previous question answering (QA) datasets have concentrated on formal text like news and Wikipedia, the first large-scale dataset for QA over social media data is presented. To make sure the tweets are meaningful and contain interesting information, tweets used by journalists to write news articles are gathered. Then human annotators are asked to write questions and answers upon these tweets. Unlike other QA datasets like SQuAD in which the answers are extractive, the answer are allowed to be abstractive. The task requires model to read a short tweet and a question and outputs a text phrase (does not need to be in the tweet) as the answer.", "### Supported Tasks and Leaderboards\n\n- 'question-answering': The dataset can be used to train a model for Open-Domain Question Answering where the task is to answer the given questions for a tweet. The performance is measured by comparing the model answers to the the annoted groundtruth and calculating the BLEU-1/Meteor/ROUGE-L score. This task has an active leaderboard which can be found here and ranks models based on BLEU-1, Meteor and ROUGLE-L.", "### Languages\n\nEnglish.", "## Dataset Structure", "### Data Instances\n\nSample data:\n\n\n\nThe test split doesn't include answers so the Answer field is an empty list.", "### Data Fields\n\n- 'Question': a question based on information from a tweet\n- 'Answer': list of possible answers from the tweet\n- 'Tweet': source tweet\n- 'qid': question id", "### Data Splits\n\nThe dataset is split in train, validation and test set. The train set cointains 10692 examples, the validation set 1086 and the test set 1979 examples.", "## Dataset Creation", "### Curation Rationale\n\nWith social media becoming increasingly popular on which lots of news and real-time events are reported, developing automated question answering systems is critical to the effectiveness of many applications that rely on real-time knowledge. While previous question answering (QA) datasets have concentrated on formal text like news and Wikipedia, the first large-scale dataset for QA over social media data is presented. To make sure the tweets are meaningful and contain interesting information, tweets used by journalists to write news articles are gathered. Then human annotators are asked to write questions and answers upon these tweets. Unlike other QA datasets like SQuAD in which the answers are extractive, the answer are allowed to be abstractive. The task requires model to read a short tweet and a question and outputs a text phrase (does not need to be in the tweet) as the answer.", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe authors look into the the archived snapshots of two major news websites (CNN, NBC), and then extract the tweet blocks that are embedded in the news articles. In order to get enough data, they first extract the URLs of all section pages (e.g. World, Politics, Money, Tech) from the snapshot of each home page and then crawl all articles with tweets from these section pages. Then, they filter out the tweets that heavily rely on attached media to convey information, for which they utilize a state-of-the-art semantic role labeling model trained on CoNLL-2005 (He et al., 2017) to analyze the predicate-argument structure of the tweets collected from news articles and keep\nonly the tweets with more than two labeled arguments. This filtering process also automatically\nfilters out most of the short tweets. For the tweets collected from CNN, 22.8% of them were filtered\nvia semantic role labeling. For tweets from NBC, 24.1% of the tweets were filtered.", "#### Who are the source language producers?\n\nTwitter users.", "### Annotations", "#### Annotation process\n\nThe Amazon Mechanical Turk workers were used to collect question-answer\npairs for the filtered tweets. For each Human Intelligence Task (HIT), the authors ask the worker to read three tweets and write two question-answer pairs for each tweet. To ensure the quality, they require the workers to be located in major English speaking countries (i.e. Canada, US, and UK) and have an acceptance rate larger than 95%. Since the authors use tweets as context, lots of important information are contained in hashtags or even emojis. Instead of only showing the text to the workers, they use javascript to directly embed the whole tweet into each HIT. This gives workers the same experience as reading tweets via web browsers and help them to better compose questions. To avoid trivial questions that can be simply answered by superficial text matching methods or too challenging questions that require background knowledge, the authors explicitly state the following items in the HIT instructions for question writing:\n- No Yes-no questions should be asked.\n- The question should have at least five words.\n- Videos, images or inserted links should not\nbe considered.\n- No background knowledge should be required to answer the question.\nTo help the workers better follow the instructions, they also include a representative example showing both good and bad questions or answers in the instructions. As for the answers, since the context they consider is relatively shorter than the context of previous datasets, they do not restrict the answers to be in the tweet, otherwise, the task may potentially be simplified as a classification problem. The workers are allowed to write their answers in their own words, but the authors require the answers to be brief and can be directly inferred from the tweets. After they retrieve the QA pairs from all HITs, they conduct further post-filtering to filter out the pairs from workers that obviously do not follow instructions. They remove QA pairs with yes/no answers. Questions with less than five words are also filtered out. This process filtered 13% of the QA pairs. The dataset now includes 10,898 articles, 17,794 tweets, and 13,757 crowdsourced question-answer pairs. All QA pairs were written by 492 individual workers.", "#### Who are the annotators?\n\nAmazon Mechanical Turk workers.", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases\n\nFrom the paper:\n> It is also worth noting that the data collected from social media can not only capture events and developments in real-time but also capture individual opinions and thus requires reasoning related to the authorship of the content as is illustrated in Table 1.\n\n> Specifically, a significant amount of questions require certain reasoning skills that are specific to social media data:\n- Understanding authorship: Since tweets are highly personal, it is critical to understand how questions/tweets related to the authors.\n- Oral English & Tweet English: Tweets are often oral and informal. QA over tweets requires the understanding of common oral English. Our TWEETQA also requires understanding some tweet-specific English, like conversation-style English.\n- Understanding of user IDs & hashtags: Tweets often contains user IDs and hashtags, which are single special tokens. Understanding these special tokens is important to answer person- or event-related questions.", "### Other Known Limitations", "## Additional Information\n\nThe annotated answers are validated by the authors as follows:\nFor the purposes of human performance evaluation and inter-annotator agreement checking, the authors launch a different set of HITs to ask workers to answer questions in the test and development set. The workers are shown with the tweet blocks as well as the questions collected in the previous step. At this step, workers are allowed to label the questions as “NA” if they think the questions are not answerable. They find that 3.1% of the questions are labeled as unanswerable by the workers (for SQuAD, the ratio is 2.6%). Since the answers collected at this step and previous step are written by different workers, the answers can be written in different text forms even they are semantically equal to each other. For example, one answer can be “Hillary Clinton” while the other is “@HillaryClinton”. As it is not straightforward to automatically calculate the overall agreement, they manually check the agreement on a subset of 200 random samples from the development set and ask an independent human moderator to verify the result. It turns out that 90% of the answers pairs are semantically equivalent, 2% of them are partially equivalent (one of them is incomplete) and 8% are totally inconsistent. The answers collected at this step are also used to measure the human performance. 59 individual workers participated in this process.", "### Dataset Curators\n\nXiong, Wenhan and Wu, Jiawei and Wang, Hong and Kulkarni, Vivek and Yu, Mo and Guo, Xiaoxiao and Chang, Shiyu and Wang, William Yang.", "### Licensing Information\n\nCC BY-SA 4.0.", "### Contributions\n\nThanks to @anaerobeth for adding this dataset." ]
[ "TAGS\n#task_categories-question-answering #task_ids-open-domain-qa #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-sa-4.0 #arxiv-1907.06292 #region-us \n", "# Dataset Card for TweetQA", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: TweetQA homepage\n- Repository:\n- Paper: TWEETQA: A Social Media Focused Question Answering Dataset\n- Leaderboard: TweetQA Leaderboard\n- Point of Contact: Wenhan Xiong", "### Dataset Summary\n\nWith social media becoming increasingly popular on which lots of news and real-time events are reported, developing automated question answering systems is critical to the effectiveness of many applications that rely on real-time knowledge. While previous question answering (QA) datasets have concentrated on formal text like news and Wikipedia, the first large-scale dataset for QA over social media data is presented. To make sure the tweets are meaningful and contain interesting information, tweets used by journalists to write news articles are gathered. Then human annotators are asked to write questions and answers upon these tweets. Unlike other QA datasets like SQuAD in which the answers are extractive, the answer are allowed to be abstractive. The task requires model to read a short tweet and a question and outputs a text phrase (does not need to be in the tweet) as the answer.", "### Supported Tasks and Leaderboards\n\n- 'question-answering': The dataset can be used to train a model for Open-Domain Question Answering where the task is to answer the given questions for a tweet. The performance is measured by comparing the model answers to the the annoted groundtruth and calculating the BLEU-1/Meteor/ROUGE-L score. This task has an active leaderboard which can be found here and ranks models based on BLEU-1, Meteor and ROUGLE-L.", "### Languages\n\nEnglish.", "## Dataset Structure", "### Data Instances\n\nSample data:\n\n\n\nThe test split doesn't include answers so the Answer field is an empty list.", "### Data Fields\n\n- 'Question': a question based on information from a tweet\n- 'Answer': list of possible answers from the tweet\n- 'Tweet': source tweet\n- 'qid': question id", "### Data Splits\n\nThe dataset is split in train, validation and test set. The train set cointains 10692 examples, the validation set 1086 and the test set 1979 examples.", "## Dataset Creation", "### Curation Rationale\n\nWith social media becoming increasingly popular on which lots of news and real-time events are reported, developing automated question answering systems is critical to the effectiveness of many applications that rely on real-time knowledge. While previous question answering (QA) datasets have concentrated on formal text like news and Wikipedia, the first large-scale dataset for QA over social media data is presented. To make sure the tweets are meaningful and contain interesting information, tweets used by journalists to write news articles are gathered. Then human annotators are asked to write questions and answers upon these tweets. Unlike other QA datasets like SQuAD in which the answers are extractive, the answer are allowed to be abstractive. The task requires model to read a short tweet and a question and outputs a text phrase (does not need to be in the tweet) as the answer.", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe authors look into the the archived snapshots of two major news websites (CNN, NBC), and then extract the tweet blocks that are embedded in the news articles. In order to get enough data, they first extract the URLs of all section pages (e.g. World, Politics, Money, Tech) from the snapshot of each home page and then crawl all articles with tweets from these section pages. Then, they filter out the tweets that heavily rely on attached media to convey information, for which they utilize a state-of-the-art semantic role labeling model trained on CoNLL-2005 (He et al., 2017) to analyze the predicate-argument structure of the tweets collected from news articles and keep\nonly the tweets with more than two labeled arguments. This filtering process also automatically\nfilters out most of the short tweets. For the tweets collected from CNN, 22.8% of them were filtered\nvia semantic role labeling. For tweets from NBC, 24.1% of the tweets were filtered.", "#### Who are the source language producers?\n\nTwitter users.", "### Annotations", "#### Annotation process\n\nThe Amazon Mechanical Turk workers were used to collect question-answer\npairs for the filtered tweets. For each Human Intelligence Task (HIT), the authors ask the worker to read three tweets and write two question-answer pairs for each tweet. To ensure the quality, they require the workers to be located in major English speaking countries (i.e. Canada, US, and UK) and have an acceptance rate larger than 95%. Since the authors use tweets as context, lots of important information are contained in hashtags or even emojis. Instead of only showing the text to the workers, they use javascript to directly embed the whole tweet into each HIT. This gives workers the same experience as reading tweets via web browsers and help them to better compose questions. To avoid trivial questions that can be simply answered by superficial text matching methods or too challenging questions that require background knowledge, the authors explicitly state the following items in the HIT instructions for question writing:\n- No Yes-no questions should be asked.\n- The question should have at least five words.\n- Videos, images or inserted links should not\nbe considered.\n- No background knowledge should be required to answer the question.\nTo help the workers better follow the instructions, they also include a representative example showing both good and bad questions or answers in the instructions. As for the answers, since the context they consider is relatively shorter than the context of previous datasets, they do not restrict the answers to be in the tweet, otherwise, the task may potentially be simplified as a classification problem. The workers are allowed to write their answers in their own words, but the authors require the answers to be brief and can be directly inferred from the tweets. After they retrieve the QA pairs from all HITs, they conduct further post-filtering to filter out the pairs from workers that obviously do not follow instructions. They remove QA pairs with yes/no answers. Questions with less than five words are also filtered out. This process filtered 13% of the QA pairs. The dataset now includes 10,898 articles, 17,794 tweets, and 13,757 crowdsourced question-answer pairs. All QA pairs were written by 492 individual workers.", "#### Who are the annotators?\n\nAmazon Mechanical Turk workers.", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases\n\nFrom the paper:\n> It is also worth noting that the data collected from social media can not only capture events and developments in real-time but also capture individual opinions and thus requires reasoning related to the authorship of the content as is illustrated in Table 1.\n\n> Specifically, a significant amount of questions require certain reasoning skills that are specific to social media data:\n- Understanding authorship: Since tweets are highly personal, it is critical to understand how questions/tweets related to the authors.\n- Oral English & Tweet English: Tweets are often oral and informal. QA over tweets requires the understanding of common oral English. Our TWEETQA also requires understanding some tweet-specific English, like conversation-style English.\n- Understanding of user IDs & hashtags: Tweets often contains user IDs and hashtags, which are single special tokens. Understanding these special tokens is important to answer person- or event-related questions.", "### Other Known Limitations", "## Additional Information\n\nThe annotated answers are validated by the authors as follows:\nFor the purposes of human performance evaluation and inter-annotator agreement checking, the authors launch a different set of HITs to ask workers to answer questions in the test and development set. The workers are shown with the tweet blocks as well as the questions collected in the previous step. At this step, workers are allowed to label the questions as “NA” if they think the questions are not answerable. They find that 3.1% of the questions are labeled as unanswerable by the workers (for SQuAD, the ratio is 2.6%). Since the answers collected at this step and previous step are written by different workers, the answers can be written in different text forms even they are semantically equal to each other. For example, one answer can be “Hillary Clinton” while the other is “@HillaryClinton”. As it is not straightforward to automatically calculate the overall agreement, they manually check the agreement on a subset of 200 random samples from the development set and ask an independent human moderator to verify the result. It turns out that 90% of the answers pairs are semantically equivalent, 2% of them are partially equivalent (one of them is incomplete) and 8% are totally inconsistent. The answers collected at this step are also used to measure the human performance. 59 individual workers participated in this process.", "### Dataset Curators\n\nXiong, Wenhan and Wu, Jiawei and Wang, Hong and Kulkarni, Vivek and Yu, Mo and Guo, Xiaoxiao and Chang, Shiyu and Wang, William Yang.", "### Licensing Information\n\nCC BY-SA 4.0.", "### Contributions\n\nThanks to @anaerobeth for adding this dataset." ]
[ 106, 7, 120, 51, 202, 119, 6, 6, 28, 51, 44, 5, 203, 4, 246, 13, 5, 499, 15, 8, 8, 7, 212, 7, 320, 50, 12, 17 ]
[ "passage: TAGS\n#task_categories-question-answering #task_ids-open-domain-qa #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-sa-4.0 #arxiv-1907.06292 #region-us \n# Dataset Card for TweetQA## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: TweetQA homepage\n- Repository:\n- Paper: TWEETQA: A Social Media Focused Question Answering Dataset\n- Leaderboard: TweetQA Leaderboard\n- Point of Contact: Wenhan Xiong### Dataset Summary\n\nWith social media becoming increasingly popular on which lots of news and real-time events are reported, developing automated question answering systems is critical to the effectiveness of many applications that rely on real-time knowledge. While previous question answering (QA) datasets have concentrated on formal text like news and Wikipedia, the first large-scale dataset for QA over social media data is presented. To make sure the tweets are meaningful and contain interesting information, tweets used by journalists to write news articles are gathered. Then human annotators are asked to write questions and answers upon these tweets. Unlike other QA datasets like SQuAD in which the answers are extractive, the answer are allowed to be abstractive. The task requires model to read a short tweet and a question and outputs a text phrase (does not need to be in the tweet) as the answer.", "passage: ### Supported Tasks and Leaderboards\n\n- 'question-answering': The dataset can be used to train a model for Open-Domain Question Answering where the task is to answer the given questions for a tweet. The performance is measured by comparing the model answers to the the annoted groundtruth and calculating the BLEU-1/Meteor/ROUGE-L score. This task has an active leaderboard which can be found here and ranks models based on BLEU-1, Meteor and ROUGLE-L.### Languages\n\nEnglish.## Dataset Structure### Data Instances\n\nSample data:\n\n\n\nThe test split doesn't include answers so the Answer field is an empty list.### Data Fields\n\n- 'Question': a question based on information from a tweet\n- 'Answer': list of possible answers from the tweet\n- 'Tweet': source tweet\n- 'qid': question id### Data Splits\n\nThe dataset is split in train, validation and test set. The train set cointains 10692 examples, the validation set 1086 and the test set 1979 examples.## Dataset Creation### Curation Rationale\n\nWith social media becoming increasingly popular on which lots of news and real-time events are reported, developing automated question answering systems is critical to the effectiveness of many applications that rely on real-time knowledge. While previous question answering (QA) datasets have concentrated on formal text like news and Wikipedia, the first large-scale dataset for QA over social media data is presented. To make sure the tweets are meaningful and contain interesting information, tweets used by journalists to write news articles are gathered. Then human annotators are asked to write questions and answers upon these tweets. Unlike other QA datasets like SQuAD in which the answers are extractive, the answer are allowed to be abstractive. The task requires model to read a short tweet and a question and outputs a text phrase (does not need to be in the tweet) as the answer.### Source Data", "passage: #### Initial Data Collection and Normalization\n\nThe authors look into the the archived snapshots of two major news websites (CNN, NBC), and then extract the tweet blocks that are embedded in the news articles. In order to get enough data, they first extract the URLs of all section pages (e.g. World, Politics, Money, Tech) from the snapshot of each home page and then crawl all articles with tweets from these section pages. Then, they filter out the tweets that heavily rely on attached media to convey information, for which they utilize a state-of-the-art semantic role labeling model trained on CoNLL-2005 (He et al., 2017) to analyze the predicate-argument structure of the tweets collected from news articles and keep\nonly the tweets with more than two labeled arguments. This filtering process also automatically\nfilters out most of the short tweets. For the tweets collected from CNN, 22.8% of them were filtered\nvia semantic role labeling. For tweets from NBC, 24.1% of the tweets were filtered.#### Who are the source language producers?\n\nTwitter users.### Annotations", "passage: #### Annotation process\n\nThe Amazon Mechanical Turk workers were used to collect question-answer\npairs for the filtered tweets. For each Human Intelligence Task (HIT), the authors ask the worker to read three tweets and write two question-answer pairs for each tweet. To ensure the quality, they require the workers to be located in major English speaking countries (i.e. Canada, US, and UK) and have an acceptance rate larger than 95%. Since the authors use tweets as context, lots of important information are contained in hashtags or even emojis. Instead of only showing the text to the workers, they use javascript to directly embed the whole tweet into each HIT. This gives workers the same experience as reading tweets via web browsers and help them to better compose questions. To avoid trivial questions that can be simply answered by superficial text matching methods or too challenging questions that require background knowledge, the authors explicitly state the following items in the HIT instructions for question writing:\n- No Yes-no questions should be asked.\n- The question should have at least five words.\n- Videos, images or inserted links should not\nbe considered.\n- No background knowledge should be required to answer the question.\nTo help the workers better follow the instructions, they also include a representative example showing both good and bad questions or answers in the instructions. As for the answers, since the context they consider is relatively shorter than the context of previous datasets, they do not restrict the answers to be in the tweet, otherwise, the task may potentially be simplified as a classification problem. The workers are allowed to write their answers in their own words, but the authors require the answers to be brief and can be directly inferred from the tweets. After they retrieve the QA pairs from all HITs, they conduct further post-filtering to filter out the pairs from workers that obviously do not follow instructions. They remove QA pairs with yes/no answers. Questions with less than five words are also filtered out. This process filtered 13% of the QA pairs. The dataset now includes 10,898 articles, 17,794 tweets, and 13,757 crowdsourced question-answer pairs. All QA pairs were written by 492 individual workers.#### Who are the annotators?\n\nAmazon Mechanical Turk workers.### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases\n\nFrom the paper:\n> It is also worth noting that the data collected from social media can not only capture events and developments in real-time but also capture individual opinions and thus requires reasoning related to the authorship of the content as is illustrated in Table 1.\n\n> Specifically, a significant amount of questions require certain reasoning skills that are specific to social media data:\n- Understanding authorship: Since tweets are highly personal, it is critical to understand how questions/tweets related to the authors.\n- Oral English & Tweet English: Tweets are often oral and informal. QA over tweets requires the understanding of common oral English. Our TWEETQA also requires understanding some tweet-specific English, like conversation-style English.\n- Understanding of user IDs & hashtags: Tweets often contains user IDs and hashtags, which are single special tokens. Understanding these special tokens is important to answer person- or event-related questions.### Other Known Limitations" ]
9a54a59781aa1c0ebdb603592691ce0ca11b44af
# Dataset Card for Bilingual Corpus of Arabic-English Parallel Tweets ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Bilingual Corpus of Arabic-English Parallel Tweets](https://alt.qcri.org/resources/bilingual_corpus_of_parallel_tweets) - **Repository:** - **Paper:** [Aclweb](https://www.aclweb.org/anthology/2020.bucc-1.3/) - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Twitter users often post parallel tweets—tweets that contain the same content but are written in different languages. Parallel tweets can be an important resource for developing machine translation (MT) systems among other natural language processing (NLP) tasks. This resource is a result of a generic method for collecting parallel tweets. Using the method, we compiled a bilingual corpus of English-Arabic parallel tweets and a list of Twitter accounts who post English-Arabic tweets regularly. Additionally, we annotate a subset of Twitter accounts with their countries of origin and topic of interest, which provides insights about the population who post parallel tweets. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances parallelTweets: ``` { "ArabicTweetID": 981111245209243600, "EnglishTweetID": 981111450432401400 } ``` accountList: ``` { 'account': 'HukoomiQatar' } ``` countryTopicAnnotation: ``` { 'account': 'HukoomiQatar', 'country': 'QA', 'topic': 'Gov' } ``` ### Data Fields parallelTweets: - `ArabicTweetID` (int) - `EnglishTweetID` (int) accountList: - `account` (str) countryTopicAnnotation: - `account` (str) - `country` (class label): One of: - "QA", - "BH", - "AE", - "OM", - "SA", - "PL", - "JO", - "IQ", - "Other", - "EG", - "KW", - "SY" - `topic` (class label): One of: - "Gov", - "Culture", - "Education", - "Sports", - "Travel", - "Events", - "Business", - "Science", - "Politics", - "Health", - "Governoment", - "Media", ### Data Splits All configuration have only one split: "test". ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information It is licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0). ### Citation Information ``` @inproceedings{Mubarak2020bilingualtweets, title={Constructing a Bilingual Corpus of Parallel Tweets}, author={Mubarak, Hamdy and Hassan, Sabit and Abdelali, Ahmed}, booktitle={Proceedings of 13th Workshop on Building and Using Comparable Corpora (BUCC)}, address={Marseille, France}, year={2020} } ``` [More Information Needed] ### Contributions Thanks to [@sumanthd17](https://github.com/sumanthd17) for adding this dataset.
tweets_ar_en_parallel
[ "task_categories:translation", "annotations_creators:expert-generated", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:translation", "size_categories:100K<n<1M", "source_datasets:original", "language:ar", "language:en", "license:apache-2.0", "tweets-translation", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated", "no-annotation"], "language_creators": ["found"], "language": ["ar", "en"], "license": ["apache-2.0"], "multilinguality": ["translation"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["translation"], "task_ids": [], "paperswithcode_id": "bilingual-corpus-of-arabic-english-parallel", "pretty_name": "Bilingual Corpus of Arabic-English Parallel Tweets", "tags": ["tweets-translation"], "dataset_info": [{"config_name": "parallelTweets", "features": [{"name": "ArabicTweetID", "dtype": "int64"}, {"name": "EnglishTweetID", "dtype": "int64"}], "splits": [{"name": "test", "num_bytes": 2667296, "num_examples": 166706}], "download_size": 2937626, "dataset_size": 2667296}, {"config_name": "accountList", "features": [{"name": "account", "dtype": "string"}], "splits": [{"name": "test", "num_bytes": 20108, "num_examples": 1389}], "download_size": 2937626, "dataset_size": 20108}, {"config_name": "countryTopicAnnotation", "features": [{"name": "account", "dtype": "string"}, {"name": "country", "dtype": {"class_label": {"names": {"0": "QA", "1": "BH", "2": "AE", "3": "OM", "4": "SA", "5": "PL", "6": "JO", "7": "IQ", "8": "Other", "9": "EG", "10": "KW", "11": "SY"}}}}, {"name": "topic", "dtype": {"class_label": {"names": {"0": "Gov", "1": "Culture", "2": "Education", "3": "Sports", "4": "Travel", "5": "Events", "6": "Business", "7": "Science", "8": "Politics", "9": "Health", "10": "Governoment", "11": "Media"}}}}], "splits": [{"name": "test", "num_bytes": 6036, "num_examples": 200}], "download_size": 2937626, "dataset_size": 6036}]}
2024-01-18T11:17:35+00:00
[]
[ "ar", "en" ]
TAGS #task_categories-translation #annotations_creators-expert-generated #annotations_creators-no-annotation #language_creators-found #multilinguality-translation #size_categories-100K<n<1M #source_datasets-original #language-Arabic #language-English #license-apache-2.0 #tweets-translation #region-us
# Dataset Card for Bilingual Corpus of Arabic-English Parallel Tweets ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: Bilingual Corpus of Arabic-English Parallel Tweets - Repository: - Paper: Aclweb - Leaderboard: - Point of Contact: ### Dataset Summary Twitter users often post parallel tweets—tweets that contain the same content but are written in different languages. Parallel tweets can be an important resource for developing machine translation (MT) systems among other natural language processing (NLP) tasks. This resource is a result of a generic method for collecting parallel tweets. Using the method, we compiled a bilingual corpus of English-Arabic parallel tweets and a list of Twitter accounts who post English-Arabic tweets regularly. Additionally, we annotate a subset of Twitter accounts with their countries of origin and topic of interest, which provides insights about the population who post parallel tweets. ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances parallelTweets: accountList: countryTopicAnnotation: ### Data Fields parallelTweets: - 'ArabicTweetID' (int) - 'EnglishTweetID' (int) accountList: - 'account' (str) countryTopicAnnotation: - 'account' (str) - 'country' (class label): One of: - "QA", - "BH", - "AE", - "OM", - "SA", - "PL", - "JO", - "IQ", - "Other", - "EG", - "KW", - "SY" - 'topic' (class label): One of: - "Gov", - "Culture", - "Education", - "Sports", - "Travel", - "Events", - "Business", - "Science", - "Politics", - "Health", - "Governoment", - "Media", ### Data Splits All configuration have only one split: "test". ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information It is licensed under the Apache License, Version 2.0. ### Contributions Thanks to @sumanthd17 for adding this dataset.
[ "# Dataset Card for Bilingual Corpus of Arabic-English Parallel Tweets", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Bilingual Corpus of Arabic-English Parallel Tweets\n- Repository:\n- Paper: Aclweb\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nTwitter users often post parallel tweets—tweets that contain the same content but are written in different languages. Parallel tweets can be an important resource for developing machine translation (MT) systems among other natural language processing (NLP) tasks. This resource is a result of a generic method for collecting parallel tweets. Using the method, we compiled a bilingual corpus of English-Arabic parallel tweets and a list of Twitter accounts who post English-Arabic tweets regularly. Additionally, we annotate a subset of Twitter accounts with their countries of origin and topic of interest, which provides insights about the population who post parallel tweets.", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances\n\nparallelTweets:\n\n\naccountList:\n\n\ncountryTopicAnnotation:", "### Data Fields\n\nparallelTweets:\n- 'ArabicTweetID' (int)\n- 'EnglishTweetID' (int)\n\naccountList:\n- 'account' (str)\n\ncountryTopicAnnotation:\n- 'account' (str)\n- 'country' (class label): One of:\n - \"QA\",\n - \"BH\",\n - \"AE\",\n - \"OM\",\n - \"SA\",\n - \"PL\",\n - \"JO\",\n - \"IQ\",\n - \"Other\",\n - \"EG\",\n - \"KW\",\n - \"SY\"\n- 'topic' (class label): One of:\n - \"Gov\",\n - \"Culture\",\n - \"Education\",\n - \"Sports\",\n - \"Travel\",\n - \"Events\",\n - \"Business\",\n - \"Science\",\n - \"Politics\",\n - \"Health\",\n - \"Governoment\",\n - \"Media\",", "### Data Splits\n\nAll configuration have only one split: \"test\".", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nIt is licensed under the Apache License, Version 2.0.", "### Contributions\n\nThanks to @sumanthd17 for adding this dataset." ]
[ "TAGS\n#task_categories-translation #annotations_creators-expert-generated #annotations_creators-no-annotation #language_creators-found #multilinguality-translation #size_categories-100K<n<1M #source_datasets-original #language-Arabic #language-English #license-apache-2.0 #tweets-translation #region-us \n", "# Dataset Card for Bilingual Corpus of Arabic-English Parallel Tweets", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Bilingual Corpus of Arabic-English Parallel Tweets\n- Repository:\n- Paper: Aclweb\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nTwitter users often post parallel tweets—tweets that contain the same content but are written in different languages. Parallel tweets can be an important resource for developing machine translation (MT) systems among other natural language processing (NLP) tasks. This resource is a result of a generic method for collecting parallel tweets. Using the method, we compiled a bilingual corpus of English-Arabic parallel tweets and a list of Twitter accounts who post English-Arabic tweets regularly. Additionally, we annotate a subset of Twitter accounts with their countries of origin and topic of interest, which provides insights about the population who post parallel tweets.", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances\n\nparallelTweets:\n\n\naccountList:\n\n\ncountryTopicAnnotation:", "### Data Fields\n\nparallelTweets:\n- 'ArabicTweetID' (int)\n- 'EnglishTweetID' (int)\n\naccountList:\n- 'account' (str)\n\ncountryTopicAnnotation:\n- 'account' (str)\n- 'country' (class label): One of:\n - \"QA\",\n - \"BH\",\n - \"AE\",\n - \"OM\",\n - \"SA\",\n - \"PL\",\n - \"JO\",\n - \"IQ\",\n - \"Other\",\n - \"EG\",\n - \"KW\",\n - \"SY\"\n- 'topic' (class label): One of:\n - \"Gov\",\n - \"Culture\",\n - \"Education\",\n - \"Sports\",\n - \"Travel\",\n - \"Events\",\n - \"Business\",\n - \"Science\",\n - \"Politics\",\n - \"Health\",\n - \"Governoment\",\n - \"Media\",", "### Data Splits\n\nAll configuration have only one split: \"test\".", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nIt is licensed under the Apache License, Version 2.0.", "### Contributions\n\nThanks to @sumanthd17 for adding this dataset." ]
[ 99, 16, 120, 38, 151, 10, 4, 6, 21, 188, 15, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 19, 18 ]
[ "passage: TAGS\n#task_categories-translation #annotations_creators-expert-generated #annotations_creators-no-annotation #language_creators-found #multilinguality-translation #size_categories-100K<n<1M #source_datasets-original #language-Arabic #language-English #license-apache-2.0 #tweets-translation #region-us \n# Dataset Card for Bilingual Corpus of Arabic-English Parallel Tweets## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: Bilingual Corpus of Arabic-English Parallel Tweets\n- Repository:\n- Paper: Aclweb\n- Leaderboard:\n- Point of Contact:### Dataset Summary\n\nTwitter users often post parallel tweets—tweets that contain the same content but are written in different languages. Parallel tweets can be an important resource for developing machine translation (MT) systems among other natural language processing (NLP) tasks. This resource is a result of a generic method for collecting parallel tweets. Using the method, we compiled a bilingual corpus of English-Arabic parallel tweets and a list of Twitter accounts who post English-Arabic tweets regularly. Additionally, we annotate a subset of Twitter accounts with their countries of origin and topic of interest, which provides insights about the population who post parallel tweets.### Supported Tasks and Leaderboards### Languages## Dataset Structure### Data Instances\n\nparallelTweets:\n\n\naccountList:\n\n\ncountryTopicAnnotation:" ]
498dbe69ce09573af624aabcbc3d6530ded7d631
# Dataset Card for Tweets Hate Speech Detection ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Home](https://github.com/sharmaroshan/Twitter-Sentiment-Analysis) - **Repository:** [Repo](https://github.com/sharmaroshan/Twitter-Sentiment-Analysis/blob/master/train_tweet.csv) - **Paper:** - **Leaderboard:** - **Point of Contact:** [Darshan Gandhi](darshangandhi1151@gmail.com) ### Dataset Summary The objective of this task is to detect hate speech in tweets. For the sake of simplicity, we say a tweet contains hate speech if it has a racist or sexist sentiment associated with it. So, the task is to classify racist or sexist tweets from other tweets. Formally, given a training sample of tweets and labels, where label ‘1’ denotes the tweet is racist/sexist and label ‘0’ denotes the tweet is not racist/sexist, your objective is to predict the labels on the given test dataset. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The tweets are primarily in English Language. ## Dataset Structure ### Data Instances The dataset contains a label denoting is the tweet a hate speech or not ``` {'label': 0, # not a hate speech 'tweet': ' @user when a father is dysfunctional and is so selfish he drags his kids into his dysfunction. #run'} ``` ### Data Fields * label : 1 - it is a hate speech, 0 - not a hate speech. * tweet: content of the tweet as a string. ### Data Splits The data contains training data with :31962 entries ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization Crowdsourced from tweets of users #### Who are the source language producers? Cwodsourced from twitter ### Annotations #### Annotation process The data has been precprocessed and a model has been trained to assign the relevant label to the tweet #### Who are the annotators? The data has been provided by Roshan Sharma ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset With the help of this dataset, one can understand more about the human sentiments and also analye the situations when a particular person intends to make use of hatred/racist comments ### Discussion of Biases The data could be cleaned up further for additional purposes such as applying a better feature extraction techniques [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators Roshan Sharma ### Licensing Information [Information](https://github.com/sharmaroshan/Twitter-Sentiment-Analysis/blob/master/LICENSE) ### Citation Information [Citation](https://github.com/sharmaroshan/Twitter-Sentiment-Analysis/blob/master/CONTRIBUTING.md) ### Contributions Thanks to [@darshan-gandhi](https://github.com/darshan-gandhi) for adding this dataset.
tweets_hate_speech_detection
[ "task_categories:text-classification", "task_ids:sentiment-classification", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:gpl-3.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["gpl-3.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["sentiment-classification"], "pretty_name": "Tweets Hate Speech Detection", "dataset_info": {"features": [{"name": "label", "dtype": {"class_label": {"names": {"0": "no-hate-speech", "1": "hate-speech"}}}}, {"name": "tweet", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 3191888, "num_examples": 31962}, {"name": "test", "num_bytes": 1711606, "num_examples": 17197}], "download_size": 4738708, "dataset_size": 4903494}, "train-eval-index": [{"config": "default", "task": "text-classification", "task_id": "binary_classification", "splits": {"train_split": "train"}, "col_mapping": {"tweet": "text", "label": "target", "metrics": [{"type": "accuracy", "name": "Accuracy"}, {"type": "f1", "name": "F1 binary", "args": {"average": "binary"}}, {"type": "precision", "name": "Precision macro", "args": {"average": "macro"}}, {"type": "precision", "name": "Precision micro", "args": {"average": "micro"}}, {"type": "precision", "name": "Precision weighted", "args": {"average": "weighted"}}, {"type": "recall", "name": "Recall macro", "args": {"average": "macro"}}, {"type": "recall", "name": "Recall micro", "args": {"average": "micro"}}, {"type": "recall", "name": "Recall weighted", "args": {"average": "weighted"}}]}}]}
2024-01-18T11:17:36+00:00
[]
[ "en" ]
TAGS #task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-gpl-3.0 #region-us
# Dataset Card for Tweets Hate Speech Detection ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: Home - Repository: Repo - Paper: - Leaderboard: - Point of Contact: Darshan Gandhi ### Dataset Summary The objective of this task is to detect hate speech in tweets. For the sake of simplicity, we say a tweet contains hate speech if it has a racist or sexist sentiment associated with it. So, the task is to classify racist or sexist tweets from other tweets. Formally, given a training sample of tweets and labels, where label ‘1’ denotes the tweet is racist/sexist and label ‘0’ denotes the tweet is not racist/sexist, your objective is to predict the labels on the given test dataset. ### Supported Tasks and Leaderboards ### Languages The tweets are primarily in English Language. ## Dataset Structure ### Data Instances The dataset contains a label denoting is the tweet a hate speech or not ### Data Fields * label : 1 - it is a hate speech, 0 - not a hate speech. * tweet: content of the tweet as a string. ### Data Splits The data contains training data with :31962 entries ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization Crowdsourced from tweets of users #### Who are the source language producers? Cwodsourced from twitter ### Annotations #### Annotation process The data has been precprocessed and a model has been trained to assign the relevant label to the tweet #### Who are the annotators? The data has been provided by Roshan Sharma ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset With the help of this dataset, one can understand more about the human sentiments and also analye the situations when a particular person intends to make use of hatred/racist comments ### Discussion of Biases The data could be cleaned up further for additional purposes such as applying a better feature extraction techniques ### Other Known Limitations ## Additional Information ### Dataset Curators Roshan Sharma ### Licensing Information Information Citation ### Contributions Thanks to @darshan-gandhi for adding this dataset.
[ "# Dataset Card for Tweets Hate Speech Detection", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Home\n- Repository: Repo\n- Paper:\n- Leaderboard:\n- Point of Contact: Darshan Gandhi", "### Dataset Summary\n\nThe objective of this task is to detect hate speech in tweets. For the sake of simplicity, we say a tweet contains hate speech if it has a racist or sexist sentiment associated with it. So, the task is to classify racist or sexist tweets from other tweets.\n\nFormally, given a training sample of tweets and labels, where label ‘1’ denotes the tweet is racist/sexist and label ‘0’ denotes the tweet is not racist/sexist, your objective is to predict the labels on the given test dataset.", "### Supported Tasks and Leaderboards", "### Languages\nThe tweets are primarily in English Language.", "## Dataset Structure", "### Data Instances\n\nThe dataset contains a label denoting is the tweet a hate speech or not", "### Data Fields\n\n* label : 1 - it is a hate speech, 0 - not a hate speech.\n* tweet: content of the tweet as a string.", "### Data Splits\n\nThe data contains training data with :31962 entries", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization\n\nCrowdsourced from tweets of users", "#### Who are the source language producers?\n\nCwodsourced from twitter", "### Annotations", "#### Annotation process\n\nThe data has been precprocessed and a model has been trained to assign the relevant label to the tweet", "#### Who are the annotators?\n\nThe data has been provided by Roshan Sharma", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset\n\nWith the help of this dataset, one can understand more about the human sentiments and also analye the situations when a particular person intends to make use of hatred/racist comments", "### Discussion of Biases\n\nThe data could be cleaned up further for additional purposes such as applying a better feature extraction techniques", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nRoshan Sharma", "### Licensing Information\n\nInformation\n\n\n\nCitation", "### Contributions\n\nThanks to @darshan-gandhi for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-gpl-3.0 #region-us \n", "# Dataset Card for Tweets Hate Speech Detection", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Home\n- Repository: Repo\n- Paper:\n- Leaderboard:\n- Point of Contact: Darshan Gandhi", "### Dataset Summary\n\nThe objective of this task is to detect hate speech in tweets. For the sake of simplicity, we say a tweet contains hate speech if it has a racist or sexist sentiment associated with it. So, the task is to classify racist or sexist tweets from other tweets.\n\nFormally, given a training sample of tweets and labels, where label ‘1’ denotes the tweet is racist/sexist and label ‘0’ denotes the tweet is not racist/sexist, your objective is to predict the labels on the given test dataset.", "### Supported Tasks and Leaderboards", "### Languages\nThe tweets are primarily in English Language.", "## Dataset Structure", "### Data Instances\n\nThe dataset contains a label denoting is the tweet a hate speech or not", "### Data Fields\n\n* label : 1 - it is a hate speech, 0 - not a hate speech.\n* tweet: content of the tweet as a string.", "### Data Splits\n\nThe data contains training data with :31962 entries", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization\n\nCrowdsourced from tweets of users", "#### Who are the source language producers?\n\nCwodsourced from twitter", "### Annotations", "#### Annotation process\n\nThe data has been precprocessed and a model has been trained to assign the relevant label to the tweet", "#### Who are the annotators?\n\nThe data has been provided by Roshan Sharma", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset\n\nWith the help of this dataset, one can understand more about the human sentiments and also analye the situations when a particular person intends to make use of hatred/racist comments", "### Discussion of Biases\n\nThe data could be cleaned up further for additional purposes such as applying a better feature extraction techniques", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nRoshan Sharma", "### Licensing Information\n\nInformation\n\n\n\nCitation", "### Contributions\n\nThanks to @darshan-gandhi for adding this dataset." ]
[ 92, 12, 120, 30, 128, 10, 14, 6, 24, 34, 17, 5, 7, 4, 19, 16, 5, 27, 18, 8, 8, 46, 30, 7, 5, 9, 9, 19 ]
[ "passage: TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-gpl-3.0 #region-us \n# Dataset Card for Tweets Hate Speech Detection## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: Home\n- Repository: Repo\n- Paper:\n- Leaderboard:\n- Point of Contact: Darshan Gandhi### Dataset Summary\n\nThe objective of this task is to detect hate speech in tweets. For the sake of simplicity, we say a tweet contains hate speech if it has a racist or sexist sentiment associated with it. So, the task is to classify racist or sexist tweets from other tweets.\n\nFormally, given a training sample of tweets and labels, where label ‘1’ denotes the tweet is racist/sexist and label ‘0’ denotes the tweet is not racist/sexist, your objective is to predict the labels on the given test dataset.### Supported Tasks and Leaderboards### Languages\nThe tweets are primarily in English Language.## Dataset Structure### Data Instances\n\nThe dataset contains a label denoting is the tweet a hate speech or not### Data Fields\n\n* label : 1 - it is a hate speech, 0 - not a hate speech.\n* tweet: content of the tweet as a string.### Data Splits\n\nThe data contains training data with :31962 entries## Dataset Creation### Curation Rationale### Source Data" ]
50b00a370769822fbd6b9437664511425891283a
# Dataset Card for Twi Text C3 ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.aclweb.org/anthology/2020.lrec-1.335 - **Repository:** https://github.com/ajesujoba/YorubaTwi-Embedding/ - **Paper:** https://www.aclweb.org/anthology/2020.lrec-1.335 - **Leaderboard:** - **Point of Contact:** [Kwabena Amponsah-Kaakyire](mailto:s8kwampo@stud.uni-saarland.de) ### Dataset Summary Twi Text C3 was collected from various sources from the web (Bible, JW300, wikipedia, etc) to compare pre-trained word embeddings (Fasttext) and embeddings and embeddings trained on curated Twi Texts. The dataset consists of clean texts (i.e the Bible) and noisy texts (with incorrect orthography and mixed dialects) from other online sources like Wikipedia and JW300 ### Supported Tasks and Leaderboards For training word embeddings and language models on Twi texts. ### Languages The language supported is Twi. ## Dataset Structure ### Data Instances A data point is a sentence in each line. { 'text': 'mfitiaseɛ no onyankopɔn bɔɔ ɔsoro ne asaase' } ### Data Fields - `text`: a `string` feature. a sentence text per line ### Data Splits Contains only the training split. ## Dataset Creation ### Curation Rationale The data was created to help introduce resources to new language - Twi. ### Source Data #### Initial Data Collection and Normalization The dataset comes from various sources of the web: Bible, JW300, and wikipedia. See Table 1 in the [paper](https://www.aclweb.org/anthology/2020.lrec-1.335/) for the summary of the dataset and statistics #### Who are the source language producers? [Jehovah Witness](https://www.jw.org/) (JW300) [Twi Bible](http://www.bible.com/) [Yorùbá Wikipedia](dumps.wikimedia.org/twwiki) ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases The dataset is biased to the religion domain (Christianity) because of the inclusion of JW300 and the Bible. ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators The data sets were curated by Kwabena Amponsah-Kaakyire, Jesujoba Alabi, and David Adelani, students of Saarland University, Saarbrücken, Germany . ### Licensing Information The data is under the [Creative Commons Attribution-NonCommercial 4.0 ](https://creativecommons.org/licenses/by-nc/4.0/legalcode) ### Citation Information ``` @inproceedings{alabi-etal-2020-massive, title = "Massive vs. Curated Embeddings for Low-Resourced Languages: the Case of {Y}or{\`u}b{\'a} and {T}wi", author = "Alabi, Jesujoba and Amponsah-Kaakyire, Kwabena and Adelani, David and Espa{\~n}a-Bonet, Cristina", booktitle = "Proceedings of the 12th Language Resources and Evaluation Conference", month = may, year = "2020", address = "Marseille, France", publisher = "European Language Resources Association", url = "https://www.aclweb.org/anthology/2020.lrec-1.335", pages = "2754--2762", abstract = "The success of several architectures to learn semantic representations from unannotated text and the availability of these kind of texts in online multilingual resources such as Wikipedia has facilitated the massive and automatic creation of resources for multiple languages. The evaluation of such resources is usually done for the high-resourced languages, where one has a smorgasbord of tasks and test sets to evaluate on. For low-resourced languages, the evaluation is more difficult and normally ignored, with the hope that the impressive capability of deep learning architectures to learn (multilingual) representations in the high-resourced setting holds in the low-resourced setting too. In this paper we focus on two African languages, Yor{\`u}b{\'a} and Twi, and compare the word embeddings obtained in this way, with word embeddings obtained from curated corpora and a language-dependent processing. We analyse the noise in the publicly available corpora, collect high quality and noisy data for the two languages and quantify the improvements that depend not only on the amount of data but on the quality too. We also use different architectures that learn word representations both from surface forms and characters to further exploit all the available information which showed to be important for these languages. For the evaluation, we manually translate the wordsim-353 word pairs dataset from English into Yor{\`u}b{\'a} and Twi. We extend the analysis to contextual word embeddings and evaluate multilingual BERT on a named entity recognition task. For this, we annotate with named entities the Global Voices corpus for Yor{\`u}b{\'a}. As output of the work, we provide corpora, embeddings and the test suits for both languages.", language = "English", ISBN = "979-10-95546-34-4", } ``` ### Contributions Thanks to [@dadelani](https://github.com/dadelani) for adding this dataset.
twi_text_c3
[ "task_categories:text-generation", "task_categories:fill-mask", "task_ids:language-modeling", "task_ids:masked-language-modeling", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:tw", "license:cc-by-nc-4.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found"], "language": ["tw"], "license": ["cc-by-nc-4.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["text-generation", "fill-mask"], "task_ids": ["language-modeling", "masked-language-modeling"], "pretty_name": "Twi Text C3", "dataset_info": {"features": [{"name": "text", "dtype": "string"}], "config_name": "plain_text", "splits": [{"name": "train", "num_bytes": 71198430, "num_examples": 675772}], "download_size": 69170842, "dataset_size": 71198430}}
2024-01-18T11:17:37+00:00
[]
[ "tw" ]
TAGS #task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Twi #license-cc-by-nc-4.0 #region-us
# Dataset Card for Twi Text C3 ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: URL - Paper: URL - Leaderboard: - Point of Contact: Kwabena Amponsah-Kaakyire ### Dataset Summary Twi Text C3 was collected from various sources from the web (Bible, JW300, wikipedia, etc) to compare pre-trained word embeddings (Fasttext) and embeddings and embeddings trained on curated Twi Texts. The dataset consists of clean texts (i.e the Bible) and noisy texts (with incorrect orthography and mixed dialects) from other online sources like Wikipedia and JW300 ### Supported Tasks and Leaderboards For training word embeddings and language models on Twi texts. ### Languages The language supported is Twi. ## Dataset Structure ### Data Instances A data point is a sentence in each line. { 'text': 'mfitiaseɛ no onyankopɔn bɔɔ ɔsoro ne asaase' } ### Data Fields - 'text': a 'string' feature. a sentence text per line ### Data Splits Contains only the training split. ## Dataset Creation ### Curation Rationale The data was created to help introduce resources to new language - Twi. ### Source Data #### Initial Data Collection and Normalization The dataset comes from various sources of the web: Bible, JW300, and wikipedia. See Table 1 in the paper for the summary of the dataset and statistics #### Who are the source language producers? Jehovah Witness (JW300) Twi Bible Yorùbá Wikipedia ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases The dataset is biased to the religion domain (Christianity) because of the inclusion of JW300 and the Bible. ### Other Known Limitations ## Additional Information ### Dataset Curators The data sets were curated by Kwabena Amponsah-Kaakyire, Jesujoba Alabi, and David Adelani, students of Saarland University, Saarbrücken, Germany . ### Licensing Information The data is under the Creative Commons Attribution-NonCommercial 4.0 ### Contributions Thanks to @dadelani for adding this dataset.
[ "# Dataset Card for Twi Text C3", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: URL\n- Leaderboard:\n- Point of Contact: Kwabena Amponsah-Kaakyire", "### Dataset Summary\n\nTwi Text C3 was collected from various sources from the web (Bible, JW300, wikipedia, etc)\nto compare pre-trained word embeddings (Fasttext) and embeddings and embeddings trained on curated Twi Texts. \nThe dataset consists of clean texts (i.e the Bible) and noisy texts (with incorrect orthography and mixed dialects)\nfrom other online sources like Wikipedia and JW300", "### Supported Tasks and Leaderboards\n\nFor training word embeddings and language models on Twi texts.", "### Languages\n\nThe language supported is Twi.", "## Dataset Structure", "### Data Instances\n\nA data point is a sentence in each line.\n{\n 'text': 'mfitiaseɛ no onyankopɔn bɔɔ ɔsoro ne asaase'\n}", "### Data Fields\n\n- 'text': a 'string' feature.\na sentence text per line", "### Data Splits\n\nContains only the training split.", "## Dataset Creation", "### Curation Rationale\n\nThe data was created to help introduce resources to new language - Twi.", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe dataset comes from various sources of the web: Bible, JW300, and wikipedia. \nSee Table 1 in the paper for the summary of the dataset and statistics", "#### Who are the source language producers?\n\nJehovah Witness (JW300)\nTwi Bible\nYorùbá Wikipedia", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases\n\nThe dataset is biased to the religion domain (Christianity) because of the inclusion of JW300 and the Bible.", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThe data sets were curated by Kwabena Amponsah-Kaakyire, Jesujoba Alabi, and David Adelani, students of Saarland University, Saarbrücken, Germany .", "### Licensing Information\n\n\nThe data is under the Creative Commons Attribution-NonCommercial 4.0", "### Contributions\n\nThanks to @dadelani for adding this dataset." ]
[ "TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Twi #license-cc-by-nc-4.0 #region-us \n", "# Dataset Card for Twi Text C3", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: URL\n- Leaderboard:\n- Point of Contact: Kwabena Amponsah-Kaakyire", "### Dataset Summary\n\nTwi Text C3 was collected from various sources from the web (Bible, JW300, wikipedia, etc)\nto compare pre-trained word embeddings (Fasttext) and embeddings and embeddings trained on curated Twi Texts. \nThe dataset consists of clean texts (i.e the Bible) and noisy texts (with incorrect orthography and mixed dialects)\nfrom other online sources like Wikipedia and JW300", "### Supported Tasks and Leaderboards\n\nFor training word embeddings and language models on Twi texts.", "### Languages\n\nThe language supported is Twi.", "## Dataset Structure", "### Data Instances\n\nA data point is a sentence in each line.\n{\n 'text': 'mfitiaseɛ no onyankopɔn bɔɔ ɔsoro ne asaase'\n}", "### Data Fields\n\n- 'text': a 'string' feature.\na sentence text per line", "### Data Splits\n\nContains only the training split.", "## Dataset Creation", "### Curation Rationale\n\nThe data was created to help introduce resources to new language - Twi.", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe dataset comes from various sources of the web: Bible, JW300, and wikipedia. \nSee Table 1 in the paper for the summary of the dataset and statistics", "#### Who are the source language producers?\n\nJehovah Witness (JW300)\nTwi Bible\nYorùbá Wikipedia", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases\n\nThe dataset is biased to the religion domain (Christianity) because of the inclusion of JW300 and the Bible.", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThe data sets were curated by Kwabena Amponsah-Kaakyire, Jesujoba Alabi, and David Adelani, students of Saarland University, Saarbrücken, Germany .", "### Licensing Information\n\n\nThe data is under the Creative Commons Attribution-NonCommercial 4.0", "### Contributions\n\nThanks to @dadelani for adding this dataset." ]
[ 116, 10, 120, 38, 110, 26, 12, 6, 45, 21, 12, 5, 22, 4, 46, 26, 5, 5, 9, 8, 8, 7, 34, 7, 5, 50, 20, 16 ]
[ "passage: TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Twi #license-cc-by-nc-4.0 #region-us \n# Dataset Card for Twi Text C3## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper: URL\n- Leaderboard:\n- Point of Contact: Kwabena Amponsah-Kaakyire### Dataset Summary\n\nTwi Text C3 was collected from various sources from the web (Bible, JW300, wikipedia, etc)\nto compare pre-trained word embeddings (Fasttext) and embeddings and embeddings trained on curated Twi Texts. \nThe dataset consists of clean texts (i.e the Bible) and noisy texts (with incorrect orthography and mixed dialects)\nfrom other online sources like Wikipedia and JW300### Supported Tasks and Leaderboards\n\nFor training word embeddings and language models on Twi texts.### Languages\n\nThe language supported is Twi.## Dataset Structure### Data Instances\n\nA data point is a sentence in each line.\n{\n 'text': 'mfitiaseɛ no onyankopɔn bɔɔ ɔsoro ne asaase'\n}### Data Fields\n\n- 'text': a 'string' feature.\na sentence text per line" ]
79142c9dc65d27f8190a589d8c854cdc0bed7555
# Dataset Card for Yorùbá Wordsim-353 ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** -https://www.aclweb.org/anthology/2020.lrec-1.335/ - **Repository:** https://github.com/ajesujoba/YorubaTwi-Embedding - **Paper:** https://www.aclweb.org/anthology/2020.lrec-1.335/ - **Leaderboard:** - - **Point of Contact:** [Kwabena Amponsah-Kaakyire](mailto:s8kwampo@stud.uni-saarland.de) ### Dataset Summary A translation of the word pair similarity dataset wordsim-353 to Twi. However, only 274 (out of 353) pairs of words were translated ### Supported Tasks and Leaderboards [More Information Needed] ### Languages Twi (ISO 639-1: tw) ## Dataset Structure ### Data Instances An instance consists of a pair of words as well as their similarity. The dataset contains both the original English words (from wordsim-353) as well as their translation to Twi. ### Data Fields - `twi1`: the first word of the pair; translation to Twi - `twi2`: the second word of the pair; translation to Twi - `similarity`: similarity rating according to the English dataset ### Data Splits Only the test data is available ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ``` @inproceedings{alabi-etal-2020-massive, title = "Massive vs. Curated Embeddings for Low-Resourced Languages: the Case of {Y}or{\`u}b{\'a} and {T}wi", author = "Alabi, Jesujoba and Amponsah-Kaakyire, Kwabena and Adelani, David and Espa{\~n}a-Bonet, Cristina", booktitle = "Proceedings of the 12th Language Resources and Evaluation Conference", month = may, year = "2020", address = "Marseille, France", publisher = "European Language Resources Association", url = "https://www.aclweb.org/anthology/2020.lrec-1.335", pages = "2754--2762", abstract = "The success of several architectures to learn semantic representations from unannotated text and the availability of these kind of texts in online multilingual resources such as Wikipedia has facilitated the massive and automatic creation of resources for multiple languages. The evaluation of such resources is usually done for the high-resourced languages, where one has a smorgasbord of tasks and test sets to evaluate on. For low-resourced languages, the evaluation is more difficult and normally ignored, with the hope that the impressive capability of deep learning architectures to learn (multilingual) representations in the high-resourced setting holds in the low-resourced setting too. In this paper we focus on two African languages, Yor{\`u}b{\'a} and Twi, and compare the word embeddings obtained in this way, with word embeddings obtained from curated corpora and a language-dependent processing. We analyse the noise in the publicly available corpora, collect high quality and noisy data for the two languages and quantify the improvements that depend not only on the amount of data but on the quality too. We also use different architectures that learn word representations both from surface forms and characters to further exploit all the available information which showed to be important for these languages. For the evaluation, we manually translate the wordsim-353 word pairs dataset from English into Yor{\`u}b{\'a} and Twi. We extend the analysis to contextual word embeddings and evaluate multilingual BERT on a named entity recognition task. For this, we annotate with named entities the Global Voices corpus for Yor{\`u}b{\'a}. As output of the work, we provide corpora, embeddings and the test suits for both languages.", language = "English", ISBN = "979-10-95546-34-4", } ``` ### Contributions Thanks to [@dadelani](https://github.com/dadelani) for adding this dataset.
twi_wordsim353
[ "task_categories:text-classification", "task_ids:text-scoring", "task_ids:semantic-similarity-scoring", "annotations_creators:crowdsourced", "language_creators:expert-generated", "multilinguality:multilingual", "size_categories:n<1K", "source_datasets:original", "language:en", "language:tw", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["expert-generated"], "language": ["en", "tw"], "license": ["unknown"], "multilinguality": ["multilingual"], "size_categories": ["n<1K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["text-scoring", "semantic-similarity-scoring"], "pretty_name": "Yor\u00f9b\u00e1 Wordsim-353", "dataset_info": {"features": [{"name": "twi1", "dtype": "string"}, {"name": "twi2", "dtype": "string"}, {"name": "similarity", "dtype": "float32"}], "splits": [{"name": "test", "num_bytes": 7285, "num_examples": 274}], "download_size": 6141, "dataset_size": 7285}}
2024-01-18T11:17:38+00:00
[]
[ "en", "tw" ]
TAGS #task_categories-text-classification #task_ids-text-scoring #task_ids-semantic-similarity-scoring #annotations_creators-crowdsourced #language_creators-expert-generated #multilinguality-multilingual #size_categories-n<1K #source_datasets-original #language-English #language-Twi #license-unknown #region-us
# Dataset Card for Yorùbá Wordsim-353 ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: -URL - Repository: URL - Paper: URL - Leaderboard: - - Point of Contact: Kwabena Amponsah-Kaakyire ### Dataset Summary A translation of the word pair similarity dataset wordsim-353 to Twi. However, only 274 (out of 353) pairs of words were translated ### Supported Tasks and Leaderboards ### Languages Twi (ISO 639-1: tw) ## Dataset Structure ### Data Instances An instance consists of a pair of words as well as their similarity. The dataset contains both the original English words (from wordsim-353) as well as their translation to Twi. ### Data Fields - 'twi1': the first word of the pair; translation to Twi - 'twi2': the second word of the pair; translation to Twi - 'similarity': similarity rating according to the English dataset ### Data Splits Only the test data is available ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @dadelani for adding this dataset.
[ "# Dataset Card for Yorùbá Wordsim-353", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: -URL\n- Repository: URL\n- Paper: URL\n- Leaderboard: -\n- Point of Contact: Kwabena Amponsah-Kaakyire", "### Dataset Summary\n\nA translation of the word pair similarity dataset wordsim-353 to Twi. However, only 274 (out of 353) pairs of words were translated", "### Supported Tasks and Leaderboards", "### Languages\n\nTwi (ISO 639-1: tw)", "## Dataset Structure", "### Data Instances\n\nAn instance consists of a pair of words as well as their similarity. The dataset contains both the original English words (from wordsim-353) as well as their translation to Twi.", "### Data Fields\n\n- 'twi1': the first word of the pair; translation to Twi\n- 'twi2': the second word of the pair; translation to Twi\n- 'similarity': similarity rating according to the English dataset", "### Data Splits\n\nOnly the test data is available", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @dadelani for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #task_ids-text-scoring #task_ids-semantic-similarity-scoring #annotations_creators-crowdsourced #language_creators-expert-generated #multilinguality-multilingual #size_categories-n<1K #source_datasets-original #language-English #language-Twi #license-unknown #region-us \n", "# Dataset Card for Yorùbá Wordsim-353", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: -URL\n- Repository: URL\n- Paper: URL\n- Leaderboard: -\n- Point of Contact: Kwabena Amponsah-Kaakyire", "### Dataset Summary\n\nA translation of the word pair similarity dataset wordsim-353 to Twi. However, only 274 (out of 353) pairs of words were translated", "### Supported Tasks and Leaderboards", "### Languages\n\nTwi (ISO 639-1: tw)", "## Dataset Structure", "### Data Instances\n\nAn instance consists of a pair of words as well as their similarity. The dataset contains both the original English words (from wordsim-353) as well as their translation to Twi.", "### Data Fields\n\n- 'twi1': the first word of the pair; translation to Twi\n- 'twi2': the second word of the pair; translation to Twi\n- 'similarity': similarity rating according to the English dataset", "### Data Splits\n\nOnly the test data is available", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @dadelani for adding this dataset." ]
[ 108, 13, 120, 40, 42, 10, 13, 6, 47, 57, 11, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 16 ]
[ "passage: TAGS\n#task_categories-text-classification #task_ids-text-scoring #task_ids-semantic-similarity-scoring #annotations_creators-crowdsourced #language_creators-expert-generated #multilinguality-multilingual #size_categories-n<1K #source_datasets-original #language-English #language-Twi #license-unknown #region-us \n# Dataset Card for Yorùbá Wordsim-353## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: -URL\n- Repository: URL\n- Paper: URL\n- Leaderboard: -\n- Point of Contact: Kwabena Amponsah-Kaakyire### Dataset Summary\n\nA translation of the word pair similarity dataset wordsim-353 to Twi. However, only 274 (out of 353) pairs of words were translated### Supported Tasks and Leaderboards### Languages\n\nTwi (ISO 639-1: tw)## Dataset Structure### Data Instances\n\nAn instance consists of a pair of words as well as their similarity. The dataset contains both the original English words (from wordsim-353) as well as their translation to Twi.### Data Fields\n\n- 'twi1': the first word of the pair; translation to Twi\n- 'twi2': the second word of the pair; translation to Twi\n- 'similarity': similarity rating according to the English dataset### Data Splits\n\nOnly the test data is available## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?" ]
824c1b749da46e73930be9142d3b6815f2dded02
# Dataset Card for "tydiqa" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://github.com/google-research-datasets/tydiqa](https://github.com/google-research-datasets/tydiqa) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 3.91 GB - **Size of the generated dataset:** 6.10 GB - **Total amount of disk used:** 10.00 GB ### Dataset Summary TyDi QA is a question answering dataset covering 11 typologically diverse languages with 204K question-answer pairs. The languages of TyDi QA are diverse with regard to their typology -- the set of linguistic features that each language expresses -- such that we expect models performing well on this set to generalize across a large number of the languages in the world. It contains language phenomena that would not be found in English-only corpora. To provide a realistic information-seeking task and avoid priming effects, questions are written by people who want to know the answer, but don’t know the answer yet, (unlike SQuAD and its descendents) and the data is collected directly in each language without the use of translation (unlike MLQA and XQuAD). ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### primary_task - **Size of downloaded dataset files:** 1.95 GB - **Size of the generated dataset:** 6.04 GB - **Total amount of disk used:** 7.99 GB An example of 'validation' looks as follows. ``` This example was too long and was cropped: { "annotations": { "minimal_answers_end_byte": [-1, -1, -1], "minimal_answers_start_byte": [-1, -1, -1], "passage_answer_candidate_index": [-1, -1, -1], "yes_no_answer": ["NONE", "NONE", "NONE"] }, "document_plaintext": "\"\\nรองศาสตราจารย์[1] หม่อมราชวงศ์สุขุมพันธุ์ บริพัตร (22 กันยายน 2495 -) ผู้ว่าราชการกรุงเทพมหานครคนที่ 15 อดีตรองหัวหน้าพรรคปร...", "document_title": "หม่อมราชวงศ์สุขุมพันธุ์ บริพัตร", "document_url": "\"https://th.wikipedia.org/wiki/%E0%B8%AB%E0%B8%A1%E0%B9%88%E0%B8%AD%E0%B8%A1%E0%B8%A3%E0%B8%B2%E0%B8%8A%E0%B8%A7%E0%B8%87%E0%B8%...", "language": "thai", "passage_answer_candidates": "{\"plaintext_end_byte\": [494, 1779, 2931, 3904, 4506, 5588, 6383, 7122, 8224, 9375, 10473, 12563, 15134, 17765, 19863, 21902, 229...", "question_text": "\"หม่อมราชวงศ์สุขุมพันธุ์ บริพัตร เรียนจบจากที่ไหน ?\"..." } ``` #### secondary_task - **Size of downloaded dataset files:** 1.95 GB - **Size of the generated dataset:** 58.03 MB - **Total amount of disk used:** 2.01 GB An example of 'validation' looks as follows. ``` This example was too long and was cropped: { "answers": { "answer_start": [394], "text": ["بطولتين"] }, "context": "\"أقيمت البطولة 21 مرة، شارك في النهائيات 78 دولة، وعدد الفرق التي فازت بالبطولة حتى الآن 8 فرق، ويعد المنتخب البرازيلي الأكثر تت...", "id": "arabic-2387335860751143628-1", "question": "\"كم عدد مرات فوز الأوروغواي ببطولة كاس العالم لكرو القدم؟\"...", "title": "قائمة نهائيات كأس العالم" } ``` ### Data Fields The data fields are the same among all splits. #### primary_task - `passage_answer_candidates`: a dictionary feature containing: - `plaintext_start_byte`: a `int32` feature. - `plaintext_end_byte`: a `int32` feature. - `question_text`: a `string` feature. - `document_title`: a `string` feature. - `language`: a `string` feature. - `annotations`: a dictionary feature containing: - `passage_answer_candidate_index`: a `int32` feature. - `minimal_answers_start_byte`: a `int32` feature. - `minimal_answers_end_byte`: a `int32` feature. - `yes_no_answer`: a `string` feature. - `document_plaintext`: a `string` feature. - `document_url`: a `string` feature. #### secondary_task - `id`: a `string` feature. - `title`: a `string` feature. - `context`: a `string` feature. - `question`: a `string` feature. - `answers`: a dictionary feature containing: - `text`: a `string` feature. - `answer_start`: a `int32` feature. ### Data Splits | name | train | validation | | -------------- | -----: | ---------: | | primary_task | 166916 | 18670 | | secondary_task | 49881 | 5077 | ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @article{tydiqa, title = {TyDi QA: A Benchmark for Information-Seeking Question Answering in Typologically Diverse Languages}, author = {Jonathan H. Clark and Eunsol Choi and Michael Collins and Dan Garrette and Tom Kwiatkowski and Vitaly Nikolaev and Jennimaria Palomaki} year = {2020}, journal = {Transactions of the Association for Computational Linguistics} } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@albertvillanova](https://github.com/albertvillanova), [@lewtun](https://github.com/lewtun), [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset.
tydiqa
[ "task_categories:question-answering", "task_ids:extractive-qa", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:multilingual", "size_categories:unknown", "source_datasets:extended|wikipedia", "language:ar", "language:bn", "language:en", "language:fi", "language:id", "language:ja", "language:ko", "language:ru", "language:sw", "language:te", "language:th", "license:apache-2.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["ar", "bn", "en", "fi", "id", "ja", "ko", "ru", "sw", "te", "th"], "license": ["apache-2.0"], "multilinguality": ["multilingual"], "size_categories": ["unknown"], "source_datasets": ["extended|wikipedia"], "task_categories": ["question-answering"], "task_ids": ["extractive-qa"], "paperswithcode_id": "tydi-qa", "pretty_name": "TyDi QA", "dataset_info": [{"config_name": "primary_task", "features": [{"name": "passage_answer_candidates", "sequence": [{"name": "plaintext_start_byte", "dtype": "int32"}, {"name": "plaintext_end_byte", "dtype": "int32"}]}, {"name": "question_text", "dtype": "string"}, {"name": "document_title", "dtype": "string"}, {"name": "language", "dtype": "string"}, {"name": "annotations", "sequence": [{"name": "passage_answer_candidate_index", "dtype": "int32"}, {"name": "minimal_answers_start_byte", "dtype": "int32"}, {"name": "minimal_answers_end_byte", "dtype": "int32"}, {"name": "yes_no_answer", "dtype": "string"}]}, {"name": "document_plaintext", "dtype": "string"}, {"name": "document_url", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 5550574617, "num_examples": 166916}, {"name": "validation", "num_bytes": 484380443, "num_examples": 18670}], "download_size": 1953887429, "dataset_size": 6034955060}, {"config_name": "secondary_task", "features": [{"name": "id", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "context", "dtype": "string"}, {"name": "question", "dtype": "string"}, {"name": "answers", "sequence": [{"name": "text", "dtype": "string"}, {"name": "answer_start", "dtype": "int32"}]}], "splits": [{"name": "train", "num_bytes": 52948607, "num_examples": 49881}, {"name": "validation", "num_bytes": 5006461, "num_examples": 5077}], "download_size": 1953887429, "dataset_size": 57955068}]}
2024-01-18T11:17:40+00:00
[]
[ "ar", "bn", "en", "fi", "id", "ja", "ko", "ru", "sw", "te", "th" ]
TAGS #task_categories-question-answering #task_ids-extractive-qa #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-multilingual #size_categories-unknown #source_datasets-extended|wikipedia #language-Arabic #language-Bengali #language-English #language-Finnish #language-Indonesian #language-Japanese #language-Korean #language-Russian #language-Swahili (macrolanguage) #language-Telugu #language-Thai #license-apache-2.0 #region-us
Dataset Card for "tydiqa" ========================= Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: * Point of Contact: * Size of downloaded dataset files: 3.91 GB * Size of the generated dataset: 6.10 GB * Total amount of disk used: 10.00 GB ### Dataset Summary TyDi QA is a question answering dataset covering 11 typologically diverse languages with 204K question-answer pairs. The languages of TyDi QA are diverse with regard to their typology -- the set of linguistic features that each language expresses -- such that we expect models performing well on this set to generalize across a large number of the languages in the world. It contains language phenomena that would not be found in English-only corpora. To provide a realistic information-seeking task and avoid priming effects, questions are written by people who want to know the answer, but don’t know the answer yet, (unlike SQuAD and its descendents) and the data is collected directly in each language without the use of translation (unlike MLQA and XQuAD). ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- ### Data Instances #### primary\_task * Size of downloaded dataset files: 1.95 GB * Size of the generated dataset: 6.04 GB * Total amount of disk used: 7.99 GB An example of 'validation' looks as follows. #### secondary\_task * Size of downloaded dataset files: 1.95 GB * Size of the generated dataset: 58.03 MB * Total amount of disk used: 2.01 GB An example of 'validation' looks as follows. ### Data Fields The data fields are the same among all splits. #### primary\_task * 'passage\_answer\_candidates': a dictionary feature containing: + 'plaintext\_start\_byte': a 'int32' feature. + 'plaintext\_end\_byte': a 'int32' feature. * 'question\_text': a 'string' feature. * 'document\_title': a 'string' feature. * 'language': a 'string' feature. * 'annotations': a dictionary feature containing: + 'passage\_answer\_candidate\_index': a 'int32' feature. + 'minimal\_answers\_start\_byte': a 'int32' feature. + 'minimal\_answers\_end\_byte': a 'int32' feature. + 'yes\_no\_answer': a 'string' feature. * 'document\_plaintext': a 'string' feature. * 'document\_url': a 'string' feature. #### secondary\_task * 'id': a 'string' feature. * 'title': a 'string' feature. * 'context': a 'string' feature. * 'question': a 'string' feature. * 'answers': a dictionary feature containing: + 'text': a 'string' feature. + 'answer\_start': a 'int32' feature. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @thomwolf, @albertvillanova, @lewtun, @patrickvonplaten for adding this dataset.
[ "### Dataset Summary\n\n\nTyDi QA is a question answering dataset covering 11 typologically diverse languages with 204K question-answer pairs.\nThe languages of TyDi QA are diverse with regard to their typology -- the set of linguistic features that each language\nexpresses -- such that we expect models performing well on this set to generalize across a large number of the languages\nin the world. It contains language phenomena that would not be found in English-only corpora. To provide a realistic\ninformation-seeking task and avoid priming effects, questions are written by people who want to know the answer, but\ndon’t know the answer yet, (unlike SQuAD and its descendents) and the data is collected directly in each language without\nthe use of translation (unlike MLQA and XQuAD).", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### primary\\_task\n\n\n* Size of downloaded dataset files: 1.95 GB\n* Size of the generated dataset: 6.04 GB\n* Total amount of disk used: 7.99 GB\n\n\nAn example of 'validation' looks as follows.", "#### secondary\\_task\n\n\n* Size of downloaded dataset files: 1.95 GB\n* Size of the generated dataset: 58.03 MB\n* Total amount of disk used: 2.01 GB\n\n\nAn example of 'validation' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### primary\\_task\n\n\n* 'passage\\_answer\\_candidates': a dictionary feature containing:\n\t+ 'plaintext\\_start\\_byte': a 'int32' feature.\n\t+ 'plaintext\\_end\\_byte': a 'int32' feature.\n* 'question\\_text': a 'string' feature.\n* 'document\\_title': a 'string' feature.\n* 'language': a 'string' feature.\n* 'annotations': a dictionary feature containing:\n\t+ 'passage\\_answer\\_candidate\\_index': a 'int32' feature.\n\t+ 'minimal\\_answers\\_start\\_byte': a 'int32' feature.\n\t+ 'minimal\\_answers\\_end\\_byte': a 'int32' feature.\n\t+ 'yes\\_no\\_answer': a 'string' feature.\n* 'document\\_plaintext': a 'string' feature.\n* 'document\\_url': a 'string' feature.", "#### secondary\\_task\n\n\n* 'id': a 'string' feature.\n* 'title': a 'string' feature.\n* 'context': a 'string' feature.\n* 'question': a 'string' feature.\n* 'answers': a dictionary feature containing:\n\t+ 'text': a 'string' feature.\n\t+ 'answer\\_start': a 'int32' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @thomwolf, @albertvillanova, @lewtun, @patrickvonplaten for adding this dataset." ]
[ "TAGS\n#task_categories-question-answering #task_ids-extractive-qa #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-multilingual #size_categories-unknown #source_datasets-extended|wikipedia #language-Arabic #language-Bengali #language-English #language-Finnish #language-Indonesian #language-Japanese #language-Korean #language-Russian #language-Swahili (macrolanguage) #language-Telugu #language-Thai #license-apache-2.0 #region-us \n", "### Dataset Summary\n\n\nTyDi QA is a question answering dataset covering 11 typologically diverse languages with 204K question-answer pairs.\nThe languages of TyDi QA are diverse with regard to their typology -- the set of linguistic features that each language\nexpresses -- such that we expect models performing well on this set to generalize across a large number of the languages\nin the world. It contains language phenomena that would not be found in English-only corpora. To provide a realistic\ninformation-seeking task and avoid priming effects, questions are written by people who want to know the answer, but\ndon’t know the answer yet, (unlike SQuAD and its descendents) and the data is collected directly in each language without\nthe use of translation (unlike MLQA and XQuAD).", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### primary\\_task\n\n\n* Size of downloaded dataset files: 1.95 GB\n* Size of the generated dataset: 6.04 GB\n* Total amount of disk used: 7.99 GB\n\n\nAn example of 'validation' looks as follows.", "#### secondary\\_task\n\n\n* Size of downloaded dataset files: 1.95 GB\n* Size of the generated dataset: 58.03 MB\n* Total amount of disk used: 2.01 GB\n\n\nAn example of 'validation' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### primary\\_task\n\n\n* 'passage\\_answer\\_candidates': a dictionary feature containing:\n\t+ 'plaintext\\_start\\_byte': a 'int32' feature.\n\t+ 'plaintext\\_end\\_byte': a 'int32' feature.\n* 'question\\_text': a 'string' feature.\n* 'document\\_title': a 'string' feature.\n* 'language': a 'string' feature.\n* 'annotations': a dictionary feature containing:\n\t+ 'passage\\_answer\\_candidate\\_index': a 'int32' feature.\n\t+ 'minimal\\_answers\\_start\\_byte': a 'int32' feature.\n\t+ 'minimal\\_answers\\_end\\_byte': a 'int32' feature.\n\t+ 'yes\\_no\\_answer': a 'string' feature.\n* 'document\\_plaintext': a 'string' feature.\n* 'document\\_url': a 'string' feature.", "#### secondary\\_task\n\n\n* 'id': a 'string' feature.\n* 'title': a 'string' feature.\n* 'context': a 'string' feature.\n* 'question': a 'string' feature.\n* 'answers': a dictionary feature containing:\n\t+ 'text': a 'string' feature.\n\t+ 'answer\\_start': a 'int32' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @thomwolf, @albertvillanova, @lewtun, @patrickvonplaten for adding this dataset." ]
[ 151, 184, 10, 11, 6, 54, 56, 17, 242, 95, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 6, 34 ]
[ "passage: TAGS\n#task_categories-question-answering #task_ids-extractive-qa #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-multilingual #size_categories-unknown #source_datasets-extended|wikipedia #language-Arabic #language-Bengali #language-English #language-Finnish #language-Indonesian #language-Japanese #language-Korean #language-Russian #language-Swahili (macrolanguage) #language-Telugu #language-Thai #license-apache-2.0 #region-us \n### Dataset Summary\n\n\nTyDi QA is a question answering dataset covering 11 typologically diverse languages with 204K question-answer pairs.\nThe languages of TyDi QA are diverse with regard to their typology -- the set of linguistic features that each language\nexpresses -- such that we expect models performing well on this set to generalize across a large number of the languages\nin the world. It contains language phenomena that would not be found in English-only corpora. To provide a realistic\ninformation-seeking task and avoid priming effects, questions are written by people who want to know the answer, but\ndon’t know the answer yet, (unlike SQuAD and its descendents) and the data is collected directly in each language without\nthe use of translation (unlike MLQA and XQuAD).### Supported Tasks and Leaderboards### Languages\n\n\nDataset Structure\n-----------------### Data Instances#### primary\\_task\n\n\n* Size of downloaded dataset files: 1.95 GB\n* Size of the generated dataset: 6.04 GB\n* Total amount of disk used: 7.99 GB\n\n\nAn example of 'validation' looks as follows.#### secondary\\_task\n\n\n* Size of downloaded dataset files: 1.95 GB\n* Size of the generated dataset: 58.03 MB\n* Total amount of disk used: 2.01 GB\n\n\nAn example of 'validation' looks as follows.### Data Fields\n\n\nThe data fields are the same among all splits." ]
47e5a7623dea983573573940b00e244ea64f6771
# Dataset Card for "ubuntu_dialogs_corpus" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Repository:** https://github.com/rkadlec/ubuntu-ranking-dataset-creator - **Paper:** [The Ubuntu Dialogue Corpus: A Large Dataset for Research in Unstructured Multi-Turn Dialogue Systems](https://arxiv.org/abs/1506.08909) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 0.00 MB - **Size of the generated dataset:** 65.49 MB - **Total amount of disk used:** 65.49 MB ### Dataset Summary Ubuntu Dialogue Corpus, a dataset containing almost 1 million multi-turn dialogues, with a total of over 7 million utterances and 100 million words. This provides a unique resource for research into building dialogue managers based on neural language models that can make use of large amounts of unlabeled data. The dataset has both the multi-turn property of conversations in the Dialog State Tracking Challenge datasets, and the unstructured nature of interactions from microblog services such as Twitter. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### train - **Size of downloaded dataset files:** 0.00 MB - **Size of the generated dataset:** 65.49 MB - **Total amount of disk used:** 65.49 MB An example of 'train' looks as follows. ``` This example was too long and was cropped: { "Context": "\"i think we could import the old comment via rsync , but from there we need to go via email . i think it be easier than cach the...", "Label": 1, "Utterance": "basic each xfree86 upload will not forc user to upgrad 100mb of font for noth __eou__ no someth i do in my spare time . __eou__" } ``` ### Data Fields The data fields are the same among all splits. #### train - `Context`: a `string` feature. - `Utterance`: a `string` feature. - `Label`: a `int32` feature. ### Data Splits |name |train | |-----|-----:| |train|127422| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @article{DBLP:journals/corr/LowePSP15, author = {Ryan Lowe and Nissan Pow and Iulian Serban and Joelle Pineau}, title = {The Ubuntu Dialogue Corpus: {A} Large Dataset for Research in Unstructured Multi-Turn Dialogue Systems}, journal = {CoRR}, volume = {abs/1506.08909}, year = {2015}, url = {http://arxiv.org/abs/1506.08909}, archivePrefix = {arXiv}, eprint = {1506.08909}, timestamp = {Mon, 13 Aug 2018 16:48:23 +0200}, biburl = {https://dblp.org/rec/journals/corr/LowePSP15.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@patrickvonplaten](https://github.com/patrickvonplaten), [@lewtun](https://github.com/lewtun) for adding this dataset.
ubuntu_dialogs_corpus
[ "task_categories:conversational", "task_ids:dialogue-generation", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:1M<n<10M", "source_datasets:original", "language:en", "license:unknown", "arxiv:1506.08909", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["1M<n<10M"], "source_datasets": ["original"], "task_categories": ["conversational"], "task_ids": ["dialogue-generation"], "paperswithcode_id": "ubuntu-dialogue-corpus", "pretty_name": "UDC (Ubuntu Dialogue Corpus)", "dataset_info": [{"config_name": "train", "features": [{"name": "Context", "dtype": "string"}, {"name": "Utterance", "dtype": "string"}, {"name": "Label", "dtype": "int32"}], "splits": [{"name": "train", "num_bytes": 525126729, "num_examples": 1000000}], "download_size": 0, "dataset_size": 525126729}, {"config_name": "dev_test", "features": [{"name": "Context", "dtype": "string"}, {"name": "Ground Truth Utterance", "dtype": "string"}, {"name": "Distractor_0", "dtype": "string"}, {"name": "Distractor_1", "dtype": "string"}, {"name": "Distractor_2", "dtype": "string"}, {"name": "Distractor_3", "dtype": "string"}, {"name": "Distractor_4", "dtype": "string"}, {"name": "Distractor_5", "dtype": "string"}, {"name": "Distractor_6", "dtype": "string"}, {"name": "Distractor_7", "dtype": "string"}, {"name": "Distractor_8", "dtype": "string"}], "splits": [{"name": "test", "num_bytes": 27060502, "num_examples": 18920}, {"name": "validation", "num_bytes": 27663181, "num_examples": 19560}], "download_size": 0, "dataset_size": 54723683}]}
2024-01-18T11:17:41+00:00
[ "1506.08909" ]
[ "en" ]
TAGS #task_categories-conversational #task_ids-dialogue-generation #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-1M<n<10M #source_datasets-original #language-English #license-unknown #arxiv-1506.08909 #region-us
Dataset Card for "ubuntu\_dialogs\_corpus" ========================================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Repository: URL * Paper: The Ubuntu Dialogue Corpus: A Large Dataset for Research in Unstructured Multi-Turn Dialogue Systems * Point of Contact: * Size of downloaded dataset files: 0.00 MB * Size of the generated dataset: 65.49 MB * Total amount of disk used: 65.49 MB ### Dataset Summary Ubuntu Dialogue Corpus, a dataset containing almost 1 million multi-turn dialogues, with a total of over 7 million utterances and 100 million words. This provides a unique resource for research into building dialogue managers based on neural language models that can make use of large amounts of unlabeled data. The dataset has both the multi-turn property of conversations in the Dialog State Tracking Challenge datasets, and the unstructured nature of interactions from microblog services such as Twitter. ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- ### Data Instances #### train * Size of downloaded dataset files: 0.00 MB * Size of the generated dataset: 65.49 MB * Total amount of disk used: 65.49 MB An example of 'train' looks as follows. ### Data Fields The data fields are the same among all splits. #### train * 'Context': a 'string' feature. * 'Utterance': a 'string' feature. * 'Label': a 'int32' feature. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @thomwolf, @patrickvonplaten, @lewtun for adding this dataset.
[ "### Dataset Summary\n\n\nUbuntu Dialogue Corpus, a dataset containing almost 1 million multi-turn dialogues, with a total of over 7 million utterances and 100 million words. This provides a unique resource for research into building dialogue managers based on neural language models that can make use of large amounts of unlabeled data. The dataset has both the multi-turn property of conversations in the Dialog State Tracking Challenge datasets, and the unstructured nature of interactions from microblog services such as Twitter.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### train\n\n\n* Size of downloaded dataset files: 0.00 MB\n* Size of the generated dataset: 65.49 MB\n* Total amount of disk used: 65.49 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### train\n\n\n* 'Context': a 'string' feature.\n* 'Utterance': a 'string' feature.\n* 'Label': a 'int32' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @thomwolf, @patrickvonplaten, @lewtun for adding this dataset." ]
[ "TAGS\n#task_categories-conversational #task_ids-dialogue-generation #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-1M<n<10M #source_datasets-original #language-English #license-unknown #arxiv-1506.08909 #region-us \n", "### Dataset Summary\n\n\nUbuntu Dialogue Corpus, a dataset containing almost 1 million multi-turn dialogues, with a total of over 7 million utterances and 100 million words. This provides a unique resource for research into building dialogue managers based on neural language models that can make use of large amounts of unlabeled data. The dataset has both the multi-turn property of conversations in the Dialog State Tracking Challenge datasets, and the unstructured nature of interactions from microblog services such as Twitter.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### train\n\n\n* Size of downloaded dataset files: 0.00 MB\n* Size of the generated dataset: 65.49 MB\n* Total amount of disk used: 65.49 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### train\n\n\n* 'Context': a 'string' feature.\n* 'Utterance': a 'string' feature.\n* 'Label': a 'int32' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @thomwolf, @patrickvonplaten, @lewtun for adding this dataset." ]
[ 94, 115, 10, 11, 6, 51, 17, 41, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 6, 28 ]
[ "passage: TAGS\n#task_categories-conversational #task_ids-dialogue-generation #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-1M<n<10M #source_datasets-original #language-English #license-unknown #arxiv-1506.08909 #region-us \n### Dataset Summary\n\n\nUbuntu Dialogue Corpus, a dataset containing almost 1 million multi-turn dialogues, with a total of over 7 million utterances and 100 million words. This provides a unique resource for research into building dialogue managers based on neural language models that can make use of large amounts of unlabeled data. The dataset has both the multi-turn property of conversations in the Dialog State Tracking Challenge datasets, and the unstructured nature of interactions from microblog services such as Twitter.### Supported Tasks and Leaderboards### Languages\n\n\nDataset Structure\n-----------------### Data Instances#### train\n\n\n* Size of downloaded dataset files: 0.00 MB\n* Size of the generated dataset: 65.49 MB\n* Total amount of disk used: 65.49 MB\n\n\nAn example of 'train' looks as follows.### Data Fields\n\n\nThe data fields are the same among all splits.#### train\n\n\n* 'Context': a 'string' feature.\n* 'Utterance': a 'string' feature.\n* 'Label': a 'int32' feature.### Data Splits\n\n\n\nDataset Creation\n----------------### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------### Social Impact of Dataset### Discussion of Biases### Other Known Limitations\n\n\nAdditional Information\n----------------------### Dataset Curators### Licensing Information### Contributions\n\n\nThanks to @thomwolf, @patrickvonplaten, @lewtun for adding this dataset." ]
bd8e3b36f543b2adb1959678450a7092811ebb24
# Dataset Card for The Universal Declaration of Human Rights (UDHR) ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.ohchr.org/en/universal-declaration-of-human-rights, https://unicode.org/udhr/index.html - **Repository:** https://github.com/unicode-org/udhr - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary The Universal Declaration of Human Rights (UDHR) is a milestone document in the history of human rights. Drafted by representatives with different legal and cultural backgrounds from all regions of the world, it set out, for the first time, fundamental human rights to be universally protected. The Declaration was adopted by the UN General Assembly in Paris on 10 December 1948 during its 183rd plenary meeting. © 1996 – 2009 The Office of the High Commissioner for Human Rights This plain text version prepared by the "UDHR in Unicode" project, https://www.unicode.org/udhr. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The dataset includes translations of the document in over 400 languages and dialects. The list of languages can be found [here](https://unicode.org/udhr/translations.html). ## Dataset Structure ### Data Instances Each instance corresponds to a different language and includes information about the language and the full document text. ### Data Fields - `text`: The full document text with each line of text delimited by a newline (`\n`). - `lang_key`: The unique identifier of a given translation. - `lang_name`: The textual description of language/dialect. - `iso639-3`: The [iso639-3](https://iso639-3.sil.org/) language identifier. - `iso15924`: The [iso15924](https://unicode.org/iso15924/iso15924-codes.html) language identifier. - `bcp47`: The [BCP 47](https://www.rfc-editor.org/info/bcp47) language identifier. ### Data Splits Only a `train` split included which includes the full document in all languages. | | train | |--------------------|------:| | Number of examples | 488 | ## Dataset Creation ### Curation Rationale In addition to its social significance, the document set a world record in 1999 for being the most translated document in the world and as such can be useful for settings requiring paired text between many languages. ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset In addition to the social and political significance of the United Nations' Universal Declaration of Human Rights, the document set a world record in 1999 for being the most translated document in the world and as such can be useful for settings requiring paired text between many languages including those that are low resource and significantly underrepresented in NLP research. ### Discussion of Biases [More Information Needed] ### Other Known Limitations Although the document is translated into a very large number of languages, the text is very short and therefore may have limited usefulness for most types of modeling and evaluation. ## Additional Information ### Dataset Curators The txt/xml data files used here were compiled by The Unicode Consortium, which can be found [here](https://unicode.org/udhr/index.html). The original texts can be found on the [United Nations website](https://www.ohchr.org/EN/UDHR/Pages/UDHRIndex.aspx). ### Licensing Information Source text © 1996 – 2022 The Office of the High Commissioner for Human Rights The [Unicode license](https://www.unicode.org/license.txt) applies to these translations. ### Citation Information United Nations. (1998). The Universal Declaration of Human Rights, 1948-1998. New York: United Nations Dept. of Public Information. ### Contributions Thanks to [@joeddav](https://github.com/joeddav) for adding this dataset. Updated May 2022 [@leondz](https://github.com/leondz).
udhr
[ "task_categories:translation", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:multilingual", "size_categories:n<1K", "source_datasets:original", "language:aa", "language:ab", "language:ace", "language:acu", "language:ada", "language:ady", "language:af", "language:agr", "language:aii", "language:ajg", "language:als", "language:alt", "language:am", "language:amc", "language:ame", "language:ami", "language:amr", "language:ar", "language:arl", "language:arn", "language:ast", "language:auc", "language:ay", "language:az", "language:ban", "language:bax", "language:bba", "language:bci", "language:be", "language:bem", "language:bfa", "language:bg", "language:bho", "language:bi", "language:bik", "language:bin", "language:blt", "language:bm", "language:bn", "language:bo", "language:boa", "language:br", "language:bs", "language:buc", "language:bug", "language:bum", "language:ca", "language:cab", "language:cak", "language:cbi", "language:cbr", "language:cbs", "language:cbt", "language:cbu", "language:ccp", "language:ceb", "language:cfm", "language:ch", "language:chj", "language:chk", "language:chr", "language:cic", "language:cjk", "language:cjs", "language:cjy", "language:ckb", "language:cnh", "language:cni", "language:cnr", "language:co", "language:cof", "language:cot", "language:cpu", "language:crh", "language:cri", "language:crs", "language:cs", "language:csa", "language:csw", "language:ctd", "language:cy", "language:da", "language:dag", "language:ddn", "language:de", "language:dga", "language:dip", "language:duu", "language:dv", "language:dyo", "language:dyu", "language:dz", "language:ee", "language:el", "language:en", "language:eo", "language:es", "language:ese", "language:et", "language:eu", "language:eve", "language:evn", "language:fa", "language:fat", "language:fi", "language:fj", "language:fkv", "language:fo", "language:fon", "language:fr", "language:fuf", "language:fur", "language:fuv", "language:fvr", "language:fy", "language:ga", "language:gaa", "language:gag", "language:gan", "language:gd", "language:gjn", "language:gkp", "language:gl", "language:gld", "language:gn", "language:gsw", "language:gu", "language:guc", "language:guu", "language:gv", "language:gyr", "language:ha", "language:hak", "language:haw", "language:he", "language:hi", "language:hil", "language:hlt", "language:hmn", "language:hms", "language:hna", "language:hni", "language:hnj", "language:hns", "language:hr", "language:hsb", "language:hsn", "language:ht", "language:hu", "language:hus", "language:huu", "language:hy", "language:ia", "language:ibb", "language:id", "language:idu", "language:ig", "language:ii", "language:ijs", "language:ilo", "language:io", "language:is", "language:it", "language:iu", "language:ja", "language:jiv", "language:jv", "language:ka", "language:kaa", "language:kbd", "language:kbp", "language:kde", "language:kdh", "language:kea", "language:kek", "language:kg", "language:kha", "language:kjh", "language:kk", "language:kkh", "language:kl", "language:km", "language:kmb", "language:kn", "language:ko", "language:koi", "language:koo", "language:kqn", "language:kqs", "language:kr", "language:kri", "language:krl", "language:ktu", "language:ku", "language:kwi", "language:ky", "language:la", "language:lad", "language:lah", "language:lb", "language:lg", "language:lia", "language:lij", "language:lld", "language:ln", "language:lns", "language:lo", "language:lob", "language:lot", "language:loz", "language:lt", "language:lua", "language:lue", "language:lun", "language:lus", "language:lv", "language:mad", "language:mag", "language:mai", "language:mam", "language:man", "language:maz", "language:mcd", "language:mcf", "language:men", "language:mfq", "language:mg", "language:mh", "language:mi", "language:mic", "language:min", "language:miq", "language:mk", "language:ml", "language:mn", "language:mnw", "language:mor", "language:mos", "language:mr", "language:mt", "language:mto", "language:mxi", "language:mxv", "language:my", "language:mzi", "language:nan", "language:nb", "language:nba", "language:nds", "language:ne", "language:ng", "language:nhn", "language:nio", "language:niu", "language:niv", "language:njo", "language:nku", "language:nl", "language:nn", "language:not", "language:nr", "language:nso", "language:nv", "language:ny", "language:nym", "language:nyn", "language:nzi", "language:oaa", "language:oc", "language:ojb", "language:oki", "language:om", "language:orh", "language:os", "language:ote", "language:pa", "language:pam", "language:pap", "language:pau", "language:pbb", "language:pcd", "language:pcm", "language:pis", "language:piu", "language:pl", "language:pon", "language:pov", "language:ppl", "language:prq", "language:ps", "language:pt", "language:qu", "language:quc", "language:qug", "language:quh", "language:quy", "language:qva", "language:qvc", "language:qvh", "language:qvm", "language:qvn", "language:qwh", "language:qxn", "language:qxu", "language:rar", "language:rgn", "language:rm", "language:rmn", "language:rn", "language:ro", "language:ru", "language:rup", "language:rw", "language:sa", "language:sah", "language:sc", "language:sco", "language:se", "language:sey", "language:sg", "language:shk", "language:shn", "language:shp", "language:si", "language:sk", "language:skr", "language:sl", "language:slr", "language:sm", "language:sn", "language:snk", "language:snn", "language:so", "language:sr", "language:srr", "language:ss", "language:st", "language:su", "language:suk", "language:sus", "language:sv", "language:sw", "language:swb", "language:ta", "language:taj", "language:tbz", "language:tca", "language:tdt", "language:te", "language:tem", "language:tet", "language:tg", "language:th", "language:ti", "language:tiv", "language:tk", "language:tl", "language:tly", "language:tn", "language:to", "language:tob", "language:toi", "language:toj", "language:top", "language:tpi", "language:tr", "language:ts", "language:tsz", "language:tt", "language:tw", "language:ty", "language:tyv", "language:tzh", "language:tzm", "language:tzo", "language:udu", "language:ug", "language:uk", "language:umb", "language:und", "language:ur", "language:ura", "language:uz", "language:vai", "language:ve", "language:vec", "language:vep", "language:vi", "language:vmw", "language:wa", "language:war", "language:wo", "language:wuu", "language:wwa", "language:xh", "language:xsm", "language:yad", "language:yao", "language:yap", "language:yi", "language:ykg", "language:yo", "language:yrk", "language:yua", "language:yue", "language:za", "language:zam", "language:zdj", "language:zgh", "language:zh", "language:zlm", "language:zro", "language:ztu", "language:zu", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["found"], "language": ["aa", "ab", "ace", "acu", "ada", "ady", "af", "agr", "aii", "ajg", "als", "alt", "am", "amc", "ame", "ami", "amr", "ar", "arl", "arn", "ast", "auc", "ay", "az", "ban", "bax", "bba", "bci", "be", "bem", "bfa", "bg", "bho", "bi", "bik", "bin", "blt", "bm", "bn", "bo", "boa", "br", "bs", "buc", "bug", "bum", "ca", "cab", "cak", "cbi", "cbr", "cbs", "cbt", "cbu", "ccp", "ceb", "cfm", "ch", "chj", "chk", "chr", "cic", "cjk", "cjs", "cjy", "ckb", "cnh", "cni", "cnr", "co", "cof", "cot", "cpu", "crh", "cri", "crs", "cs", "csa", "csw", "ctd", "cy", "da", "dag", "ddn", "de", "dga", "dip", "duu", "dv", "dyo", "dyu", "dz", "ee", "el", "en", "eo", "es", "ese", "et", "eu", "eve", "evn", "fa", "fat", "fi", "fj", "fkv", "fo", "fon", "fr", "fuf", "fur", "fuv", "fvr", "fy", "ga", "gaa", "gag", "gan", "gd", "gjn", "gkp", "gl", "gld", "gn", "gsw", "gu", "guc", "guu", "gv", "gyr", "ha", "hak", "haw", "he", "hi", "hil", "hlt", "hmn", "hms", "hna", "hni", "hnj", "hns", "hr", "hsb", "hsn", "ht", "hu", "hus", "huu", "hy", "ia", "ibb", "id", "idu", "ig", "ii", "ijs", "ilo", "io", "is", "it", "iu", "ja", "jiv", "jv", "ka", "kaa", "kbd", "kbp", "kde", "kdh", "kea", "kek", "kg", "kha", "kjh", "kk", "kkh", "kl", "km", "kmb", "kn", "ko", "koi", "koo", "kqn", "kqs", "kr", "kri", "krl", "ktu", "ku", "kwi", "ky", "la", "lad", "lah", "lb", "lg", "lia", "lij", "lld", "ln", "lns", "lo", "lob", "lot", "loz", "lt", "lua", "lue", "lun", "lus", "lv", "mad", "mag", "mai", "mam", "man", "maz", "mcd", "mcf", "men", "mfq", "mg", "mh", "mi", "mic", "min", "miq", "mk", "ml", "mn", "mnw", "mor", "mos", "mr", "mt", "mto", "mxi", "mxv", "my", "mzi", "nan", "nb", "nba", "nds", "ne", "ng", "nhn", "nio", "niu", "niv", "njo", "nku", "nl", "nn", "not", "nr", "nso", "nv", "ny", "nym", "nyn", "nzi", "oaa", "oc", "ojb", "oki", "om", "orh", "os", "ote", "pa", "pam", "pap", "pau", "pbb", "pcd", "pcm", "pis", "piu", "pl", "pon", "pov", "ppl", "prq", "ps", "pt", "qu", "quc", "qug", "quh", "quy", "qva", "qvc", "qvh", "qvm", "qvn", "qwh", "qxn", "qxu", "rar", "rgn", "rm", "rmn", "rn", "ro", "ru", "rup", "rw", "sa", "sah", "sc", "sco", "se", "sey", "sg", "shk", "shn", "shp", "si", "sk", "skr", "sl", "slr", "sm", "sn", "snk", "snn", "so", "sr", "srr", "ss", "st", "su", "suk", "sus", "sv", "sw", "swb", "ta", "taj", "tbz", "tca", "tdt", "te", "tem", "tet", "tg", "th", "ti", "tiv", "tk", "tl", "tly", "tn", "to", "tob", "toi", "toj", "top", "tpi", "tr", "ts", "tsz", "tt", "tw", "ty", "tyv", "tzh", "tzm", "tzo", "udu", "ug", "uk", "umb", "und", "ur", "ura", "uz", "vai", "ve", "vec", "vep", "vi", "vmw", "wa", "war", "wo", "wuu", "wwa", "xh", "xsm", "yad", "yao", "yap", "yi", "ykg", "yo", "yrk", "yua", "yue", "za", "zam", "zdj", "zgh", "zh", "zlm", "zro", "ztu", "zu"], "license": ["unknown"], "multilinguality": ["multilingual"], "size_categories": ["n<1K"], "source_datasets": ["original"], "task_categories": ["translation"], "task_ids": [], "pretty_name": "The Universal Declaration of Human Rights (UDHR)", "language_bcp47": ["az-Cyrl", "az-Latn", "bs-Cyrl", "bs-Latn", "ckb-Latn", "de-1901", "de-1996", "el-monoton", "el-polyton", "fa-AF", "fuf-Adlm", "ha-NE", "ha-NG", "jv-Java", "kg-AO", "kkh-Lana", "mn-Cyrl", "pt-BR", "pt-PT", "rm-puter", "rm-rumgr", "rm-surmiran", "rm-sursilv", "rm-sutsilv", "rm-vallader", "sa-Gran", "sr-Cyrl", "sr-Latn", "ta-LK", "tk-Cyrl", "tk-Latn", "tw-akuapem", "tw-asante", "ug-Arab", "ug-Latn", "uz-Cyrl", "uz-Latn", "vi-Hani", "zh-Hant", "zlm-Arab", "zlm-Latn"], "dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "lang_key", "dtype": "string"}, {"name": "lang_name", "dtype": "string"}, {"name": "iso639-3", "dtype": "string"}, {"name": "iso15924", "dtype": "string"}, {"name": "bcp47", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 6753383, "num_examples": 488}], "download_size": 2389690, "dataset_size": 6753383}}
2024-01-18T11:17:42+00:00
[]
[ "aa", "ab", "ace", "acu", "ada", "ady", "af", "agr", "aii", "ajg", "als", "alt", "am", "amc", "ame", "ami", "amr", "ar", "arl", "arn", "ast", "auc", "ay", "az", "ban", "bax", "bba", "bci", "be", "bem", "bfa", "bg", "bho", "bi", "bik", "bin", "blt", "bm", "bn", "bo", "boa", "br", "bs", "buc", "bug", "bum", "ca", "cab", "cak", "cbi", "cbr", "cbs", "cbt", "cbu", "ccp", "ceb", "cfm", "ch", "chj", "chk", "chr", "cic", "cjk", "cjs", "cjy", "ckb", "cnh", "cni", "cnr", "co", "cof", "cot", "cpu", "crh", "cri", "crs", "cs", "csa", "csw", "ctd", "cy", "da", "dag", "ddn", "de", "dga", "dip", "duu", "dv", "dyo", "dyu", "dz", "ee", "el", "en", "eo", "es", "ese", "et", "eu", "eve", "evn", "fa", "fat", "fi", "fj", "fkv", "fo", "fon", "fr", "fuf", "fur", "fuv", "fvr", "fy", "ga", "gaa", "gag", "gan", "gd", "gjn", "gkp", "gl", "gld", "gn", "gsw", "gu", "guc", "guu", "gv", "gyr", "ha", "hak", "haw", "he", "hi", "hil", "hlt", "hmn", "hms", "hna", "hni", "hnj", "hns", "hr", "hsb", "hsn", "ht", "hu", "hus", "huu", "hy", "ia", "ibb", "id", "idu", "ig", "ii", "ijs", "ilo", "io", "is", "it", "iu", "ja", "jiv", "jv", "ka", "kaa", "kbd", "kbp", "kde", "kdh", "kea", "kek", "kg", "kha", "kjh", "kk", "kkh", "kl", "km", "kmb", "kn", "ko", "koi", "koo", "kqn", "kqs", "kr", "kri", "krl", "ktu", "ku", "kwi", "ky", "la", "lad", "lah", "lb", "lg", "lia", "lij", "lld", "ln", "lns", "lo", "lob", "lot", "loz", "lt", "lua", "lue", "lun", "lus", "lv", "mad", "mag", "mai", "mam", "man", "maz", "mcd", "mcf", "men", "mfq", "mg", "mh", "mi", "mic", "min", "miq", "mk", "ml", "mn", "mnw", "mor", "mos", "mr", "mt", "mto", "mxi", "mxv", "my", "mzi", "nan", "nb", "nba", "nds", "ne", "ng", "nhn", "nio", "niu", "niv", "njo", "nku", "nl", "nn", "not", "nr", "nso", "nv", "ny", "nym", "nyn", "nzi", "oaa", "oc", "ojb", "oki", "om", "orh", "os", "ote", "pa", "pam", "pap", "pau", "pbb", "pcd", "pcm", "pis", "piu", "pl", "pon", "pov", "ppl", "prq", "ps", "pt", "qu", "quc", "qug", "quh", "quy", "qva", "qvc", "qvh", "qvm", "qvn", "qwh", "qxn", "qxu", "rar", "rgn", "rm", "rmn", "rn", "ro", "ru", "rup", "rw", "sa", "sah", "sc", "sco", "se", "sey", "sg", "shk", "shn", "shp", "si", "sk", "skr", "sl", "slr", "sm", "sn", "snk", "snn", "so", "sr", "srr", "ss", "st", "su", "suk", "sus", "sv", "sw", "swb", "ta", "taj", "tbz", "tca", "tdt", "te", "tem", "tet", "tg", "th", "ti", "tiv", "tk", "tl", "tly", "tn", "to", "tob", "toi", "toj", "top", "tpi", "tr", "ts", "tsz", "tt", "tw", "ty", "tyv", "tzh", "tzm", "tzo", "udu", "ug", "uk", "umb", "und", "ur", "ura", "uz", "vai", "ve", "vec", "vep", "vi", "vmw", "wa", "war", "wo", "wuu", "wwa", "xh", "xsm", "yad", "yao", "yap", "yi", "ykg", "yo", "yrk", "yua", "yue", "za", "zam", "zdj", "zgh", "zh", "zlm", "zro", "ztu", "zu" ]
TAGS #task_categories-translation #annotations_creators-no-annotation #language_creators-found #multilinguality-multilingual #size_categories-n<1K #source_datasets-original #language-Afar #language-Abkhazian #language-Achinese #language-Achuar-Shiwiar #language-Adangme #language-Adyghe #language-Afrikaans #language-Aguaruna #language-Assyrian Neo-Aramaic #language-Aja (Benin) #language-Tosk Albanian #language-Southern Altai #language-Amharic #language-Amahuaca #language-Yanesha' #language-Amis #language-Amarakaeri #language-Arabic #language-Arabela #language-Mapudungun #language-Asturian #language-Waorani #language-Aymara #language-Azerbaijani #language-Balinese #language-Bamun #language-Baatonum #language-Baoulé #language-Belarusian #language-Bemba (Zambia) #language-Bari #language-Bulgarian #language-Bhojpuri #language-Bislama #language-Bikol #language-Bini #language-Tai Dam #language-Bambara #language-Bengali #language-Tibetan #language-Bora #language-Breton #language-Bosnian #language-Bushi #language-Buginese #language-Bulu (Cameroon) #language-Catalan #language-Garifuna #language-Kaqchikel #language-Chachi #language-Cashibo-Cacataibo #language-Cashinahua #language-Chayahuita #language-Candoshi-Shapra #language-Chakma #language-Cebuano #language-Falam Chin #language-Chamorro #language-Ojitlán Chinantec #language-Chuukese #language-Cherokee #language-Chickasaw #language-Chokwe #language-Shor #language-Jinyu Chinese #language-Central Kurdish #language-Hakha Chin #language-Asháninka #language-Montenegrin #language-Corsican #language-Colorado #language-Caquinte #language-Pichis Ashéninka #language-Crimean Tatar #language-Sãotomense #language-Seselwa Creole French #language-Czech #language-Chiltepec Chinantec #language-Swampy Cree #language-Tedim Chin #language-Welsh #language-Danish #language-Dagbani #language-Dendi (Benin) #language-German #language-Southern Dagaare #language-Northeastern Dinka #language-Drung #language-Dhivehi #language-Jola-Fonyi #language-Dyula #language-Dzongkha #language-Ewe #language-Modern Greek (1453-) #language-English #language-Esperanto #language-Spanish #language-Ese Ejja #language-Estonian #language-Basque #language-Even #language-Evenki #language-Persian #language-Fanti #language-Finnish #language-Fijian #language-Kven Finnish #language-Faroese #language-Fon #language-French #language-Pular #language-Friulian #language-Nigerian Fulfulde #language-Fur #language-Western Frisian #language-Irish #language-Ga #language-Gagauz #language-Gan Chinese #language-Scottish Gaelic #language-Gonja #language-Guinea Kpelle #language-Galician #language-Nanai #language-Guarani #language-Swiss German #language-Gujarati #language-Wayuu #language-Yanomamö #language-Manx #language-Guarayu #language-Hausa #language-Hakka Chinese #language-Hawaiian #language-Hebrew #language-Hindi #language-Hiligaynon #language-Matu Chin #language-Hmong #language-Southern Qiandong Miao #language-Mina (Cameroon) #language-Hani #language-Hmong Njua #language-Caribbean Hindustani #language-Croatian #language-Upper Sorbian #language-Xiang Chinese #language-Haitian #language-Hungarian #language-Huastec #language-Murui Huitoto #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Ibibio #language-Indonesian #language-Idoma #language-Igbo #language-Sichuan Yi #language-Southeast Ijo #language-Iloko #language-Ido #language-Icelandic #language-Italian #language-Inuktitut #language-Japanese #language-Shuar #language-Javanese #language-Georgian #language-Kara-Kalpak #language-Kabardian #language-Kabiyè #language-Makonde #language-Tem #language-Kabuverdianu #language-Kekchí #language-Kongo #language-Khasi #language-Khakas #language-Kazakh #language-Khün #language-Kalaallisut #language-Khmer #language-Kimbundu #language-Kannada #language-Korean #language-Komi-Permyak #language-Konzo #language-Kaonde #language-Northern Kissi #language-Kanuri #language-Krio #language-Karelian #language-Kituba (Democratic Republic of Congo) #language-Kurdish #language-Awa-Cuaiquer #language-Kirghiz #language-Latin #language-Ladino #language-Lahnda #language-Luxembourgish #language-Ganda #language-West-Central Limba #language-Ligurian #language-Ladin #language-Lingala #language-Lamnso' #language-Lao #language-Lobi #language-Otuho #language-Lozi #language-Lithuanian #language-Luba-Lulua #language-Luvale #language-Lunda #language-Lushai #language-Latvian #language-Madurese #language-Magahi #language-Maithili #language-Mam #language-Mandingo #language-Central Mazahua #language-Sharanahua #language-Matsés #language-Mende (Sierra Leone) #language-Moba #language-Malagasy #language-Marshallese #language-Maori #language-Mi'kmaq #language-Minangkabau #language-Mískito #language-Macedonian #language-Malayalam #language-Mongolian #language-Mon #language-Moro #language-Mossi #language-Marathi #language-Maltese #language-Totontepec Mixe #language-Mozarabic #language-Metlatónoc Mixtec #language-Burmese #language-Ixcatlán Mazatec #language-Min Nan Chinese #language-Norwegian Bokmål #language-Nyemba #language-Low German #language-Nepali (macrolanguage) #language-Ndonga #language-Central Nahuatl #language-Nganasan #language-Niuean #language-Gilyak #language-Ao Naga #language-Bouna Kulango #language-Dutch #language-Norwegian Nynorsk #language-Nomatsiguenga #language-South Ndebele #language-Pedi #language-Navajo #language-Nyanja #language-Nyamwezi #language-Nyankole #language-Nzima #language-Orok #language-Occitan (post 1500) #language-Northwestern Ojibwa #language-Okiek #language-Oromo #language-Oroqen #language-Ossetian #language-Mezquital Otomi #language-Panjabi #language-Pampanga #language-Papiamento #language-Palauan #language-Páez #language-Picard #language-Nigerian Pidgin #language-Pijin #language-Pintupi-Luritja #language-Polish #language-Pohnpeian #language-Upper Guinea Crioulo #language-Pipil #language-Ashéninka Perené #language-Pushto #language-Portuguese #language-Quechua #language-K'iche' #language-Chimborazo Highland Quichua #language-South Bolivian Quechua #language-Ayacucho Quechua #language-Ambo-Pasco Quechua #language-Cajamarca Quechua #language-Huamalíes-Dos de Mayo Huánuco Quechua #language-Margos-Yarowilca-Lauricocha Quechua #language-North Junín Quechua #language-Huaylas Ancash Quechua #language-Northern Conchucos Ancash Quechua #language-Arequipa-La Unión Quechua #language-Rarotongan #language-Romagnol #language-Romansh #language-Balkan Romani #language-Rundi #language-Romanian #language-Russian #language-Macedo-Romanian #language-Kinyarwanda #language-Sanskrit #language-Yakut #language-Sardinian #language-Scots #language-Northern Sami #language-Secoya #language-Sango #language-Shilluk #language-Shan #language-Shipibo-Conibo #language-Sinhala #language-Slovak #language-Saraiki #language-Slovenian #language-Salar #language-Samoan #language-Shona #language-Soninke #language-Siona #language-Somali #language-Serbian #language-Serer #language-Swati #language-Southern Sotho #language-Sundanese #language-Sukuma #language-Susu #language-Swedish #language-Swahili (macrolanguage) #language-Maore Comorian #language-Tamil #language-Eastern Tamang #language-Ditammari #language-Ticuna #language-Tetun Dili #language-Telugu #language-Timne #language-Tetum #language-Tajik #language-Thai #language-Tigrinya #language-Tiv #language-Turkmen #language-Tagalog #language-Talysh #language-Tswana #language-Tonga (Tonga Islands) #language-Toba #language-Tonga (Zambia) #language-Tojolabal #language-Papantla Totonac #language-Tok Pisin #language-Turkish #language-Tsonga #language-Purepecha #language-Tatar #language-Twi #language-Tahitian #language-Tuvinian #language-Tzeltal #language-Central Atlas Tamazight #language-Tzotzil #language-Uduk #language-Uighur #language-Ukrainian #language-Umbundu #language-Undetermined #language-Urdu #language-Urarina #language-Uzbek #language-Vai #language-Venda #language-Venetian #language-Veps #language-Vietnamese #language-Makhuwa #language-Walloon #language-Waray (Philippines) #language-Wolof #language-Wu Chinese #language-Waama #language-Xhosa #language-Kasem #language-Yagua #language-Yao #language-Yapese #language-Yiddish #language-Northern Yukaghir #language-Yoruba #language-Nenets #language-Yucateco #language-Yue Chinese #language-Zhuang #language-Miahuatlán Zapotec #language-Ngazidja Comorian #language-Standard Moroccan Tamazight #language-Chinese #language-Malay (individual language) #language-Záparo #language-Güilá Zapotec #language-Zulu #license-unknown #region-us
Dataset Card for The Universal Declaration of Human Rights (UDHR) ================================================================= Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL URL * Repository: URL * Paper: * Leaderboard: * Point of Contact: ### Dataset Summary The Universal Declaration of Human Rights (UDHR) is a milestone document in the history of human rights. Drafted by representatives with different legal and cultural backgrounds from all regions of the world, it set out, for the first time, fundamental human rights to be universally protected. The Declaration was adopted by the UN General Assembly in Paris on 10 December 1948 during its 183rd plenary meeting. © 1996 – 2009 The Office of the High Commissioner for Human Rights This plain text version prepared by the "UDHR in Unicode" project, URL ### Supported Tasks and Leaderboards ### Languages The dataset includes translations of the document in over 400 languages and dialects. The list of languages can be found here. Dataset Structure ----------------- ### Data Instances Each instance corresponds to a different language and includes information about the language and the full document text. ### Data Fields * 'text': The full document text with each line of text delimited by a newline ('\n'). * 'lang\_key': The unique identifier of a given translation. * 'lang\_name': The textual description of language/dialect. * 'iso639-3': The iso639-3 language identifier. * 'iso15924': The iso15924 language identifier. * 'bcp47': The BCP 47 language identifier. ### Data Splits Only a 'train' split included which includes the full document in all languages. Dataset Creation ---------------- ### Curation Rationale In addition to its social significance, the document set a world record in 1999 for being the most translated document in the world and as such can be useful for settings requiring paired text between many languages. ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset In addition to the social and political significance of the United Nations' Universal Declaration of Human Rights, the document set a world record in 1999 for being the most translated document in the world and as such can be useful for settings requiring paired text between many languages including those that are low resource and significantly underrepresented in NLP research. ### Discussion of Biases ### Other Known Limitations Although the document is translated into a very large number of languages, the text is very short and therefore may have limited usefulness for most types of modeling and evaluation. Additional Information ---------------------- ### Dataset Curators The txt/xml data files used here were compiled by The Unicode Consortium, which can be found here. The original texts can be found on the United Nations website. ### Licensing Information Source text © 1996 – 2022 The Office of the High Commissioner for Human Rights The Unicode license applies to these translations. United Nations. (1998). The Universal Declaration of Human Rights, 1948-1998. New York: United Nations Dept. of Public Information. ### Contributions Thanks to @joeddav for adding this dataset. Updated May 2022 @leondz.
[ "### Dataset Summary\n\n\nThe Universal Declaration of Human Rights (UDHR) is a milestone document in the history of human rights. Drafted by\nrepresentatives with different legal and cultural backgrounds from all regions of the world, it set out, for the\nfirst time, fundamental human rights to be universally protected. The Declaration was adopted by the UN General\nAssembly in Paris on 10 December 1948 during its 183rd plenary meeting.\n\n\n© 1996 – 2009 The Office of the High Commissioner for Human Rights\n\n\nThis plain text version prepared by the \"UDHR in Unicode\" project, URL", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe dataset includes translations of the document in over 400 languages and dialects. The list of languages can be found\nhere.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nEach instance corresponds to a different language and includes information about the language and the full document\ntext.", "### Data Fields\n\n\n* 'text': The full document text with each line of text delimited by a newline ('\\n').\n* 'lang\\_key': The unique identifier of a given translation.\n* 'lang\\_name': The textual description of language/dialect.\n* 'iso639-3': The iso639-3 language identifier.\n* 'iso15924': The iso15924 language identifier.\n* 'bcp47': The BCP 47 language identifier.", "### Data Splits\n\n\nOnly a 'train' split included which includes the full document in all languages.\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nIn addition to its social significance, the document set a world record in 1999 for being the most translated\ndocument in the world and as such can be useful for settings requiring paired text between many languages.", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset\n\n\nIn addition to the social and political significance of the United Nations' Universal Declaration of Human Rights,\nthe document set a world record in 1999 for being the most translated document in the world and as such can be useful\nfor settings requiring paired text between many languages including those that are low resource and significantly\nunderrepresented in NLP research.", "### Discussion of Biases", "### Other Known Limitations\n\n\nAlthough the document is translated into a very large number of languages, the text is very short and therefore may\nhave limited usefulness for most types of modeling and evaluation.\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThe txt/xml data files used here were compiled by The Unicode Consortium, which can be found\nhere. The original texts can be found on the\nUnited Nations website.", "### Licensing Information\n\n\nSource text © 1996 – 2022 The Office of the High Commissioner for Human Rights\n\n\nThe Unicode license applies to these translations.\n\n\nUnited Nations. (1998). The Universal Declaration of Human Rights, 1948-1998. New York: United Nations Dept. of Public Information.", "### Contributions\n\n\nThanks to @joeddav for adding this dataset. Updated May 2022 @leondz." ]
[ "TAGS\n#task_categories-translation #annotations_creators-no-annotation #language_creators-found #multilinguality-multilingual #size_categories-n<1K #source_datasets-original #language-Afar #language-Abkhazian #language-Achinese #language-Achuar-Shiwiar #language-Adangme #language-Adyghe #language-Afrikaans #language-Aguaruna #language-Assyrian Neo-Aramaic #language-Aja (Benin) #language-Tosk Albanian #language-Southern Altai #language-Amharic #language-Amahuaca #language-Yanesha' #language-Amis #language-Amarakaeri #language-Arabic #language-Arabela #language-Mapudungun #language-Asturian #language-Waorani #language-Aymara #language-Azerbaijani #language-Balinese #language-Bamun #language-Baatonum #language-Baoulé #language-Belarusian #language-Bemba (Zambia) #language-Bari #language-Bulgarian #language-Bhojpuri #language-Bislama #language-Bikol #language-Bini #language-Tai Dam #language-Bambara #language-Bengali #language-Tibetan #language-Bora #language-Breton #language-Bosnian #language-Bushi #language-Buginese #language-Bulu (Cameroon) #language-Catalan #language-Garifuna #language-Kaqchikel #language-Chachi #language-Cashibo-Cacataibo #language-Cashinahua #language-Chayahuita #language-Candoshi-Shapra #language-Chakma #language-Cebuano #language-Falam Chin #language-Chamorro #language-Ojitlán Chinantec #language-Chuukese #language-Cherokee #language-Chickasaw #language-Chokwe #language-Shor #language-Jinyu Chinese #language-Central Kurdish #language-Hakha Chin #language-Asháninka #language-Montenegrin #language-Corsican #language-Colorado #language-Caquinte #language-Pichis Ashéninka #language-Crimean Tatar #language-Sãotomense #language-Seselwa Creole French #language-Czech #language-Chiltepec Chinantec #language-Swampy Cree #language-Tedim Chin #language-Welsh #language-Danish #language-Dagbani #language-Dendi (Benin) #language-German #language-Southern Dagaare #language-Northeastern Dinka #language-Drung #language-Dhivehi #language-Jola-Fonyi #language-Dyula #language-Dzongkha #language-Ewe #language-Modern Greek (1453-) #language-English #language-Esperanto #language-Spanish #language-Ese Ejja #language-Estonian #language-Basque #language-Even #language-Evenki #language-Persian #language-Fanti #language-Finnish #language-Fijian #language-Kven Finnish #language-Faroese #language-Fon #language-French #language-Pular #language-Friulian #language-Nigerian Fulfulde #language-Fur #language-Western Frisian #language-Irish #language-Ga #language-Gagauz #language-Gan Chinese #language-Scottish Gaelic #language-Gonja #language-Guinea Kpelle #language-Galician #language-Nanai #language-Guarani #language-Swiss German #language-Gujarati #language-Wayuu #language-Yanomamö #language-Manx #language-Guarayu #language-Hausa #language-Hakka Chinese #language-Hawaiian #language-Hebrew #language-Hindi #language-Hiligaynon #language-Matu Chin #language-Hmong #language-Southern Qiandong Miao #language-Mina (Cameroon) #language-Hani #language-Hmong Njua #language-Caribbean Hindustani #language-Croatian #language-Upper Sorbian #language-Xiang Chinese #language-Haitian #language-Hungarian #language-Huastec #language-Murui Huitoto #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Ibibio #language-Indonesian #language-Idoma #language-Igbo #language-Sichuan Yi #language-Southeast Ijo #language-Iloko #language-Ido #language-Icelandic #language-Italian #language-Inuktitut #language-Japanese #language-Shuar #language-Javanese #language-Georgian #language-Kara-Kalpak #language-Kabardian #language-Kabiyè #language-Makonde #language-Tem #language-Kabuverdianu #language-Kekchí #language-Kongo #language-Khasi #language-Khakas #language-Kazakh #language-Khün #language-Kalaallisut #language-Khmer #language-Kimbundu #language-Kannada #language-Korean #language-Komi-Permyak #language-Konzo #language-Kaonde #language-Northern Kissi #language-Kanuri #language-Krio #language-Karelian #language-Kituba (Democratic Republic of Congo) #language-Kurdish #language-Awa-Cuaiquer #language-Kirghiz #language-Latin #language-Ladino #language-Lahnda #language-Luxembourgish #language-Ganda #language-West-Central Limba #language-Ligurian #language-Ladin #language-Lingala #language-Lamnso' #language-Lao #language-Lobi #language-Otuho #language-Lozi #language-Lithuanian #language-Luba-Lulua #language-Luvale #language-Lunda #language-Lushai #language-Latvian #language-Madurese #language-Magahi #language-Maithili #language-Mam #language-Mandingo #language-Central Mazahua #language-Sharanahua #language-Matsés #language-Mende (Sierra Leone) #language-Moba #language-Malagasy #language-Marshallese #language-Maori #language-Mi'kmaq #language-Minangkabau #language-Mískito #language-Macedonian #language-Malayalam #language-Mongolian #language-Mon #language-Moro #language-Mossi #language-Marathi #language-Maltese #language-Totontepec Mixe #language-Mozarabic #language-Metlatónoc Mixtec #language-Burmese #language-Ixcatlán Mazatec #language-Min Nan Chinese #language-Norwegian Bokmål #language-Nyemba #language-Low German #language-Nepali (macrolanguage) #language-Ndonga #language-Central Nahuatl #language-Nganasan #language-Niuean #language-Gilyak #language-Ao Naga #language-Bouna Kulango #language-Dutch #language-Norwegian Nynorsk #language-Nomatsiguenga #language-South Ndebele #language-Pedi #language-Navajo #language-Nyanja #language-Nyamwezi #language-Nyankole #language-Nzima #language-Orok #language-Occitan (post 1500) #language-Northwestern Ojibwa #language-Okiek #language-Oromo #language-Oroqen #language-Ossetian #language-Mezquital Otomi #language-Panjabi #language-Pampanga #language-Papiamento #language-Palauan #language-Páez #language-Picard #language-Nigerian Pidgin #language-Pijin #language-Pintupi-Luritja #language-Polish #language-Pohnpeian #language-Upper Guinea Crioulo #language-Pipil #language-Ashéninka Perené #language-Pushto #language-Portuguese #language-Quechua #language-K'iche' #language-Chimborazo Highland Quichua #language-South Bolivian Quechua #language-Ayacucho Quechua #language-Ambo-Pasco Quechua #language-Cajamarca Quechua #language-Huamalíes-Dos de Mayo Huánuco Quechua #language-Margos-Yarowilca-Lauricocha Quechua #language-North Junín Quechua #language-Huaylas Ancash Quechua #language-Northern Conchucos Ancash Quechua #language-Arequipa-La Unión Quechua #language-Rarotongan #language-Romagnol #language-Romansh #language-Balkan Romani #language-Rundi #language-Romanian #language-Russian #language-Macedo-Romanian #language-Kinyarwanda #language-Sanskrit #language-Yakut #language-Sardinian #language-Scots #language-Northern Sami #language-Secoya #language-Sango #language-Shilluk #language-Shan #language-Shipibo-Conibo #language-Sinhala #language-Slovak #language-Saraiki #language-Slovenian #language-Salar #language-Samoan #language-Shona #language-Soninke #language-Siona #language-Somali #language-Serbian #language-Serer #language-Swati #language-Southern Sotho #language-Sundanese #language-Sukuma #language-Susu #language-Swedish #language-Swahili (macrolanguage) #language-Maore Comorian #language-Tamil #language-Eastern Tamang #language-Ditammari #language-Ticuna #language-Tetun Dili #language-Telugu #language-Timne #language-Tetum #language-Tajik #language-Thai #language-Tigrinya #language-Tiv #language-Turkmen #language-Tagalog #language-Talysh #language-Tswana #language-Tonga (Tonga Islands) #language-Toba #language-Tonga (Zambia) #language-Tojolabal #language-Papantla Totonac #language-Tok Pisin #language-Turkish #language-Tsonga #language-Purepecha #language-Tatar #language-Twi #language-Tahitian #language-Tuvinian #language-Tzeltal #language-Central Atlas Tamazight #language-Tzotzil #language-Uduk #language-Uighur #language-Ukrainian #language-Umbundu #language-Undetermined #language-Urdu #language-Urarina #language-Uzbek #language-Vai #language-Venda #language-Venetian #language-Veps #language-Vietnamese #language-Makhuwa #language-Walloon #language-Waray (Philippines) #language-Wolof #language-Wu Chinese #language-Waama #language-Xhosa #language-Kasem #language-Yagua #language-Yao #language-Yapese #language-Yiddish #language-Northern Yukaghir #language-Yoruba #language-Nenets #language-Yucateco #language-Yue Chinese #language-Zhuang #language-Miahuatlán Zapotec #language-Ngazidja Comorian #language-Standard Moroccan Tamazight #language-Chinese #language-Malay (individual language) #language-Záparo #language-Güilá Zapotec #language-Zulu #license-unknown #region-us \n", "### Dataset Summary\n\n\nThe Universal Declaration of Human Rights (UDHR) is a milestone document in the history of human rights. Drafted by\nrepresentatives with different legal and cultural backgrounds from all regions of the world, it set out, for the\nfirst time, fundamental human rights to be universally protected. The Declaration was adopted by the UN General\nAssembly in Paris on 10 December 1948 during its 183rd plenary meeting.\n\n\n© 1996 – 2009 The Office of the High Commissioner for Human Rights\n\n\nThis plain text version prepared by the \"UDHR in Unicode\" project, URL", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe dataset includes translations of the document in over 400 languages and dialects. The list of languages can be found\nhere.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nEach instance corresponds to a different language and includes information about the language and the full document\ntext.", "### Data Fields\n\n\n* 'text': The full document text with each line of text delimited by a newline ('\\n').\n* 'lang\\_key': The unique identifier of a given translation.\n* 'lang\\_name': The textual description of language/dialect.\n* 'iso639-3': The iso639-3 language identifier.\n* 'iso15924': The iso15924 language identifier.\n* 'bcp47': The BCP 47 language identifier.", "### Data Splits\n\n\nOnly a 'train' split included which includes the full document in all languages.\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nIn addition to its social significance, the document set a world record in 1999 for being the most translated\ndocument in the world and as such can be useful for settings requiring paired text between many languages.", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset\n\n\nIn addition to the social and political significance of the United Nations' Universal Declaration of Human Rights,\nthe document set a world record in 1999 for being the most translated document in the world and as such can be useful\nfor settings requiring paired text between many languages including those that are low resource and significantly\nunderrepresented in NLP research.", "### Discussion of Biases", "### Other Known Limitations\n\n\nAlthough the document is translated into a very large number of languages, the text is very short and therefore may\nhave limited usefulness for most types of modeling and evaluation.\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThe txt/xml data files used here were compiled by The Unicode Consortium, which can be found\nhere. The original texts can be found on the\nUnited Nations website.", "### Licensing Information\n\n\nSource text © 1996 – 2022 The Office of the High Commissioner for Human Rights\n\n\nThe Unicode license applies to these translations.\n\n\nUnited Nations. (1998). The Universal Declaration of Human Rights, 1948-1998. New York: United Nations Dept. of Public Information.", "### Contributions\n\n\nThanks to @joeddav for adding this dataset. Updated May 2022 @leondz." ]
[ 2739, 124, 10, 39, 26, 117, 29, 53, 4, 10, 10, 5, 5, 9, 18, 82, 8, 51, 45, 61, 26 ]
[ "passage: ", "passage: TAGS\n#task_categories-translation #annotations_creators-no-annotation #language_creators-found #multilinguality-multilingual #size_categories-n<1K #source_datasets-original #language-Afar #language-Abkhazian #language-Achinese #language-Achuar-Shiwiar #language-Adangme #language-Adyghe #language-Afrikaans #language-Aguaruna #language-Assyrian Neo-Aramaic #language-Aja (Benin) #language-Tosk Albanian #language-Southern Altai #language-Amharic #language-Amahuaca #language-Yanesha' #language-Amis #language-Amarakaeri #language-Arabic #language-Arabela #language-Mapudungun #language-Asturian #language-Waorani #language-Aymara #language-Azerbaijani #language-Balinese #language-Bamun #language-Baatonum #language-Baoulé #language-Belarusian #language-Bemba (Zambia) #language-Bari #language-Bulgarian #language-Bhojpuri #language-Bislama #language-Bikol #language-Bini #language-Tai Dam #language-Bambara #language-Bengali #language-Tibetan #language-Bora #language-Breton #language-Bosnian #language-Bushi #language-Buginese #language-Bulu (Cameroon) #language-Catalan #language-Garifuna #language-Kaqchikel #language-Chachi #language-Cashibo-Cacataibo #language-Cashinahua #language-Chayahuita #language-Candoshi-Shapra #language-Chakma #language-Cebuano #language-Falam Chin #language-Chamorro #language-Ojitlán Chinantec #language-Chuukese #language-Cherokee #language-Chickasaw #language-Chokwe #language-Shor #language-Jinyu Chinese #language-Central Kurdish #language-Hakha Chin #language-Asháninka #language-Montenegrin #language-Corsican #language-Colorado #language-Caquinte #language-Pichis Ashéninka #language-Crimean Tatar #language-Sãotomense #language-Seselwa Creole French #language-Czech #language-Chiltepec Chinantec #language-Swampy Cree #language-Tedim Chin #language-Welsh #language-Danish #language-Dagbani #language-Dendi (Benin) #language-German #language-Southern Dagaare #language-Northeastern Dinka #language-Drung #language-Dhivehi #language-Jola-Fonyi #language-Dyula #language-Dzongkha #language-Ewe #language-Modern Greek (1453-) #language-English #language-Esperanto #language-Spanish #language-Ese Ejja #language-Estonian #language-Basque #language-Even #language-Evenki #language-Persian #language-Fanti #language-Finnish #language-Fijian #language-Kven Finnish #language-Faroese #language-Fon #language-French #language-Pular #language-Friulian #language-Nigerian Fulfulde #language-Fur #language-Western Frisian #language-Irish #language-Ga #language-Gagauz #language-Gan Chinese #language-Scottish Gaelic #language-Gonja #language-Guinea Kpelle #language-Galician #language-Nanai #language-Guarani #language-Swiss German #language-Gujarati #language-Wayuu #language-Yanomamö #language-Manx #language-Guarayu #language-Hausa #language-Hakka Chinese #language-Hawaiian #language-Hebrew #language-Hindi #language-Hiligaynon #language-Matu Chin #language-Hmong #language-Southern Qiandong Miao #language-Mina (Cameroon) #language-Hani #language-Hmong Njua #language-Caribbean Hindustani #language-Croatian #language-Upper Sorbian #language-Xiang Chinese #language-Haitian #language-Hungarian #language-Huastec #language-Murui Huitoto #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Ibibio #language-Indonesian #language-Idoma #language-Igbo #language-Sichuan Yi #language-Southeast Ijo #language-Iloko #language-Ido #language-Icelandic #language-Italian #language-Inuktitut #language-Japanese #language-Shuar #language-Javanese #language-Georgian #language-Kara-Kalpak #language-Kabardian #language-Kabiyè #language-Makonde #language-Tem #language-Kabuverdianu #language-Kekchí #language-Kongo #language-Khasi #language-Khakas #language-Kazakh #language-Khün #language-Kalaallisut #language-Khmer #language-Kimbundu #language-Kannada #language-Korean #language-Komi-Permyak #language-Konzo #language-Kaonde #language-Northern Kissi #language-Kanuri #language-Krio #language-Karelian #language-Kituba (Democratic Republic of Congo) #language-Kurdish #language-Awa-Cuaiquer #language-Kirghiz #language-Latin #language-Ladino #language-Lahnda #language-Luxembourgish #language-Ganda #language-West-Central Limba #language-Ligurian #language-Ladin #language-Lingala #language-Lamnso' #language-Lao #language-Lobi #language-Otuho #language-Lozi #language-Lithuanian #language-Luba-Lulua #language-Luvale #language-Lunda #language-Lushai #language-Latvian #language-Madurese #language-Magahi #language-Maithili #language-Mam #language-Mandingo #language-Central Mazahua #language-Sharanahua #language-Matsés #language-Mende (Sierra Leone) #language-Moba #language-Malagasy #language-Marshallese #language-Maori #language-Mi'kmaq #language-Minangkabau #language-Mískito #language-Macedonian #language-Malayalam #language-Mongolian #language-Mon #language-Moro #language-Mossi #language-Marathi #language-Maltese #language-Totontepec Mixe #language-Mozarabic #language-Metlatónoc Mixtec #language-Burmese #language-Ixcatlán Mazatec #language-Min Nan Chinese #language-Norwegian Bokmål #language-Nyemba #language-Low German #language-Nepali (macrolanguage) #language-Ndonga #language-Central Nahuatl #language-Nganasan #language-Niuean #language-Gilyak #language-Ao Naga #language-Bouna Kulango #language-Dutch #language-Norwegian Nynorsk #language-Nomatsiguenga #language-South Ndebele #language-Pedi #language-Navajo #language-Nyanja #language-Nyamwezi #language-Nyankole #language-Nzima #language-Orok #language-Occitan (post 1500) #language-Northwestern Ojibwa #language-Okiek #language-Oromo #language-Oroqen #language-Ossetian #language-Mezquital Otomi #language-Panjabi #language-Pampanga #language-Papiamento #language-Palauan #language-Páez #language-Picard #language-Nigerian Pidgin #language-Pijin #language-Pintupi-Luritja #language-Polish #language-Pohnpeian #language-Upper Guinea Crioulo #language-Pipil #language-Ashéninka Perené #language-Pushto #language-Portuguese #language-Quechua #language-K'iche' #language-Chimborazo Highland Quichua #language-South Bolivian Quechua #language-Ayacucho Quechua #language-Ambo-Pasco Quechua #language-Cajamarca Quechua #language-Huamalíes-Dos de Mayo Huánuco Quechua #language-Margos-Yarowilca-Lauricocha Quechua #language-North Junín Quechua #language-Huaylas Ancash Quechua #language-Northern Conchucos Ancash Quechua #language-Arequipa-La Unión Quechua #language-Rarotongan #language-Romagnol #language-Romansh #language-Balkan Romani #language-Rundi #language-Romanian #language-Russian #language-Macedo-Romanian #language-Kinyarwanda #language-Sanskrit #language-Yakut #language-Sardinian #language-Scots #language-Northern Sami #language-Secoya #language-Sango #language-Shilluk #language-Shan #language-Shipibo-Conibo #language-Sinhala #language-Slovak #language-Saraiki #language-Slovenian #language-Salar #language-Samoan #language-Shona #language-Soninke #language-Siona #language-Somali #language-Serbian #language-Serer #language-Swati #language-Southern Sotho #language-Sundanese #language-Sukuma #language-Susu #language-Swedish #language-Swahili (macrolanguage) #language-Maore Comorian #language-Tamil #language-Eastern Tamang #language-Ditammari #language-Ticuna #language-Tetun Dili #language-Telugu #language-Timne #language-Tetum #language-Tajik #language-Thai #language-Tigrinya #language-Tiv #language-Turkmen #language-Tagalog #language-Talysh #language-Tswana #language-Tonga (Tonga Islands) #language-Toba #language-Tonga (Zambia) #language-Tojolabal #language-Papantla Totonac #language-Tok Pisin #language-Turkish #language-Tsonga #language-Purepecha #language-Tatar #language-Twi #language-Tahitian #language-Tuvinian #language-Tzeltal #language-Central Atlas Tamazight #language-Tzotzil #language-Uduk #language-Uighur #language-Ukrainian #language-Umbundu #language-Undetermined #language-Urdu #language-Urarina #language-Uzbek #language-Vai #language-Venda #language-Venetian #language-Veps #language-Vietnamese #language-Makhuwa #language-Walloon #language-Waray (Philippines) #language-Wolof #language-Wu Chinese #language-Waama #language-Xhosa #language-Kasem #language-Yagua #language-Yao #language-Yapese #language-Yiddish #language-Northern Yukaghir #language-Yoruba #language-Nenets #language-Yucateco #language-Yue Chinese #language-Zhuang #language-Miahuatlán Zapotec #language-Ngazidja Comorian #language-Standard Moroccan Tamazight #language-Chinese #language-Malay (individual language) #language-Záparo #language-Güilá Zapotec #language-Zulu #license-unknown #region-us \n### Dataset Summary\n\n\nThe Universal Declaration of Human Rights (UDHR) is a milestone document in the history of human rights. Drafted by\nrepresentatives with different legal and cultural backgrounds from all regions of the world, it set out, for the\nfirst time, fundamental human rights to be universally protected. The Declaration was adopted by the UN General\nAssembly in Paris on 10 December 1948 during its 183rd plenary meeting.\n\n\n© 1996 – 2009 The Office of the High Commissioner for Human Rights\n\n\nThis plain text version prepared by the \"UDHR in Unicode\" project, URL### Supported Tasks and Leaderboards### Languages\n\n\nThe dataset includes translations of the document in over 400 languages and dialects. The list of languages can be found\nhere.\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nEach instance corresponds to a different language and includes information about the language and the full document\ntext.### Data Fields\n\n\n* 'text': The full document text with each line of text delimited by a newline ('\\n').\n* 'lang\\_key': The unique identifier of a given translation.\n* 'lang\\_name': The textual description of language/dialect.\n* 'iso639-3': The iso639-3 language identifier.\n* 'iso15924': The iso15924 language identifier.\n* 'bcp47': The BCP 47 language identifier.### Data Splits\n\n\nOnly a 'train' split included which includes the full document in all languages.\n\n\n\nDataset Creation\n----------------### Curation Rationale\n\n\nIn addition to its social significance, the document set a world record in 1999 for being the most translated\ndocument in the world and as such can be useful for settings requiring paired text between many languages.### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------" ]
16e284f69547dc12c65476c7f37db7c63957b021
# Dataset Card for UMC005 English-Urdu ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** http://ufal.ms.mff.cuni.cz/umc/005-en-ur/ - **Repository:** None - **Paper:** https://www.researchgate.net/publication/268008206_Word-Order_Issues_in_English-to-Urdu_Statistical_Machine_Translation - **Leaderboard:** [If the dataset supports an active leaderboard, add link here]() - **Point of Contact:** Bushra Jawaid and Daniel Zeman ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@abhishekkrthakur](https://github.com/abhishekkrthakur) for adding this dataset.
um005
[ "task_categories:translation", "annotations_creators:no-annotation", "language_creators:other", "multilinguality:multilingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "language:ur", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["other"], "language": ["en", "ur"], "license": ["unknown"], "multilinguality": ["multilingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["translation"], "task_ids": [], "paperswithcode_id": "umc005-english-urdu", "pretty_name": "UMC005 English-Urdu", "dataset_info": [{"config_name": "bible", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["ur", "en"]}}}], "splits": [{"name": "train", "num_bytes": 2350730, "num_examples": 7400}, {"name": "validation", "num_bytes": 113476, "num_examples": 300}, {"name": "test", "num_bytes": 104678, "num_examples": 257}], "download_size": 3683565, "dataset_size": 2568884}, {"config_name": "quran", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["ur", "en"]}}}], "splits": [{"name": "train", "num_bytes": 2929711, "num_examples": 6000}, {"name": "validation", "num_bytes": 43499, "num_examples": 214}, {"name": "test", "num_bytes": 44413, "num_examples": 200}], "download_size": 3683565, "dataset_size": 3017623}, {"config_name": "all", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["ur", "en"]}}}], "splits": [{"name": "train", "num_bytes": 5280441, "num_examples": 13400}, {"name": "validation", "num_bytes": 156963, "num_examples": 514}, {"name": "test", "num_bytes": 149079, "num_examples": 457}], "download_size": 3683565, "dataset_size": 5586483}]}
2024-01-18T11:17:42+00:00
[]
[ "en", "ur" ]
TAGS #task_categories-translation #annotations_creators-no-annotation #language_creators-other #multilinguality-multilingual #size_categories-1K<n<10K #source_datasets-original #language-English #language-Urdu #license-unknown #region-us
# Dataset Card for UMC005 English-Urdu ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: None - Paper: URL - Leaderboard: [If the dataset supports an active leaderboard, add link here]() - Point of Contact: Bushra Jawaid and Daniel Zeman ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @abhishekkrthakur for adding this dataset.
[ "# Dataset Card for UMC005 English-Urdu", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: URL\n- Leaderboard: [If the dataset supports an active leaderboard, add link here]()\n- Point of Contact: Bushra Jawaid and Daniel Zeman", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ "TAGS\n#task_categories-translation #annotations_creators-no-annotation #language_creators-other #multilinguality-multilingual #size_categories-1K<n<10K #source_datasets-original #language-English #language-Urdu #license-unknown #region-us \n", "# Dataset Card for UMC005 English-Urdu", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: URL\n- Leaderboard: [If the dataset supports an active leaderboard, add link here]()\n- Point of Contact: Bushra Jawaid and Daniel Zeman", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ 80, 12, 120, 53, 6, 10, 4, 6, 6, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 20 ]
[ "passage: TAGS\n#task_categories-translation #annotations_creators-no-annotation #language_creators-other #multilinguality-multilingual #size_categories-1K<n<10K #source_datasets-original #language-English #language-Urdu #license-unknown #region-us \n# Dataset Card for UMC005 English-Urdu## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: URL\n- Leaderboard: [If the dataset supports an active leaderboard, add link here]()\n- Point of Contact: Bushra Jawaid and Daniel Zeman### Dataset Summary### Supported Tasks and Leaderboards### Languages## Dataset Structure### Data Instances### Data Fields### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators### Licensing Information### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
8740ebaf90f279644d75b5ddd018e5f6c0aac21f
<div class="course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400"> <p><b>Deprecated:</b> Dataset "un_ga" is deprecated due to the the unavailability of its source data. It has been superseded by the official United Nations Parallel Corpus, which is recommended for use in its place: <a href="https://huggingface.co/datasets/un_pc">un_pc</a></p> </div> # Dataset Card for [Dataset Name] ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** http://opus.nlpl.eu/UN.php - **Repository:** - **Paper:** https://www.researchgate.net/publication/228579662_United_nations_general_assembly_resolutions_A_six-language_parallel_corpus - **Leaderboard:** - **Point of Contact:** ### Dataset Summary This is a collection of translated documents from the United Nations originally compiled into a translation memory by Alexandre Rafalovitch, Robert Dale (see http://uncorpora.org). ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information @inproceedings{title = "United Nations General Assembly Resolutions: a six-language parallel corpus", abstract = "In this paper we describe a six-ways parallel public-domain corpus consisting of 2100 United Nations General Assembly Resolutions with translations in the six official languages of the United Nations, with an average of around 3 million tokens per language. The corpus is available in a preprocessed, formatting-normalized TMX format with paragraphs aligned across multiple languages. We describe the background to the corpus and its content, the process of its construction, and some of its interesting properties.", author = "Alexandre Rafalovitch and Robert Dale", year = "2009", language = "English", booktitle = "MT Summit XII proceedings", publisher = "International Association of Machine Translation", } ### Contributions Thanks to [@param087](https://github.com/param087) for adding this dataset.
un_ga
[ "task_categories:translation", "annotations_creators:found", "language_creators:found", "multilinguality:translation", "size_categories:10K<n<100K", "source_datasets:original", "language:ar", "language:en", "language:es", "language:fr", "language:ru", "language:zh", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["ar", "en", "es", "fr", "ru", "zh"], "license": ["unknown"], "multilinguality": ["translation"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["translation"], "task_ids": [], "pretty_name": "UnGa", "config_names": ["ar-to-en", "ar-to-es", "ar-to-fr", "ar-to-ru", "ar-to-zh", "en-to-es", "en-to-fr", "en-to-ru", "en-to-zh", "es-to-fr", "es-to-ru", "es-to-zh", "fr-to-ru", "fr-to-zh", "ru-to-zh"], "dataset_info": [{"config_name": "ar_to_en", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["ar", "en"]}}}], "splits": [{"name": "train", "num_bytes": 53122776, "num_examples": 74067}], "download_size": 21418697, "dataset_size": 53122776}, {"config_name": "ar_to_es", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["ar", "es"]}}}], "splits": [{"name": "train", "num_bytes": 55728615, "num_examples": 74067}], "download_size": 22724976, "dataset_size": 55728615}, {"config_name": "ar_to_fr", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["ar", "fr"]}}}], "splits": [{"name": "train", "num_bytes": 55930802, "num_examples": 74067}], "download_size": 23035904, "dataset_size": 55930802}, {"config_name": "ar_to_ru", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["ar", "ru"]}}}], "splits": [{"name": "train", "num_bytes": 72657625, "num_examples": 74067}], "download_size": 28279669, "dataset_size": 72657625}, {"config_name": "ar_to_zh", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["ar", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 48217579, "num_examples": 74067}], "download_size": 20391116, "dataset_size": 48217579}, {"config_name": "en_to_es", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "es"]}}}], "splits": [{"name": "train", "num_bytes": 45358770, "num_examples": 74067}], "download_size": 19229141, "dataset_size": 45358770}, {"config_name": "en_to_fr", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "fr"]}}}], "splits": [{"name": "train", "num_bytes": 45560957, "num_examples": 74067}], "download_size": 19540063, "dataset_size": 45560957}, {"config_name": "en_to_ru", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "ru"]}}}], "splits": [{"name": "train", "num_bytes": 62287780, "num_examples": 74067}], "download_size": 24783812, "dataset_size": 62287780}, {"config_name": "en_to_zh", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 37847734, "num_examples": 74067}], "download_size": 16895275, "dataset_size": 37847734}, {"config_name": "es_to_fr", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["es", "fr"]}}}], "splits": [{"name": "train", "num_bytes": 48166796, "num_examples": 74067}], "download_size": 20846355, "dataset_size": 48166796}, {"config_name": "es_to_ru", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["es", "ru"]}}}], "splits": [{"name": "train", "num_bytes": 64893619, "num_examples": 74067}], "download_size": 26090092, "dataset_size": 64893619}, {"config_name": "es_to_zh", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["es", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 40453573, "num_examples": 74067}], "download_size": 18201560, "dataset_size": 40453573}, {"config_name": "fr_to_ru", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["fr", "ru"]}}}], "splits": [{"name": "train", "num_bytes": 65095806, "num_examples": 74067}], "download_size": 26401015, "dataset_size": 65095806}, {"config_name": "fr_to_zh", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["fr", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 40655760, "num_examples": 74067}], "download_size": 18512482, "dataset_size": 40655760}, {"config_name": "ru_to_zh", "features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["ru", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 57382583, "num_examples": 74067}], "download_size": 23756229, "dataset_size": 57382583}], "configs": [{"config_name": "ar_to_en", "data_files": [{"split": "train", "path": "ar_to_en/train-*"}]}, {"config_name": "ar_to_es", "data_files": [{"split": "train", "path": "ar_to_es/train-*"}]}, {"config_name": "ar_to_fr", "data_files": [{"split": "train", "path": "ar_to_fr/train-*"}]}, {"config_name": "ar_to_ru", "data_files": [{"split": "train", "path": "ar_to_ru/train-*"}]}, {"config_name": "ar_to_zh", "data_files": [{"split": "train", "path": "ar_to_zh/train-*"}]}, {"config_name": "en_to_es", "data_files": [{"split": "train", "path": "en_to_es/train-*"}]}, {"config_name": "en_to_fr", "data_files": [{"split": "train", "path": "en_to_fr/train-*"}]}, {"config_name": "en_to_ru", "data_files": [{"split": "train", "path": "en_to_ru/train-*"}]}, {"config_name": "en_to_zh", "data_files": [{"split": "train", "path": "en_to_zh/train-*"}]}, {"config_name": "es_to_fr", "data_files": [{"split": "train", "path": "es_to_fr/train-*"}]}, {"config_name": "es_to_ru", "data_files": [{"split": "train", "path": "es_to_ru/train-*"}]}, {"config_name": "es_to_zh", "data_files": [{"split": "train", "path": "es_to_zh/train-*"}]}, {"config_name": "fr_to_ru", "data_files": [{"split": "train", "path": "fr_to_ru/train-*"}]}, {"config_name": "fr_to_zh", "data_files": [{"split": "train", "path": "fr_to_zh/train-*"}]}, {"config_name": "ru_to_zh", "data_files": [{"split": "train", "path": "ru_to_zh/train-*"}]}]}
2024-02-13T09:14:43+00:00
[]
[ "ar", "en", "es", "fr", "ru", "zh" ]
TAGS #task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-translation #size_categories-10K<n<100K #source_datasets-original #language-Arabic #language-English #language-Spanish #language-French #language-Russian #language-Chinese #license-unknown #region-us
<div class="course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400"> <p><b>Deprecated:</b> Dataset "un_ga" is deprecated due to the the unavailability of its source data. It has been superseded by the official United Nations Parallel Corpus, which is recommended for use in its place: <a href="URL </div> # Dataset Card for [Dataset Name] ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: URL - Leaderboard: - Point of Contact: ### Dataset Summary This is a collection of translated documents from the United Nations originally compiled into a translation memory by Alexandre Rafalovitch, Robert Dale (see URL). ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information @inproceedings{title = "United Nations General Assembly Resolutions: a six-language parallel corpus", abstract = "In this paper we describe a six-ways parallel public-domain corpus consisting of 2100 United Nations General Assembly Resolutions with translations in the six official languages of the United Nations, with an average of around 3 million tokens per language. The corpus is available in a preprocessed, formatting-normalized TMX format with paragraphs aligned across multiple languages. We describe the background to the corpus and its content, the process of its construction, and some of its interesting properties.", author = "Alexandre Rafalovitch and Robert Dale", year = "2009", language = "English", booktitle = "MT Summit XII proceedings", publisher = "International Association of Machine Translation", } ### Contributions Thanks to @param087 for adding this dataset.
[ "# Dataset Card for [Dataset Name]", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper: URL\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\nThis is a collection of translated documents from the United Nations originally compiled into a translation memory by Alexandre Rafalovitch, Robert Dale (see URL).", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\n\n\n\n\n@inproceedings{title = \"United Nations General Assembly Resolutions: a six-language parallel corpus\",\nabstract = \"In this paper we describe a six-ways parallel public-domain corpus consisting of 2100 United Nations General Assembly Resolutions with translations in the six official languages of the United Nations, with an average of around 3 million tokens per language. The corpus is available in a preprocessed, formatting-normalized TMX format with paragraphs aligned across multiple languages. We describe the background to the corpus and its content, the process of its construction, and some of its interesting properties.\",\nauthor = \"Alexandre Rafalovitch and Robert Dale\",\nyear = \"2009\",\nlanguage = \"English\",\nbooktitle = \"MT Summit XII proceedings\",\npublisher = \"International Association of Machine Translation\",\n}", "### Contributions\n\nThanks to @param087 for adding this dataset." ]
[ "TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-translation #size_categories-10K<n<100K #source_datasets-original #language-Arabic #language-English #language-Spanish #language-French #language-Russian #language-Chinese #license-unknown #region-us \n", "# Dataset Card for [Dataset Name]", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper: URL\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\nThis is a collection of translated documents from the United Nations originally compiled into a translation memory by Alexandre Rafalovitch, Robert Dale (see URL).", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\n\n\n\n\n@inproceedings{title = \"United Nations General Assembly Resolutions: a six-language parallel corpus\",\nabstract = \"In this paper we describe a six-ways parallel public-domain corpus consisting of 2100 United Nations General Assembly Resolutions with translations in the six official languages of the United Nations, with an average of around 3 million tokens per language. The corpus is available in a preprocessed, formatting-normalized TMX format with paragraphs aligned across multiple languages. We describe the background to the corpus and its content, the process of its construction, and some of its interesting properties.\",\nauthor = \"Alexandre Rafalovitch and Robert Dale\",\nyear = \"2009\",\nlanguage = \"English\",\nbooktitle = \"MT Summit XII proceedings\",\npublisher = \"International Association of Machine Translation\",\n}", "### Contributions\n\nThanks to @param087 for adding this dataset." ]
[ 97, 10, 120, 26, 42, 10, 4, 6, 6, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 184, 18 ]
[ "passage: TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-translation #size_categories-10K<n<100K #source_datasets-original #language-Arabic #language-English #language-Spanish #language-French #language-Russian #language-Chinese #license-unknown #region-us \n# Dataset Card for [Dataset Name]## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper: URL\n- Leaderboard:\n- Point of Contact:### Dataset Summary\nThis is a collection of translated documents from the United Nations originally compiled into a translation memory by Alexandre Rafalovitch, Robert Dale (see URL).### Supported Tasks and Leaderboards### Languages## Dataset Structure### Data Instances### Data Fields### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators" ]
5131760d55d28eab51145989738003b811ea2113
# Dataset Card for MultiUN ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://opus.nlpl.eu/MultiUN/corpus/version/MultiUN - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** https://aclanthology.org/L10-1473/ - **Leaderboard:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Dataset Summary The MultiUN parallel corpus is extracted from the United Nations Website , and then cleaned and converted to XML at Language Technology Lab in DFKI GmbH (LT-DFKI), Germany. The documents were published by UN from 2000 to 2009. This is a collection of translated documents from the United Nations originally compiled by Andreas Eisele and Yu Chen (see http://www.euromatrixplus.net/multi-un/). This corpus is available in all 6 official languages of the UN consisting of around 300 million words per language ### Supported Tasks and Leaderboards The underlying task is machine translation. ### Languages Parallel texts are present in all six official languages, namely Arabic, Chinese, English, French, Russian and Spanish, with a small part of the documents available also in German. ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization Original MultiUN source data: http://www.euromatrixplus.net/multi-unp #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information If you use this corpus in your work, please cite the paper: ``` @inproceedings{eisele-chen-2010-multiun, title = "{M}ulti{UN}: A Multilingual Corpus from United Nation Documents", author = "Eisele, Andreas and Chen, Yu", booktitle = "Proceedings of the Seventh International Conference on Language Resources and Evaluation ({LREC}'10)", month = may, year = "2010", address = "Valletta, Malta", publisher = "European Language Resources Association (ELRA)", url = "http://www.lrec-conf.org/proceedings/lrec2010/pdf/686_Paper.pdf", abstract = "This paper describes the acquisition, preparation and properties of a corpus extracted from the official documents of the United Nations (UN). This corpus is available in all 6 official languages of the UN, consisting of around 300 million words per language. We describe the methods we used for crawling, document formatting, and sentence alignment. This corpus also includes a common test set for machine translation. We present the results of a French-Chinese machine translation experiment performed on this corpus.", } ``` If you use any part of the corpus (hosted in OPUS) in your own work, please cite the following article: ``` @inproceedings{tiedemann-2012-parallel, title = "Parallel Data, Tools and Interfaces in {OPUS}", author = {Tiedemann, J{\"o}rg}, editor = "Calzolari, Nicoletta and Choukri, Khalid and Declerck, Thierry and Do{\u{g}}an, Mehmet U{\u{g}}ur and Maegaard, Bente and Mariani, Joseph and Moreno, Asuncion and Odijk, Jan and Piperidis, Stelios", booktitle = "Proceedings of the Eighth International Conference on Language Resources and Evaluation ({LREC}'12)", month = may, year = "2012", address = "Istanbul, Turkey", publisher = "European Language Resources Association (ELRA)", url = "http://www.lrec-conf.org/proceedings/lrec2012/pdf/463_Paper.pdf", pages = "2214--2218", abstract = "This paper presents the current status of OPUS, a growing language resource of parallel corpora and related tools. The focus in OPUS is to provide freely available data sets in various formats together with basic annotation to be useful for applications in computational linguistics, translation studies and cross-linguistic corpus studies. In this paper, we report about new data sets and their features, additional annotation tools and models provided from the website and essential interfaces and on-line services included in the project.", } ``` ### Contributions Thanks to [@patil-suraj](https://github.com/patil-suraj) for adding this dataset.
un_multi
[ "task_categories:translation", "annotations_creators:found", "language_creators:found", "multilinguality:multilingual", "size_categories:100K<n<1M", "source_datasets:original", "language:ar", "language:de", "language:en", "language:es", "language:fr", "language:ru", "language:zh", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["ar", "de", "en", "es", "fr", "ru", "zh"], "license": ["unknown"], "multilinguality": ["multilingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["translation"], "task_ids": [], "paperswithcode_id": "multiun", "pretty_name": "Multilingual Corpus from United Nation Documents", "config_names": ["ar-de", "ar-en", "ar-es", "ar-fr", "ar-ru", "ar-zh", "de-en", "de-es", "de-fr", "de-ru", "de-zh", "en-es", "en-fr", "en-ru", "en-zh", "es-fr", "es-ru", "es-zh", "fr-ru", "fr-zh", "ru-zh"], "dataset_info": [{"config_name": "ar-de", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ar", "de"]}}}], "splits": [{"name": "train", "num_bytes": 94466397, "num_examples": 165090}], "download_size": 21869935, "dataset_size": 94466397}, {"config_name": "ar-en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ar", "en"]}}}], "splits": [{"name": "train", "num_bytes": 4189852369, "num_examples": 9759125}], "download_size": 1036296368, "dataset_size": 4189852369}, {"config_name": "ar-es", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ar", "es"]}}}], "splits": [{"name": "train", "num_bytes": 4509675284, "num_examples": 10119379}], "download_size": 1101206667, "dataset_size": 4509675284}, {"config_name": "ar-fr", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ar", "fr"]}}}], "splits": [{"name": "train", "num_bytes": 4516850009, "num_examples": 9929567}], "download_size": 1109705925, "dataset_size": 4516850009}, {"config_name": "ar-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ar", "ru"]}}}], "splits": [{"name": "train", "num_bytes": 5932866867, "num_examples": 10206243}], "download_size": 1261123878, "dataset_size": 5932866867}, {"config_name": "ar-zh", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ar", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 3781658413, "num_examples": 9832293}], "download_size": 1009696775, "dataset_size": 3781658413}, {"config_name": "de-en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["de", "en"]}}}], "splits": [{"name": "train", "num_bytes": 76684549, "num_examples": 162981}], "download_size": 19468529, "dataset_size": 76684549}, {"config_name": "de-es", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["de", "es"]}}}], "splits": [{"name": "train", "num_bytes": 80936653, "num_examples": 162078}], "download_size": 20266591, "dataset_size": 80936653}, {"config_name": "de-fr", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["de", "fr"]}}}], "splits": [{"name": "train", "num_bytes": 81888435, "num_examples": 164025}], "download_size": 20692837, "dataset_size": 81888435}, {"config_name": "de-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["de", "ru"]}}}], "splits": [{"name": "train", "num_bytes": 111517934, "num_examples": 164792}], "download_size": 23507789, "dataset_size": 111517934}, {"config_name": "de-zh", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["de", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 70534818, "num_examples": 176933}], "download_size": 19927209, "dataset_size": 70534818}, {"config_name": "en-es", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "es"]}}}], "splits": [{"name": "train", "num_bytes": 4128141663, "num_examples": 11350967}], "download_size": 1123164180, "dataset_size": 4128141663}, {"config_name": "en-fr", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "fr"]}}}], "splits": [{"name": "train", "num_bytes": 4678055160, "num_examples": 13172019}], "download_size": 1355002731, "dataset_size": 4678055160}, {"config_name": "en-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "ru"]}}}], "splits": [{"name": "train", "num_bytes": 5632662839, "num_examples": 11654416}], "download_size": 1285801078, "dataset_size": 5632662839}, {"config_name": "en-zh", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 2960376046, "num_examples": 9564315}], "download_size": 900076520, "dataset_size": 2960376046}, {"config_name": "es-fr", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["es", "fr"]}}}], "splits": [{"name": "train", "num_bytes": 4454712498, "num_examples": 11441889}], "download_size": 1195733510, "dataset_size": 4454712498}, {"config_name": "es-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["es", "ru"]}}}], "splits": [{"name": "train", "num_bytes": 5442655730, "num_examples": 10605056}], "download_size": 1228045966, "dataset_size": 5442655730}, {"config_name": "es-zh", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["es", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 3223871198, "num_examples": 9847770}], "download_size": 953250084, "dataset_size": 3223871198}, {"config_name": "fr-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["fr", "ru"]}}}], "splits": [{"name": "train", "num_bytes": 5979879089, "num_examples": 11761738}], "download_size": 1364307157, "dataset_size": 5979879089}, {"config_name": "fr-zh", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["fr", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 3241098333, "num_examples": 9690914}], "download_size": 962824881, "dataset_size": 3241098333}, {"config_name": "ru-zh", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ru", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 4233875537, "num_examples": 9557007}], "download_size": 1037881127, "dataset_size": 4233875537}]}
2024-02-13T17:16:38+00:00
[]
[ "ar", "de", "en", "es", "fr", "ru", "zh" ]
TAGS #task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-100K<n<1M #source_datasets-original #language-Arabic #language-German #language-English #language-Spanish #language-French #language-Russian #language-Chinese #license-unknown #region-us
# Dataset Card for MultiUN ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: URL - Leaderboard: - Point of Contact: ### Dataset Summary The MultiUN parallel corpus is extracted from the United Nations Website , and then cleaned and converted to XML at Language Technology Lab in DFKI GmbH (LT-DFKI), Germany. The documents were published by UN from 2000 to 2009. This is a collection of translated documents from the United Nations originally compiled by Andreas Eisele and Yu Chen (see URL This corpus is available in all 6 official languages of the UN consisting of around 300 million words per language ### Supported Tasks and Leaderboards The underlying task is machine translation. ### Languages Parallel texts are present in all six official languages, namely Arabic, Chinese, English, French, Russian and Spanish, with a small part of the documents available also in German. ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization Original MultiUN source data: URL #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information If you use this corpus in your work, please cite the paper: If you use any part of the corpus (hosted in OPUS) in your own work, please cite the following article: ### Contributions Thanks to @patil-suraj for adding this dataset.
[ "# Dataset Card for MultiUN", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: URL\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\nThe MultiUN parallel corpus is extracted from the United Nations Website , and then cleaned and converted to XML at Language Technology Lab in DFKI GmbH (LT-DFKI), Germany. The documents were published by UN from 2000 to 2009.\n\nThis is a collection of translated documents from the United Nations originally compiled by Andreas Eisele and Yu Chen (see URL\n\nThis corpus is available in all 6 official languages of the UN consisting of around 300 million words per language", "### Supported Tasks and Leaderboards\n\nThe underlying task is machine translation.", "### Languages\n\nParallel texts are present in all six official languages, namely Arabic, Chinese,\nEnglish, French, Russian and Spanish, with a small part of the documents available also in German.", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization\n\nOriginal MultiUN source data: URL", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\n\n\n\n\nIf you use this corpus in your work, please cite the paper:\n\n\nIf you use any part of the corpus (hosted in OPUS) in your own work, please cite the following article:", "### Contributions\n\nThanks to @patil-suraj for adding this dataset." ]
[ "TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-100K<n<1M #source_datasets-original #language-Arabic #language-German #language-English #language-Spanish #language-French #language-Russian #language-Chinese #license-unknown #region-us \n", "# Dataset Card for MultiUN", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: URL\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\nThe MultiUN parallel corpus is extracted from the United Nations Website , and then cleaned and converted to XML at Language Technology Lab in DFKI GmbH (LT-DFKI), Germany. The documents were published by UN from 2000 to 2009.\n\nThis is a collection of translated documents from the United Nations originally compiled by Andreas Eisele and Yu Chen (see URL\n\nThis corpus is available in all 6 official languages of the UN consisting of around 300 million words per language", "### Supported Tasks and Leaderboards\n\nThe underlying task is machine translation.", "### Languages\n\nParallel texts are present in all six official languages, namely Arabic, Chinese,\nEnglish, French, Russian and Spanish, with a small part of the documents available also in German.", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization\n\nOriginal MultiUN source data: URL", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\n\n\n\n\nIf you use this corpus in your work, please cite the paper:\n\n\nIf you use any part of the corpus (hosted in OPUS) in your own work, please cite the following article:", "### Contributions\n\nThanks to @patil-suraj for adding this dataset." ]
[ 102, 7, 120, 26, 106, 19, 42, 6, 6, 5, 5, 5, 7, 4, 17, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 46, 19 ]
[ "passage: TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-100K<n<1M #source_datasets-original #language-Arabic #language-German #language-English #language-Spanish #language-French #language-Russian #language-Chinese #license-unknown #region-us \n# Dataset Card for MultiUN## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: URL\n- Leaderboard: \n- Point of Contact:### Dataset Summary\n\nThe MultiUN parallel corpus is extracted from the United Nations Website , and then cleaned and converted to XML at Language Technology Lab in DFKI GmbH (LT-DFKI), Germany. The documents were published by UN from 2000 to 2009.\n\nThis is a collection of translated documents from the United Nations originally compiled by Andreas Eisele and Yu Chen (see URL\n\nThis corpus is available in all 6 official languages of the UN consisting of around 300 million words per language### Supported Tasks and Leaderboards\n\nThe underlying task is machine translation.### Languages\n\nParallel texts are present in all six official languages, namely Arabic, Chinese,\nEnglish, French, Russian and Spanish, with a small part of the documents available also in German.## Dataset Structure### Data Instances### Data Fields### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization\n\nOriginal MultiUN source data: URL#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?" ]
5db1519445ed7c07ad85b038b75289c8ad96c7bc
# Dataset Card for United Nations Parallel Corpus ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://opus.nlpl.eu/UNPC/corpus/version/UNPC - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** https://aclanthology.org/L16-1561/ - **Leaderboard:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Dataset Summary The United Nations Parallel Corpus is the first parallel corpus composed from United Nations documents published by the original data creator. The parallel corpus consists of manually translated UN documents from the last 25 years (1990 to 2014) for the six official UN languages, Arabic, Chinese, English, French, Russian, and Spanish. The corpus is freely available for download under a liberal license. ### Supported Tasks and Leaderboards The underlying task is machine translation. ### Languages The six official UN languages: Arabic, Chinese, English, French, Russian, and Spanish. ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information https://conferences.unite.un.org/UNCORPUS/#disclaimer The following disclaimer, an integral part of the United Nations Parallel Corpus, shall be respected with regard to the Corpus (no other restrictions apply): - The United Nations Parallel Corpus is made available without warranty of any kind, explicit or implied. The United Nations specifically makes no warranties or representations as to the accuracy or completeness of the information contained in the United Nations Corpus. - Under no circumstances shall the United Nations be liable for any loss, liability, injury or damage incurred or suffered that is claimed to have resulted from the use of the United Nations Corpus. The use of the United Nations Corpus is at the user's sole risk. The user specifically acknowledges and agrees that the United Nations is not liable for the conduct of any user. If the user is dissatisfied with any of the material provided in the United Nations Corpus, the user's sole and exclusive remedy is to discontinue using the United Nations Corpus. - When using the United Nations Corpus, the user must acknowledge the United Nations as the source of the information. For references, please cite this reference: Ziemski, M., Junczys-Dowmunt, M., and Pouliquen, B., (2016), The United Nations Parallel Corpus, Language Resources and Evaluation (LREC’16), Portorož, Slovenia, May 2016. - Nothing herein shall constitute or be considered to be a limitation upon or waiver, express or implied, of the privileges and immunities of the United Nations, which are specifically reserved. ### Citation Information ``` @inproceedings{ziemski-etal-2016-united, title = "The {U}nited {N}ations Parallel Corpus v1.0", author = "Ziemski, Micha{\\l} and Junczys-Dowmunt, Marcin and Pouliquen, Bruno", booktitle = "Proceedings of the Tenth International Conference on Language Resources and Evaluation ({LREC}'16)", month = may, year = "2016", address = "Portoro{\v{z}}, Slovenia", publisher = "European Language Resources Association (ELRA)", url = "https://www.aclweb.org/anthology/L16-1561", pages = "3530--3534", abstract = "This paper describes the creation process and statistics of the official United Nations Parallel Corpus, the first parallel corpus composed from United Nations documents published by the original data creator. The parallel corpus presented consists of manually translated UN documents from the last 25 years (1990 to 2014) for the six official UN languages, Arabic, Chinese, English, French, Russian, and Spanish. The corpus is freely available for download under a liberal license. Apart from the pairwise aligned documents, a fully aligned subcorpus for the six official UN languages is distributed. We provide baseline BLEU scores of our Moses-based SMT systems trained with the full data of language pairs involving English and for all possible translation directions of the six-way subcorpus.", } ``` ### Contributions Thanks to [@patil-suraj](https://github.com/patil-suraj) for adding this dataset.
un_pc
[ "task_categories:translation", "annotations_creators:found", "language_creators:found", "multilinguality:multilingual", "size_categories:10M<n<100M", "source_datasets:original", "language:ar", "language:en", "language:es", "language:fr", "language:ru", "language:zh", "license:other", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["ar", "en", "es", "fr", "ru", "zh"], "license": "other", "multilinguality": ["multilingual"], "size_categories": ["10M<n<100M"], "source_datasets": ["original"], "task_categories": ["translation"], "task_ids": [], "paperswithcode_id": "united-nations-parallel-corpus", "pretty_name": "United Nations Parallel Corpus", "config_names": ["ar-en", "ar-es", "ar-fr", "ar-ru", "ar-zh", "en-es", "en-fr", "en-ru", "en-zh", "es-fr", "es-ru", "es-zh", "fr-ru", "fr-zh", "ru-zh"], "dataset_info": [{"config_name": "ar-en", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ar", "en"]}}}], "splits": [{"name": "train", "num_bytes": 8039689939, "num_examples": 20044478}], "download_size": 2025106743, "dataset_size": 8039689939}, {"config_name": "ar-es", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ar", "es"]}}}], "splits": [{"name": "train", "num_bytes": 8715754848, "num_examples": 20532014}], "download_size": 2167791297, "dataset_size": 8715754848}, {"config_name": "ar-fr", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ar", "fr"]}}}], "splits": [{"name": "train", "num_bytes": 8897848038, "num_examples": 20281645}], "download_size": 2188765415, "dataset_size": 8897848038}, {"config_name": "ar-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ar", "ru"]}}}], "splits": [{"name": "train", "num_bytes": 11395923083, "num_examples": 20571334}], "download_size": 2476562835, "dataset_size": 11395923083}, {"config_name": "ar-zh", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ar", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 6447658008, "num_examples": 17306056}], "download_size": 1738869755, "dataset_size": 6447658008}, {"config_name": "en-es", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "es"]}}}], "splits": [{"name": "train", "num_bytes": 8241635322, "num_examples": 25227004}], "download_size": 2300411698, "dataset_size": 8241635322}, {"config_name": "en-fr", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "fr"]}}}], "splits": [{"name": "train", "num_bytes": 9718522775, "num_examples": 30340652}], "download_size": 2657208676, "dataset_size": 9718522775}, {"config_name": "en-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "ru"]}}}], "splits": [{"name": "train", "num_bytes": 11156164691, "num_examples": 25173398}], "download_size": 2589707636, "dataset_size": 11156164691}, {"config_name": "en-zh", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["en", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 4988812558, "num_examples": 17451549}], "download_size": 1535707641, "dataset_size": 4988812558}, {"config_name": "es-fr", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["es", "fr"]}}}], "splits": [{"name": "train", "num_bytes": 9230891207, "num_examples": 25887160}], "download_size": 2492342915, "dataset_size": 9230891207}, {"config_name": "es-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["es", "ru"]}}}], "splits": [{"name": "train", "num_bytes": 10789780134, "num_examples": 22294106}], "download_size": 2487664520, "dataset_size": 10789780134}, {"config_name": "es-zh", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["es", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 5475365986, "num_examples": 17599223}], "download_size": 1639717723, "dataset_size": 5475365986}, {"config_name": "fr-ru", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["fr", "ru"]}}}], "splits": [{"name": "train", "num_bytes": 12099669711, "num_examples": 25219973}], "download_size": 2762585269, "dataset_size": 12099669711}, {"config_name": "fr-zh", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["fr", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 5679222134, "num_examples": 17521170}], "download_size": 1668823634, "dataset_size": 5679222134}, {"config_name": "ru-zh", "features": [{"name": "translation", "dtype": {"translation": {"languages": ["ru", "zh"]}}}], "splits": [{"name": "train", "num_bytes": 7905443441, "num_examples": 17920922}], "download_size": 1934425373, "dataset_size": 7905443441}]}
2024-02-13T11:20:10+00:00
[]
[ "ar", "en", "es", "fr", "ru", "zh" ]
TAGS #task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-10M<n<100M #source_datasets-original #language-Arabic #language-English #language-Spanish #language-French #language-Russian #language-Chinese #license-other #region-us
# Dataset Card for United Nations Parallel Corpus ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: URL - Leaderboard: - Point of Contact: ### Dataset Summary The United Nations Parallel Corpus is the first parallel corpus composed from United Nations documents published by the original data creator. The parallel corpus consists of manually translated UN documents from the last 25 years (1990 to 2014) for the six official UN languages, Arabic, Chinese, English, French, Russian, and Spanish. The corpus is freely available for download under a liberal license. ### Supported Tasks and Leaderboards The underlying task is machine translation. ### Languages The six official UN languages: Arabic, Chinese, English, French, Russian, and Spanish. ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information URL The following disclaimer, an integral part of the United Nations Parallel Corpus, shall be respected with regard to the Corpus (no other restrictions apply): - The United Nations Parallel Corpus is made available without warranty of any kind, explicit or implied. The United Nations specifically makes no warranties or representations as to the accuracy or completeness of the information contained in the United Nations Corpus. - Under no circumstances shall the United Nations be liable for any loss, liability, injury or damage incurred or suffered that is claimed to have resulted from the use of the United Nations Corpus. The use of the United Nations Corpus is at the user's sole risk. The user specifically acknowledges and agrees that the United Nations is not liable for the conduct of any user. If the user is dissatisfied with any of the material provided in the United Nations Corpus, the user's sole and exclusive remedy is to discontinue using the United Nations Corpus. - When using the United Nations Corpus, the user must acknowledge the United Nations as the source of the information. For references, please cite this reference: Ziemski, M., Junczys-Dowmunt, M., and Pouliquen, B., (2016), The United Nations Parallel Corpus, Language Resources and Evaluation (LREC’16), Portorož, Slovenia, May 2016. - Nothing herein shall constitute or be considered to be a limitation upon or waiver, express or implied, of the privileges and immunities of the United Nations, which are specifically reserved. ### Contributions Thanks to @patil-suraj for adding this dataset.
[ "# Dataset Card for United Nations Parallel Corpus", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: URL\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\nThe United Nations Parallel Corpus is the first parallel corpus composed from United Nations documents published by the original data creator. \nThe parallel corpus consists of manually translated UN documents from the last 25 years (1990 to 2014)\nfor the six official UN languages, Arabic, Chinese, English, French, Russian, and Spanish.\nThe corpus is freely available for download under a liberal license.", "### Supported Tasks and Leaderboards\n\nThe underlying task is machine translation.", "### Languages\n\nThe six official UN languages: Arabic, Chinese, English, French, Russian, and Spanish.", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nURL\n\nThe following disclaimer, an integral part of the United Nations Parallel Corpus, shall be respected with regard to the Corpus (no other restrictions apply):\n- The United Nations Parallel Corpus is made available without warranty of any kind, explicit or implied. The United Nations specifically makes no warranties or representations as to the accuracy or completeness of the information contained in the United Nations Corpus.\n- Under no circumstances shall the United Nations be liable for any loss, liability, injury or damage incurred or suffered that is claimed to have resulted from the use of the United Nations Corpus. The use of the United Nations Corpus is at the user's sole risk. The user specifically acknowledges and agrees that the United Nations is not liable for the conduct of any user. If the user is dissatisfied with any of the material provided in the United Nations Corpus, the user's sole and exclusive remedy is to discontinue using the United Nations Corpus.\n- When using the United Nations Corpus, the user must acknowledge the United Nations as the source of the information. For references, please cite this reference: Ziemski, M., Junczys-Dowmunt, M., and Pouliquen, B., (2016), The United Nations Parallel Corpus, Language Resources and Evaluation (LREC’16), Portorož, Slovenia, May 2016.\n- Nothing herein shall constitute or be considered to be a limitation upon or waiver, express or implied, of the privileges and immunities of the United Nations, which are specifically reserved.", "### Contributions\n\nThanks to @patil-suraj for adding this dataset." ]
[ "TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-10M<n<100M #source_datasets-original #language-Arabic #language-English #language-Spanish #language-French #language-Russian #language-Chinese #license-other #region-us \n", "# Dataset Card for United Nations Parallel Corpus", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: URL\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary\n\nThe United Nations Parallel Corpus is the first parallel corpus composed from United Nations documents published by the original data creator. \nThe parallel corpus consists of manually translated UN documents from the last 25 years (1990 to 2014)\nfor the six official UN languages, Arabic, Chinese, English, French, Russian, and Spanish.\nThe corpus is freely available for download under a liberal license.", "### Supported Tasks and Leaderboards\n\nThe underlying task is machine translation.", "### Languages\n\nThe six official UN languages: Arabic, Chinese, English, French, Russian, and Spanish.", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nURL\n\nThe following disclaimer, an integral part of the United Nations Parallel Corpus, shall be respected with regard to the Corpus (no other restrictions apply):\n- The United Nations Parallel Corpus is made available without warranty of any kind, explicit or implied. The United Nations specifically makes no warranties or representations as to the accuracy or completeness of the information contained in the United Nations Corpus.\n- Under no circumstances shall the United Nations be liable for any loss, liability, injury or damage incurred or suffered that is claimed to have resulted from the use of the United Nations Corpus. The use of the United Nations Corpus is at the user's sole risk. The user specifically acknowledges and agrees that the United Nations is not liable for the conduct of any user. If the user is dissatisfied with any of the material provided in the United Nations Corpus, the user's sole and exclusive remedy is to discontinue using the United Nations Corpus.\n- When using the United Nations Corpus, the user must acknowledge the United Nations as the source of the information. For references, please cite this reference: Ziemski, M., Junczys-Dowmunt, M., and Pouliquen, B., (2016), The United Nations Parallel Corpus, Language Resources and Evaluation (LREC’16), Portorož, Slovenia, May 2016.\n- Nothing herein shall constitute or be considered to be a limitation upon or waiver, express or implied, of the privileges and immunities of the United Nations, which are specifically reserved.", "### Contributions\n\nThanks to @patil-suraj for adding this dataset." ]
[ 96, 9, 120, 26, 85, 19, 24, 6, 6, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 340, 19 ]
[ "passage: TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-10M<n<100M #source_datasets-original #language-Arabic #language-English #language-Spanish #language-French #language-Russian #language-Chinese #license-other #region-us \n# Dataset Card for United Nations Parallel Corpus## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: URL\n- Leaderboard: \n- Point of Contact:### Dataset Summary\n\nThe United Nations Parallel Corpus is the first parallel corpus composed from United Nations documents published by the original data creator. \nThe parallel corpus consists of manually translated UN documents from the last 25 years (1990 to 2014)\nfor the six official UN languages, Arabic, Chinese, English, French, Russian, and Spanish.\nThe corpus is freely available for download under a liberal license.### Supported Tasks and Leaderboards\n\nThe underlying task is machine translation.### Languages\n\nThe six official UN languages: Arabic, Chinese, English, French, Russian, and Spanish.## Dataset Structure### Data Instances### Data Fields### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators" ]
c85d2c32c39177a14ba02b1e4693f7f31f7bb975
# Dataset Card for Universal Dependencies Treebank ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Universal Dependencies](https://universaldependencies.org/) - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@patrickvonplaten](https://github.com/patrickvonplaten), [@jplu](https://github.com/jplu) for adding this dataset.
universal_dependencies
[ "task_categories:token-classification", "task_ids:parsing", "annotations_creators:expert-generated", "language_creators:crowdsourced", "multilinguality:multilingual", "size_categories:1K<n<10K", "source_datasets:original", "language:af", "language:aii", "language:ajp", "language:akk", "language:am", "language:apu", "language:aqz", "language:ar", "language:be", "language:bg", "language:bho", "language:bm", "language:br", "language:bxr", "language:ca", "language:ckt", "language:cop", "language:cs", "language:cu", "language:cy", "language:da", "language:de", "language:el", "language:en", "language:es", "language:et", "language:eu", "language:fa", "language:fi", "language:fo", "language:fr", "language:fro", "language:ga", "language:gd", "language:gl", "language:got", "language:grc", "language:gsw", "language:gun", "language:gv", "language:he", "language:hi", "language:hr", "language:hsb", "language:hu", "language:hy", "language:id", "language:is", "language:it", "language:ja", "language:kfm", "language:kk", "language:kmr", "language:ko", "language:koi", "language:kpv", "language:krl", "language:la", "language:lt", "language:lv", "language:lzh", "language:mdf", "language:mr", "language:mt", "language:myu", "language:myv", "language:nl", "language:no", "language:nyq", "language:olo", "language:orv", "language:otk", "language:pcm", "language:pl", "language:pt", "language:ro", "language:ru", "language:sa", "language:sk", "language:sl", "language:sme", "language:sms", "language:soj", "language:sq", "language:sr", "language:sv", "language:swl", "language:ta", "language:te", "language:th", "language:tl", "language:tpn", "language:tr", "language:ug", "language:uk", "language:ur", "language:vi", "language:wbp", "language:wo", "language:yo", "language:yue", "language:zh", "license:unknown", "constituency-parsing", "dependency-parsing", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["crowdsourced"], "language": ["af", "aii", "ajp", "akk", "am", "apu", "aqz", "ar", "be", "bg", "bho", "bm", "br", "bxr", "ca", "ckt", "cop", "cs", "cu", "cy", "da", "de", "el", "en", "es", "et", "eu", "fa", "fi", "fo", "fr", "fro", "ga", "gd", "gl", "got", "grc", "gsw", "gun", "gv", "he", "hi", "hr", "hsb", "hu", "hy", "id", "is", "it", "ja", "kfm", "kk", "kmr", "ko", "koi", "kpv", "krl", "la", "lt", "lv", "lzh", "mdf", "mr", "mt", "myu", "myv", "nl", "no", "nyq", "olo", "orv", "otk", "pcm", "pl", "pt", "ro", "ru", "sa", "sk", "sl", "sme", "sms", "soj", "sq", "sr", "sv", "swl", "ta", "te", "th", "tl", "tpn", "tr", "ug", "uk", "ur", "vi", "wbp", "wo", "yo", "yue", "zh"], "license": ["unknown"], "multilinguality": ["multilingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["token-classification"], "task_ids": ["parsing"], "paperswithcode_id": "universal-dependencies", "pretty_name": "Universal Dependencies Treebank", "config_names": ["af_afribooms", "aii_as", "ajp_madar", "akk_pisandub", "akk_riao", "am_att", "apu_ufpa", "aqz_tudet", "ar_nyuad", "ar_padt", "ar_pud", "be_hse", "bg_btb", "bho_bhtb", "bm_crb", "br_keb", "bxr_bdt", "ca_ancora", "ckt_hse", "cop_scriptorium", "cs_cac", "cs_cltt", "cs_fictree", "cs_pdt", "cs_pud", "cu_proiel", "cy_ccg", "da_ddt", "de_gsd", "de_hdt", "de_lit", "de_pud", "el_gdt", "en_esl", "en_ewt", "en_gum", "en_gumreddit", "en_lines", "en_partut", "en_pronouns", "en_pud", "es_ancora", "es_gsd", "es_pud", "et_edt", "et_ewt", "eu_bdt", "fa_perdt", "fa_seraji", "fi_ftb", "fi_ood", "fi_pud", "fi_tdt", "fo_farpahc", "fo_oft", "fr_fqb", "fr_ftb", "fr_gsd", "fr_partut", "fr_pud", "fr_sequoia", "fr_spoken", "fro_srcmf", "ga_idt", "gd_arcosg", "gl_ctg", "gl_treegal", "got_proiel", "grc_perseus", "grc_proiel", "gsw_uzh", "gun_dooley", "gun_thomas", "gv_cadhan", "he_htb", "hi_hdtb", "hi_pud", "hr_set", "hsb_ufal", "hu_szeged", "hy_armtdp", "id_csui", "id_gsd", "id_pud", "is_icepahc", "is_pud", "it_isdt", "it_partut", "it_postwita", "it_pud", "it_twittiro", "it_vit", "ja_bccwj", "ja_gsd", "ja_modern", "ja_pud", "kfm_aha", "kk_ktb", "kmr_mg", "ko_gsd", "ko_kaist", "ko_pud", "koi_uh", "kpv_ikdp", "kpv_lattice", "krl_kkpp", "la_ittb", "la_llct", "la_perseus", "la_proiel", "lt_alksnis", "lt_hse", "lv_lvtb", "lzh_kyoto", "mdf_jr", "mr_ufal", "mt_mudt", "myu_tudet", "myv_jr", "nl_alpino", "nl_lassysmall", "no_bokmaal", "no_nynorsk", "no_nynorsklia", "nyq_aha", "olo_kkpp", "orv_rnc", "orv_torot", "otk_tonqq", "pcm_nsc", "pl_lfg", "pl_pdb", "pl_pud", "pt_bosque", "pt_gsd", "pt_pud", "qhe_hiencs", "qtd_sagt", "ro_nonstandard", "ro_rrt", "ro_simonero", "ru_gsd", "ru_pud", "ru_syntagrus", "ru_taiga", "sa_ufal", "sa_vedic", "sk_snk", "sl_ssj", "sl_sst", "sme_giella", "sms_giellagas", "soj_aha", "sq_tsa", "sr_set", "sv_lines", "sv_pud", "sv_talbanken", "swl_sslc", "ta_mwtt", "ta_ttb", "te_mtg", "th_pud", "tl_trg", "tl_ugnayan", "tpn_tudet", "tr_boun", "tr_gb", "tr_imst", "tr_pud", "ug_udt", "uk_iu", "ur_udtb", "vi_vtb", "wbp_ufal", "wo_wtb", "yo_ytb", "yue_hk", "zh_cfl", "zh_gsd", "zh_gsdsimp", "zh_hk", "zh_pud"], "tags": ["constituency-parsing", "dependency-parsing"], "dataset_info": [{"config_name": "af_afribooms", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 3523113, "num_examples": 1315}, {"name": "validation", "num_bytes": 547285, "num_examples": 194}, {"name": "test", "num_bytes": 1050299, "num_examples": 425}], "download_size": 3088237, "dataset_size": 5120697}, {"config_name": "akk_pisandub", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 153470, "num_examples": 101}], "download_size": 101789, "dataset_size": 153470}, {"config_name": "akk_riao", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 3374577, "num_examples": 1804}], "download_size": 2022357, "dataset_size": 3374577}, {"config_name": "aqz_tudet", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 8286, "num_examples": 24}], "download_size": 5683, "dataset_size": 8286}, {"config_name": "sq_tsa", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 116034, "num_examples": 60}], "download_size": 68875, "dataset_size": 116034}, {"config_name": "am_att", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 1554859, "num_examples": 1074}], "download_size": 1019607, "dataset_size": 1554859}, {"config_name": "grc_perseus", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 22611612, "num_examples": 11476}, {"name": "validation", "num_bytes": 3152233, "num_examples": 1137}, {"name": "test", "num_bytes": 3004502, "num_examples": 1306}], "download_size": 18898313, "dataset_size": 28768347}, {"config_name": "grc_proiel", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 30938089, "num_examples": 15014}, {"name": "validation", "num_bytes": 2264551, "num_examples": 1019}, {"name": "test", "num_bytes": 2192289, "num_examples": 1047}], "download_size": 23715831, "dataset_size": 35394929}, {"config_name": "apu_ufpa", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 75578, "num_examples": 76}], "download_size": 69565, "dataset_size": 75578}, {"config_name": "ar_nyuad", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 79064476, "num_examples": 15789}, {"name": "validation", "num_bytes": 9859912, "num_examples": 1986}, {"name": "test", "num_bytes": 9880240, "num_examples": 1963}], "download_size": 58583673, "dataset_size": 98804628}, {"config_name": "ar_padt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 58537298, "num_examples": 6075}, {"name": "validation", "num_bytes": 7787253, "num_examples": 909}, {"name": "test", "num_bytes": 7428063, "num_examples": 680}], "download_size": 51208169, "dataset_size": 73752614}, {"config_name": "ar_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2816625, "num_examples": 1000}], "download_size": 2084082, "dataset_size": 2816625}, {"config_name": "hy_armtdp", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 7697891, "num_examples": 1975}, {"name": "validation", "num_bytes": 988849, "num_examples": 249}, {"name": "test", "num_bytes": 947287, "num_examples": 278}], "download_size": 6886567, "dataset_size": 9634027}, {"config_name": "aii_as", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 52540, "num_examples": 57}], "download_size": 32639, "dataset_size": 52540}, {"config_name": "bm_crb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 1502886, "num_examples": 1026}], "download_size": 892924, "dataset_size": 1502886}, {"config_name": "eu_bdt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 8199861, "num_examples": 5396}, {"name": "validation", "num_bytes": 2701073, "num_examples": 1798}, {"name": "test", "num_bytes": 2734601, "num_examples": 1799}], "download_size": 8213576, "dataset_size": 13635535}, {"config_name": "be_hse", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 34880663, "num_examples": 21555}, {"name": "validation", "num_bytes": 1745668, "num_examples": 1090}, {"name": "test", "num_bytes": 1818113, "num_examples": 889}], "download_size": 26433402, "dataset_size": 38444444}, {"config_name": "bho_bhtb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 947740, "num_examples": 357}], "download_size": 614159, "dataset_size": 947740}, {"config_name": "br_keb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 1026257, "num_examples": 888}], "download_size": 679680, "dataset_size": 1026257}, {"config_name": "bg_btb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 18545312, "num_examples": 8907}, {"name": "validation", "num_bytes": 2393174, "num_examples": 1115}, {"name": "test", "num_bytes": 2344136, "num_examples": 1116}], "download_size": 14910603, "dataset_size": 23282622}, {"config_name": "bxr_bdt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 17364, "num_examples": 19}, {"name": "test", "num_bytes": 1116630, "num_examples": 908}], "download_size": 726053, "dataset_size": 1133994}, {"config_name": "yue_hk", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 1242850, "num_examples": 1004}], "download_size": 710060, "dataset_size": 1242850}, {"config_name": "ca_ancora", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 46502842, "num_examples": 13123}, {"name": "validation", "num_bytes": 6282364, "num_examples": 1709}, {"name": "test", "num_bytes": 6441038, "num_examples": 1846}], "download_size": 35924146, "dataset_size": 59226244}, {"config_name": "zh_cfl", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 660584, "num_examples": 451}], "download_size": 384725, "dataset_size": 660584}, {"config_name": "zh_gsd", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 9268661, "num_examples": 3997}, {"name": "validation", "num_bytes": 1188371, "num_examples": 500}, {"name": "test", "num_bytes": 1130467, "num_examples": 500}], "download_size": 6828367, "dataset_size": 11587499}, {"config_name": "zh_gsdsimp", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 9268663, "num_examples": 3997}, {"name": "validation", "num_bytes": 1188383, "num_examples": 500}, {"name": "test", "num_bytes": 1130459, "num_examples": 500}], "download_size": 6828419, "dataset_size": 11587505}, {"config_name": "zh_hk", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 880193, "num_examples": 1004}], "download_size": 494447, "dataset_size": 880193}, {"config_name": "zh_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2425817, "num_examples": 1000}], "download_size": 1606982, "dataset_size": 2425817}, {"config_name": "ckt_hse", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 808669, "num_examples": 1004}], "download_size": 771943, "dataset_size": 808669}, {"config_name": "lzh_kyoto", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 26615708, "num_examples": 38669}, {"name": "validation", "num_bytes": 3770507, "num_examples": 5296}, {"name": "test", "num_bytes": 3155207, "num_examples": 4469}], "download_size": 22658287, "dataset_size": 33541422}, {"config_name": "cop_scriptorium", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 3944468, "num_examples": 1089}, {"name": "validation", "num_bytes": 1566786, "num_examples": 381}, {"name": "test", "num_bytes": 1487709, "num_examples": 403}], "download_size": 4502996, "dataset_size": 6998963}, {"config_name": "hr_set", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 19104315, "num_examples": 6914}, {"name": "validation", "num_bytes": 2787184, "num_examples": 960}, {"name": "test", "num_bytes": 3035797, "num_examples": 1136}], "download_size": 15103034, "dataset_size": 24927296}, {"config_name": "cs_cac", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 81527862, "num_examples": 23478}, {"name": "validation", "num_bytes": 1898678, "num_examples": 603}, {"name": "test", "num_bytes": 1878841, "num_examples": 628}], "download_size": 55990235, "dataset_size": 85305381}, {"config_name": "cs_cltt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 4277239, "num_examples": 860}, {"name": "validation", "num_bytes": 752253, "num_examples": 129}, {"name": "test", "num_bytes": 646103, "num_examples": 136}], "download_size": 3745656, "dataset_size": 5675595}, {"config_name": "cs_fictree", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 21490020, "num_examples": 10160}, {"name": "validation", "num_bytes": 2677727, "num_examples": 1309}, {"name": "test", "num_bytes": 2679930, "num_examples": 1291}], "download_size": 17464342, "dataset_size": 26847677}, {"config_name": "cs_pdt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 201356662, "num_examples": 68495}, {"name": "validation", "num_bytes": 27366981, "num_examples": 9270}, {"name": "test", "num_bytes": 29817339, "num_examples": 10148}], "download_size": 171506068, "dataset_size": 258540982}, {"config_name": "cs_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 3195818, "num_examples": 1000}], "download_size": 2231853, "dataset_size": 3195818}, {"config_name": "da_ddt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 8689809, "num_examples": 4383}, {"name": "validation", "num_bytes": 1117939, "num_examples": 564}, {"name": "test", "num_bytes": 1082651, "num_examples": 565}], "download_size": 6425281, "dataset_size": 10890399}, {"config_name": "nl_alpino", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 22503950, "num_examples": 12264}, {"name": "validation", "num_bytes": 1411253, "num_examples": 718}, {"name": "test", "num_bytes": 1354908, "num_examples": 596}], "download_size": 16858557, "dataset_size": 25270111}, {"config_name": "nl_lassysmall", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 9001614, "num_examples": 5787}, {"name": "validation", "num_bytes": 1361552, "num_examples": 676}, {"name": "test", "num_bytes": 1391136, "num_examples": 875}], "download_size": 8034396, "dataset_size": 11754302}, {"config_name": "en_esl", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 5335977, "num_examples": 4124}, {"name": "validation", "num_bytes": 648562, "num_examples": 500}, {"name": "test", "num_bytes": 651829, "num_examples": 500}], "download_size": 3351548, "dataset_size": 6636368}, {"config_name": "en_ewt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 22755753, "num_examples": 12543}, {"name": "validation", "num_bytes": 2829889, "num_examples": 2002}, {"name": "test", "num_bytes": 2820398, "num_examples": 2077}], "download_size": 16893922, "dataset_size": 28406040}, {"config_name": "en_gum", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 8999554, "num_examples": 4287}, {"name": "validation", "num_bytes": 1704949, "num_examples": 784}, {"name": "test", "num_bytes": 1743317, "num_examples": 890}], "download_size": 7702761, "dataset_size": 12447820}, {"config_name": "en_gumreddit", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 1365930, "num_examples": 587}, {"name": "validation", "num_bytes": 317546, "num_examples": 150}, {"name": "test", "num_bytes": 374707, "num_examples": 158}], "download_size": 1195979, "dataset_size": 2058183}, {"config_name": "en_lines", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 5728898, "num_examples": 3176}, {"name": "validation", "num_bytes": 1911762, "num_examples": 1032}, {"name": "test", "num_bytes": 1766797, "num_examples": 1035}], "download_size": 5522254, "dataset_size": 9407457}, {"config_name": "en_partut", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 4133445, "num_examples": 1781}, {"name": "validation", "num_bytes": 265039, "num_examples": 156}, {"name": "test", "num_bytes": 326834, "num_examples": 153}], "download_size": 2720286, "dataset_size": 4725318}, {"config_name": "en_pronouns", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 207364, "num_examples": 285}], "download_size": 147181, "dataset_size": 207364}, {"config_name": "en_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2282027, "num_examples": 1000}], "download_size": 1340563, "dataset_size": 2282027}, {"config_name": "myv_jr", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2763297, "num_examples": 1690}], "download_size": 1945981, "dataset_size": 2763297}, {"config_name": "et_edt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 42901059, "num_examples": 24633}, {"name": "validation", "num_bytes": 5551620, "num_examples": 3125}, {"name": "test", "num_bytes": 5994421, "num_examples": 3214}], "download_size": 32393618, "dataset_size": 54447100}, {"config_name": "et_ewt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 4199896, "num_examples": 2837}, {"name": "validation", "num_bytes": 1089459, "num_examples": 743}, {"name": "test", "num_bytes": 1600116, "num_examples": 913}], "download_size": 4044147, "dataset_size": 6889471}, {"config_name": "fo_farpahc", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 2114958, "num_examples": 1020}, {"name": "validation", "num_bytes": 809707, "num_examples": 300}, {"name": "test", "num_bytes": 798245, "num_examples": 301}], "download_size": 2186706, "dataset_size": 3722910}, {"config_name": "fo_oft", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 1220792, "num_examples": 1208}], "download_size": 802681, "dataset_size": 1220792}, {"config_name": "fi_ftb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 16800109, "num_examples": 14981}, {"name": "validation", "num_bytes": 2074201, "num_examples": 1875}, {"name": "test", "num_bytes": 2144908, "num_examples": 1867}], "download_size": 13132466, "dataset_size": 21019218}, {"config_name": "fi_ood", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2366923, "num_examples": 2122}], "download_size": 1480506, "dataset_size": 2366923}, {"config_name": "fi_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2086421, "num_examples": 1000}], "download_size": 1411514, "dataset_size": 2086421}, {"config_name": "fi_tdt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 22065448, "num_examples": 12217}, {"name": "validation", "num_bytes": 2483303, "num_examples": 1364}, {"name": "test", "num_bytes": 2855263, "num_examples": 1555}], "download_size": 16692242, "dataset_size": 27404014}, {"config_name": "fr_fqb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2674644, "num_examples": 2289}], "download_size": 1556235, "dataset_size": 2674644}, {"config_name": "fr_ftb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 44714315, "num_examples": 14759}, {"name": "validation", "num_bytes": 3929428, "num_examples": 1235}, {"name": "test", "num_bytes": 7583038, "num_examples": 2541}], "download_size": 30926802, "dataset_size": 56226781}, {"config_name": "fr_gsd", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 38329902, "num_examples": 14449}, {"name": "validation", "num_bytes": 3861548, "num_examples": 1476}, {"name": "test", "num_bytes": 1086926, "num_examples": 416}], "download_size": 25492044, "dataset_size": 43278376}, {"config_name": "fr_partut", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 2620477, "num_examples": 803}, {"name": "validation", "num_bytes": 205839, "num_examples": 107}, {"name": "test", "num_bytes": 288829, "num_examples": 110}], "download_size": 1817897, "dataset_size": 3115145}, {"config_name": "fr_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2660405, "num_examples": 1000}], "download_size": 1685033, "dataset_size": 2660405}, {"config_name": "fr_sequoia", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 5370647, "num_examples": 2231}, {"name": "validation", "num_bytes": 1065411, "num_examples": 412}, {"name": "test", "num_bytes": 1067676, "num_examples": 456}], "download_size": 4415282, "dataset_size": 7503734}, {"config_name": "fr_spoken", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 1625626, "num_examples": 1167}, {"name": "validation", "num_bytes": 1091750, "num_examples": 909}, {"name": "test", "num_bytes": 1078438, "num_examples": 730}], "download_size": 2483341, "dataset_size": 3795814}, {"config_name": "gl_ctg", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 8157432, "num_examples": 2272}, {"name": "validation", "num_bytes": 3057483, "num_examples": 860}, {"name": "test", "num_bytes": 3053764, "num_examples": 861}], "download_size": 8230649, "dataset_size": 14268679}, {"config_name": "gl_treegal", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 1804389, "num_examples": 600}, {"name": "test", "num_bytes": 1174023, "num_examples": 400}], "download_size": 1741471, "dataset_size": 2978412}, {"config_name": "de_gsd", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 32297384, "num_examples": 13814}, {"name": "validation", "num_bytes": 1504189, "num_examples": 799}, {"name": "test", "num_bytes": 2000117, "num_examples": 977}], "download_size": 21507364, "dataset_size": 35801690}, {"config_name": "de_hdt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 334214761, "num_examples": 153035}, {"name": "validation", "num_bytes": 39099013, "num_examples": 18434}, {"name": "test", "num_bytes": 39519143, "num_examples": 18459}], "download_size": 249243037, "dataset_size": 412832917}, {"config_name": "de_lit", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 3327891, "num_examples": 1922}], "download_size": 2060988, "dataset_size": 3327891}, {"config_name": "de_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2684407, "num_examples": 1000}], "download_size": 1731875, "dataset_size": 2684407}, {"config_name": "got_proiel", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 5175361, "num_examples": 3387}, {"name": "validation", "num_bytes": 1498101, "num_examples": 985}, {"name": "test", "num_bytes": 1518642, "num_examples": 1029}], "download_size": 5225655, "dataset_size": 8192104}, {"config_name": "el_gdt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 6028077, "num_examples": 1662}, {"name": "validation", "num_bytes": 1492610, "num_examples": 403}, {"name": "test", "num_bytes": 1521094, "num_examples": 456}], "download_size": 5788161, "dataset_size": 9041781}, {"config_name": "he_htb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 17324640, "num_examples": 5241}, {"name": "validation", "num_bytes": 1440985, "num_examples": 484}, {"name": "test", "num_bytes": 1550465, "num_examples": 491}], "download_size": 12054025, "dataset_size": 20316090}, {"config_name": "qhe_hiencs", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 1510145, "num_examples": 1448}, {"name": "validation", "num_bytes": 244129, "num_examples": 225}, {"name": "test", "num_bytes": 236291, "num_examples": 225}], "download_size": 914584, "dataset_size": 1990565}, {"config_name": "hi_hdtb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 61893814, "num_examples": 13304}, {"name": "validation", "num_bytes": 7748544, "num_examples": 1659}, {"name": "test", "num_bytes": 7786343, "num_examples": 1684}], "download_size": 51589681, "dataset_size": 77428701}, {"config_name": "hi_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 3384789, "num_examples": 1000}], "download_size": 2303495, "dataset_size": 3384789}, {"config_name": "hu_szeged", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 2822934, "num_examples": 910}, {"name": "validation", "num_bytes": 1584932, "num_examples": 441}, {"name": "test", "num_bytes": 1419130, "num_examples": 449}], "download_size": 3687905, "dataset_size": 5826996}, {"config_name": "is_icepahc", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 97197159, "num_examples": 34007}, {"name": "validation", "num_bytes": 18931295, "num_examples": 4865}, {"name": "test", "num_bytes": 19039838, "num_examples": 5157}], "download_size": 85106126, "dataset_size": 135168292}, {"config_name": "is_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2304432, "num_examples": 1000}], "download_size": 1525635, "dataset_size": 2304432}, {"config_name": "id_csui", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 1611334, "num_examples": 656}, {"name": "test", "num_bytes": 888832, "num_examples": 374}], "download_size": 1448601, "dataset_size": 2500166}, {"config_name": "id_gsd", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 11728948, "num_examples": 4477}, {"name": "validation", "num_bytes": 1513894, "num_examples": 559}, {"name": "test", "num_bytes": 1417208, "num_examples": 557}], "download_size": 9487349, "dataset_size": 14660050}, {"config_name": "id_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 1768596, "num_examples": 1000}], "download_size": 1149692, "dataset_size": 1768596}, {"config_name": "ga_idt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 10327215, "num_examples": 4005}, {"name": "validation", "num_bytes": 1057313, "num_examples": 451}, {"name": "test", "num_bytes": 1109028, "num_examples": 454}], "download_size": 7417728, "dataset_size": 12493556}, {"config_name": "it_isdt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 33510781, "num_examples": 13121}, {"name": "validation", "num_bytes": 1439348, "num_examples": 564}, {"name": "test", "num_bytes": 1267932, "num_examples": 482}], "download_size": 20998527, "dataset_size": 36218061}, {"config_name": "it_partut", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 5428686, "num_examples": 1781}, {"name": "validation", "num_bytes": 335085, "num_examples": 156}, {"name": "test", "num_bytes": 413752, "num_examples": 153}], "download_size": 3582155, "dataset_size": 6177523}, {"config_name": "it_postwita", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 10523322, "num_examples": 5368}, {"name": "validation", "num_bytes": 1299818, "num_examples": 671}, {"name": "test", "num_bytes": 1344079, "num_examples": 674}], "download_size": 7611319, "dataset_size": 13167219}, {"config_name": "it_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2612838, "num_examples": 1000}], "download_size": 1641073, "dataset_size": 2612838}, {"config_name": "it_twittiro", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 2536429, "num_examples": 1138}, {"name": "validation", "num_bytes": 323504, "num_examples": 144}, {"name": "test", "num_bytes": 316211, "num_examples": 142}], "download_size": 1894686, "dataset_size": 3176144}, {"config_name": "it_vit", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 24536095, "num_examples": 8277}, {"name": "validation", "num_bytes": 3144507, "num_examples": 743}, {"name": "test", "num_bytes": 2870355, "num_examples": 1067}], "download_size": 17605311, "dataset_size": 30550957}, {"config_name": "ja_bccwj", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 119164443, "num_examples": 40740}, {"name": "validation", "num_bytes": 23390188, "num_examples": 8417}, {"name": "test", "num_bytes": 21904413, "num_examples": 7871}], "download_size": 87340125, "dataset_size": 164459044}, {"config_name": "ja_gsd", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 36905139, "num_examples": 7027}, {"name": "validation", "num_bytes": 2662999, "num_examples": 501}, {"name": "test", "num_bytes": 2858141, "num_examples": 543}], "download_size": 30397358, "dataset_size": 42426279}, {"config_name": "ja_modern", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 3062149, "num_examples": 822}], "download_size": 2163988, "dataset_size": 3062149}, {"config_name": "ja_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 6322307, "num_examples": 1000}], "download_size": 4661525, "dataset_size": 6322307}, {"config_name": "krl_kkpp", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 370378, "num_examples": 228}], "download_size": 226103, "dataset_size": 370378}, {"config_name": "kk_ktb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 64737, "num_examples": 31}, {"name": "test", "num_bytes": 1263246, "num_examples": 1047}], "download_size": 849300, "dataset_size": 1327983}, {"config_name": "kfm_aha", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 8464, "num_examples": 10}], "download_size": 6290, "dataset_size": 8464}, {"config_name": "koi_uh", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 117629, "num_examples": 81}], "download_size": 91509, "dataset_size": 117629}, {"config_name": "kpv_ikdp", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 182189, "num_examples": 132}], "download_size": 121684, "dataset_size": 182189}, {"config_name": "kpv_lattice", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 685683, "num_examples": 435}], "download_size": 467085, "dataset_size": 685683}, {"config_name": "ko_gsd", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 5480313, "num_examples": 4400}, {"name": "validation", "num_bytes": 1156603, "num_examples": 950}, {"name": "test", "num_bytes": 1129555, "num_examples": 989}], "download_size": 4882238, "dataset_size": 7766471}, {"config_name": "ko_kaist", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 29037654, "num_examples": 23010}, {"name": "validation", "num_bytes": 2511880, "num_examples": 2066}, {"name": "test", "num_bytes": 2792215, "num_examples": 2287}], "download_size": 21855177, "dataset_size": 34341749}, {"config_name": "ko_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2511856, "num_examples": 1000}], "download_size": 2024810, "dataset_size": 2511856}, {"config_name": "kmr_mg", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 30374, "num_examples": 20}, {"name": "test", "num_bytes": 1248564, "num_examples": 734}], "download_size": 765158, "dataset_size": 1278938}, {"config_name": "la_ittb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 54306304, "num_examples": 22775}, {"name": "validation", "num_bytes": 4236222, "num_examples": 2101}, {"name": "test", "num_bytes": 4221459, "num_examples": 2101}], "download_size": 40247546, "dataset_size": 62763985}, {"config_name": "la_llct", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 26885433, "num_examples": 7289}, {"name": "validation", "num_bytes": 3363915, "num_examples": 850}, {"name": "test", "num_bytes": 3352500, "num_examples": 884}], "download_size": 21975884, "dataset_size": 33601848}, {"config_name": "la_perseus", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 2542043, "num_examples": 1334}, {"name": "test", "num_bytes": 1575350, "num_examples": 939}], "download_size": 2573703, "dataset_size": 4117393}, {"config_name": "la_proiel", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 24956038, "num_examples": 15917}, {"name": "validation", "num_bytes": 2020476, "num_examples": 1234}, {"name": "test", "num_bytes": 2029828, "num_examples": 1260}], "download_size": 18434442, "dataset_size": 29006342}, {"config_name": "lv_lvtb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 29167529, "num_examples": 10156}, {"name": "validation", "num_bytes": 4501172, "num_examples": 1664}, {"name": "test", "num_bytes": 4565919, "num_examples": 1823}], "download_size": 25227301, "dataset_size": 38234620}, {"config_name": "lt_alksnis", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 7272501, "num_examples": 2341}, {"name": "validation", "num_bytes": 1763901, "num_examples": 617}, {"name": "test", "num_bytes": 1648521, "num_examples": 684}], "download_size": 7008248, "dataset_size": 10684923}, {"config_name": "lt_hse", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 433214, "num_examples": 153}, {"name": "validation", "num_bytes": 433214, "num_examples": 153}, {"name": "test", "num_bytes": 433214, "num_examples": 153}], "download_size": 265619, "dataset_size": 1299642}, {"config_name": "olo_kkpp", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 18096, "num_examples": 19}, {"name": "test", "num_bytes": 175355, "num_examples": 106}], "download_size": 121837, "dataset_size": 193451}, {"config_name": "mt_mudt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 1858001, "num_examples": 1123}, {"name": "validation", "num_bytes": 826004, "num_examples": 433}, {"name": "test", "num_bytes": 892629, "num_examples": 518}], "download_size": 2011753, "dataset_size": 3576634}, {"config_name": "gv_cadhan", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 483042, "num_examples": 291}], "download_size": 287206, "dataset_size": 483042}, {"config_name": "mr_ufal", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 420345, "num_examples": 373}, {"name": "validation", "num_bytes": 60791, "num_examples": 46}, {"name": "test", "num_bytes": 56582, "num_examples": 47}], "download_size": 339354, "dataset_size": 537718}, {"config_name": "gun_dooley", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 1037858, "num_examples": 1046}], "download_size": 571571, "dataset_size": 1037858}, {"config_name": "gun_thomas", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 143111, "num_examples": 98}], "download_size": 92963, "dataset_size": 143111}, {"config_name": "mdf_jr", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 234147, "num_examples": 167}], "download_size": 162330, "dataset_size": 234147}, {"config_name": "myu_tudet", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 26202, "num_examples": 62}], "download_size": 20315, "dataset_size": 26202}, {"config_name": "pcm_nsc", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 16079391, "num_examples": 7279}, {"name": "validation", "num_bytes": 2099571, "num_examples": 991}, {"name": "test", "num_bytes": 2063685, "num_examples": 972}], "download_size": 14907410, "dataset_size": 20242647}, {"config_name": "nyq_aha", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 8723, "num_examples": 10}], "download_size": 6387, "dataset_size": 8723}, {"config_name": "sme_giella", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 1987666, "num_examples": 2257}, {"name": "test", "num_bytes": 1142396, "num_examples": 865}], "download_size": 1862302, "dataset_size": 3130062}, {"config_name": "no_bokmaal", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 25647647, "num_examples": 15696}, {"name": "validation", "num_bytes": 3828310, "num_examples": 2409}, {"name": "test", "num_bytes": 3151638, "num_examples": 1939}], "download_size": 19177350, "dataset_size": 32627595}, {"config_name": "no_nynorsk", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 25630539, "num_examples": 14174}, {"name": "validation", "num_bytes": 3277649, "num_examples": 1890}, {"name": "test", "num_bytes": 2601676, "num_examples": 1511}], "download_size": 18532495, "dataset_size": 31509864}, {"config_name": "no_nynorsklia", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 3500907, "num_examples": 3412}, {"name": "validation", "num_bytes": 1003845, "num_examples": 881}, {"name": "test", "num_bytes": 999943, "num_examples": 957}], "download_size": 3349676, "dataset_size": 5504695}, {"config_name": "cu_proiel", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 6106144, "num_examples": 4124}, {"name": "validation", "num_bytes": 1639912, "num_examples": 1073}, {"name": "test", "num_bytes": 1648459, "num_examples": 1141}], "download_size": 6239839, "dataset_size": 9394515}, {"config_name": "fro_srcmf", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 11959859, "num_examples": 13909}, {"name": "validation", "num_bytes": 1526574, "num_examples": 1842}, {"name": "test", "num_bytes": 1535923, "num_examples": 1927}], "download_size": 9043098, "dataset_size": 15022356}, {"config_name": "orv_rnc", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 1527306, "num_examples": 320}, {"name": "test", "num_bytes": 2552216, "num_examples": 637}], "download_size": 2627398, "dataset_size": 4079522}, {"config_name": "orv_torot", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 18077991, "num_examples": 13336}, {"name": "validation", "num_bytes": 2408313, "num_examples": 1852}, {"name": "test", "num_bytes": 2347934, "num_examples": 1756}], "download_size": 15296362, "dataset_size": 22834238}, {"config_name": "otk_tonqq", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 22829, "num_examples": 18}], "download_size": 14389, "dataset_size": 22829}, {"config_name": "fa_perdt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 48654947, "num_examples": 26196}, {"name": "validation", "num_bytes": 2687750, "num_examples": 1456}, {"name": "test", "num_bytes": 2600303, "num_examples": 1455}], "download_size": 33606395, "dataset_size": 53943000}, {"config_name": "fa_seraji", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 12627691, "num_examples": 4798}, {"name": "validation", "num_bytes": 1634327, "num_examples": 599}, {"name": "test", "num_bytes": 1675134, "num_examples": 600}], "download_size": 9890107, "dataset_size": 15937152}, {"config_name": "pl_lfg", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 16810910, "num_examples": 13774}, {"name": "validation", "num_bytes": 2093712, "num_examples": 1745}, {"name": "test", "num_bytes": 2100915, "num_examples": 1727}], "download_size": 14865541, "dataset_size": 21005537}, {"config_name": "pl_pdb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 44652289, "num_examples": 17722}, {"name": "validation", "num_bytes": 5494883, "num_examples": 2215}, {"name": "test", "num_bytes": 5322608, "num_examples": 2215}], "download_size": 36340919, "dataset_size": 55469780}, {"config_name": "pl_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2943603, "num_examples": 1000}], "download_size": 1943983, "dataset_size": 2943603}, {"config_name": "pt_bosque", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 22808617, "num_examples": 8328}, {"name": "validation", "num_bytes": 1201577, "num_examples": 560}, {"name": "test", "num_bytes": 1131511, "num_examples": 476}], "download_size": 15201503, "dataset_size": 25141705}, {"config_name": "pt_gsd", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 22208385, "num_examples": 9664}, {"name": "validation", "num_bytes": 2805628, "num_examples": 1210}, {"name": "test", "num_bytes": 2732063, "num_examples": 1204}], "download_size": 15300844, "dataset_size": 27746076}, {"config_name": "pt_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2431942, "num_examples": 1000}], "download_size": 1516883, "dataset_size": 2431942}, {"config_name": "ro_nonstandard", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 74489083, "num_examples": 24121}, {"name": "validation", "num_bytes": 2663152, "num_examples": 1052}, {"name": "test", "num_bytes": 3017162, "num_examples": 1052}], "download_size": 50345748, "dataset_size": 80169397}, {"config_name": "ro_rrt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 23695399, "num_examples": 8043}, {"name": "validation", "num_bytes": 2190973, "num_examples": 752}, {"name": "test", "num_bytes": 2092520, "num_examples": 729}], "download_size": 17187956, "dataset_size": 27978892}, {"config_name": "ro_simonero", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 15390734, "num_examples": 3747}, {"name": "validation", "num_bytes": 1926639, "num_examples": 443}, {"name": "test", "num_bytes": 1940787, "num_examples": 491}], "download_size": 11409378, "dataset_size": 19258160}, {"config_name": "ru_gsd", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 10504099, "num_examples": 3850}, {"name": "validation", "num_bytes": 1635884, "num_examples": 579}, {"name": "test", "num_bytes": 1597603, "num_examples": 601}], "download_size": 8830986, "dataset_size": 13737586}, {"config_name": "ru_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2695958, "num_examples": 1000}], "download_size": 1869304, "dataset_size": 2695958}, {"config_name": "ru_syntagrus", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 126305584, "num_examples": 48814}, {"name": "validation", "num_bytes": 17043673, "num_examples": 6584}, {"name": "test", "num_bytes": 16880203, "num_examples": 6491}], "download_size": 102745164, "dataset_size": 160229460}, {"config_name": "ru_taiga", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 5802733, "num_examples": 3138}, {"name": "validation", "num_bytes": 1382140, "num_examples": 945}, {"name": "test", "num_bytes": 1314084, "num_examples": 881}], "download_size": 5491427, "dataset_size": 8498957}, {"config_name": "sa_ufal", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 431697, "num_examples": 230}], "download_size": 424675, "dataset_size": 431697}, {"config_name": "sa_vedic", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 2179608, "num_examples": 2524}, {"name": "test", "num_bytes": 1209605, "num_examples": 1473}], "download_size": 2041583, "dataset_size": 3389213}, {"config_name": "gd_arcosg", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 3952356, "num_examples": 1990}, {"name": "validation", "num_bytes": 1038211, "num_examples": 645}, {"name": "test", "num_bytes": 1034788, "num_examples": 538}], "download_size": 3474087, "dataset_size": 6025355}, {"config_name": "sr_set", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 9309552, "num_examples": 3328}, {"name": "validation", "num_bytes": 1503953, "num_examples": 536}, {"name": "test", "num_bytes": 1432672, "num_examples": 520}], "download_size": 7414381, "dataset_size": 12246177}, {"config_name": "sms_giellagas", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 174744, "num_examples": 104}], "download_size": 116491, "dataset_size": 174744}, {"config_name": "sk_snk", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 12017312, "num_examples": 8483}, {"name": "validation", "num_bytes": 1863926, "num_examples": 1060}, {"name": "test", "num_bytes": 1943012, "num_examples": 1061}], "download_size": 10013420, "dataset_size": 15824250}, {"config_name": "sl_ssj", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 16713639, "num_examples": 6478}, {"name": "validation", "num_bytes": 2070847, "num_examples": 734}, {"name": "test", "num_bytes": 2083062, "num_examples": 788}], "download_size": 12455962, "dataset_size": 20867548}, {"config_name": "sl_sst", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 2903675, "num_examples": 2078}, {"name": "test", "num_bytes": 1493885, "num_examples": 1110}], "download_size": 2655777, "dataset_size": 4397560}, {"config_name": "soj_aha", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 6218, "num_examples": 8}], "download_size": 4577, "dataset_size": 6218}, {"config_name": "ajp_madar", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 71956, "num_examples": 100}], "download_size": 43174, "dataset_size": 71956}, {"config_name": "es_ancora", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 50101327, "num_examples": 14305}, {"name": "validation", "num_bytes": 5883940, "num_examples": 1654}, {"name": "test", "num_bytes": 5928986, "num_examples": 1721}], "download_size": 37668083, "dataset_size": 61914253}, {"config_name": "es_gsd", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 39582074, "num_examples": 14187}, {"name": "validation", "num_bytes": 3834443, "num_examples": 1400}, {"name": "test", "num_bytes": 1253720, "num_examples": 426}], "download_size": 26073760, "dataset_size": 44670237}, {"config_name": "es_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2595946, "num_examples": 1000}], "download_size": 1628475, "dataset_size": 2595946}, {"config_name": "swl_sslc", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 57443, "num_examples": 87}, {"name": "validation", "num_bytes": 59002, "num_examples": 82}, {"name": "test", "num_bytes": 24542, "num_examples": 34}], "download_size": 81699, "dataset_size": 140987}, {"config_name": "sv_lines", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 6731662, "num_examples": 3176}, {"name": "validation", "num_bytes": 2239951, "num_examples": 1032}, {"name": "test", "num_bytes": 2070626, "num_examples": 1035}], "download_size": 7245283, "dataset_size": 11042239}, {"config_name": "sv_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2554725, "num_examples": 1000}], "download_size": 1722516, "dataset_size": 2554725}, {"config_name": "sv_talbanken", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 9287256, "num_examples": 4303}, {"name": "validation", "num_bytes": 1361535, "num_examples": 504}, {"name": "test", "num_bytes": 2835742, "num_examples": 1219}], "download_size": 8476012, "dataset_size": 13484533}, {"config_name": "gsw_uzh", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 111357, "num_examples": 100}], "download_size": 59675, "dataset_size": 111357}, {"config_name": "tl_trg", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 86696, "num_examples": 128}], "download_size": 61344, "dataset_size": 86696}, {"config_name": "tl_ugnayan", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 90863, "num_examples": 94}], "download_size": 55207, "dataset_size": 90863}, {"config_name": "ta_mwtt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 522349, "num_examples": 534}], "download_size": 414263, "dataset_size": 522349}, {"config_name": "ta_ttb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 1538780, "num_examples": 400}, {"name": "validation", "num_bytes": 305206, "num_examples": 80}, {"name": "test", "num_bytes": 478941, "num_examples": 120}], "download_size": 1753448, "dataset_size": 2322927}, {"config_name": "te_mtg", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 703512, "num_examples": 1051}, {"name": "validation", "num_bytes": 91547, "num_examples": 131}, {"name": "test", "num_bytes": 99757, "num_examples": 146}], "download_size": 643764, "dataset_size": 894816}, {"config_name": "th_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2341697, "num_examples": 1000}], "download_size": 1606517, "dataset_size": 2341697}, {"config_name": "tpn_tudet", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 8089, "num_examples": 8}], "download_size": 5447, "dataset_size": 8089}, {"config_name": "qtd_sagt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 583697, "num_examples": 285}, {"name": "validation", "num_bytes": 1564765, "num_examples": 801}, {"name": "test", "num_bytes": 1710777, "num_examples": 805}], "download_size": 2299611, "dataset_size": 3859239}, {"config_name": "tr_boun", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 12827173, "num_examples": 7803}, {"name": "validation", "num_bytes": 1577760, "num_examples": 979}, {"name": "test", "num_bytes": 1580727, "num_examples": 979}], "download_size": 9742035, "dataset_size": 15985660}, {"config_name": "tr_gb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2146729, "num_examples": 2880}], "download_size": 1474083, "dataset_size": 2146729}, {"config_name": "tr_imst", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 5063905, "num_examples": 3664}, {"name": "validation", "num_bytes": 1342351, "num_examples": 988}, {"name": "test", "num_bytes": 1347524, "num_examples": 983}], "download_size": 4711018, "dataset_size": 7753780}, {"config_name": "tr_pud", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 2021772, "num_examples": 1000}], "download_size": 1359487, "dataset_size": 2021772}, {"config_name": "uk_iu", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 18886802, "num_examples": 5496}, {"name": "validation", "num_bytes": 2592721, "num_examples": 672}, {"name": "test", "num_bytes": 3561164, "num_examples": 892}], "download_size": 17344586, "dataset_size": 25040687}, {"config_name": "hsb_ufal", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 54257, "num_examples": 23}, {"name": "test", "num_bytes": 1246592, "num_examples": 623}], "download_size": 781067, "dataset_size": 1300849}, {"config_name": "ur_udtb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 19808745, "num_examples": 4043}, {"name": "validation", "num_bytes": 2652349, "num_examples": 552}, {"name": "test", "num_bytes": 2702596, "num_examples": 535}], "download_size": 15901007, "dataset_size": 25163690}, {"config_name": "ug_udt", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 2570856, "num_examples": 1656}, {"name": "validation", "num_bytes": 1406032, "num_examples": 900}, {"name": "test", "num_bytes": 1371993, "num_examples": 900}], "download_size": 3455092, "dataset_size": 5348881}, {"config_name": "vi_vtb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 1689772, "num_examples": 1400}, {"name": "validation", "num_bytes": 948019, "num_examples": 800}, {"name": "test", "num_bytes": 987207, "num_examples": 800}], "download_size": 2055529, "dataset_size": 3624998}, {"config_name": "wbp_ufal", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 48533, "num_examples": 55}], "download_size": 38326, "dataset_size": 48533}, {"config_name": "cy_ccg", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 1629465, "num_examples": 704}, {"name": "test", "num_bytes": 1779002, "num_examples": 953}], "download_size": 1984759, "dataset_size": 3408467}, {"config_name": "wo_wtb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 2781883, "num_examples": 1188}, {"name": "validation", "num_bytes": 1204839, "num_examples": 449}, {"name": "test", "num_bytes": 1227124, "num_examples": 470}], "download_size": 3042699, "dataset_size": 5213846}, {"config_name": "yo_ytb", "features": [{"name": "idx", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "upos", "sequence": {"class_label": {"names": {"0": "NOUN", "1": "PUNCT", "2": "ADP", "3": "NUM", "4": "SYM", "5": "SCONJ", "6": "ADJ", "7": "PART", "8": "DET", "9": "CCONJ", "10": "PROPN", "11": "PRON", "12": "X", "13": "_", "14": "ADV", "15": "INTJ", "16": "VERB", "17": "AUX"}}}}, {"name": "xpos", "sequence": "string"}, {"name": "feats", "sequence": "string"}, {"name": "head", "sequence": "string"}, {"name": "deprel", "sequence": "string"}, {"name": "deps", "sequence": "string"}, {"name": "misc", "sequence": "string"}], "splits": [{"name": "test", "num_bytes": 905766, "num_examples": 318}], "download_size": 567955, "dataset_size": 905766}]}
2024-01-18T11:17:47+00:00
[]
[ "af", "aii", "ajp", "akk", "am", "apu", "aqz", "ar", "be", "bg", "bho", "bm", "br", "bxr", "ca", "ckt", "cop", "cs", "cu", "cy", "da", "de", "el", "en", "es", "et", "eu", "fa", "fi", "fo", "fr", "fro", "ga", "gd", "gl", "got", "grc", "gsw", "gun", "gv", "he", "hi", "hr", "hsb", "hu", "hy", "id", "is", "it", "ja", "kfm", "kk", "kmr", "ko", "koi", "kpv", "krl", "la", "lt", "lv", "lzh", "mdf", "mr", "mt", "myu", "myv", "nl", "no", "nyq", "olo", "orv", "otk", "pcm", "pl", "pt", "ro", "ru", "sa", "sk", "sl", "sme", "sms", "soj", "sq", "sr", "sv", "swl", "ta", "te", "th", "tl", "tpn", "tr", "ug", "uk", "ur", "vi", "wbp", "wo", "yo", "yue", "zh" ]
TAGS #task_categories-token-classification #task_ids-parsing #annotations_creators-expert-generated #language_creators-crowdsourced #multilinguality-multilingual #size_categories-1K<n<10K #source_datasets-original #language-Afrikaans #language-Assyrian Neo-Aramaic #language-South Levantine Arabic #language-Akkadian #language-Amharic #language-Apurinã #language-Akuntsu #language-Arabic #language-Belarusian #language-Bulgarian #language-Bhojpuri #language-Bambara #language-Breton #language-Russia Buriat #language-Catalan #language-Chukot #language-Coptic #language-Czech #language-Church Slavic #language-Welsh #language-Danish #language-German #language-Modern Greek (1453-) #language-English #language-Spanish #language-Estonian #language-Basque #language-Persian #language-Finnish #language-Faroese #language-French #language-Old French (842-ca. 1400) #language-Irish #language-Scottish Gaelic #language-Galician #language-Gothic #language-Ancient Greek (to 1453) #language-Swiss German #language-Mbyá Guaraní #language-Manx #language-Hebrew #language-Hindi #language-Croatian #language-Upper Sorbian #language-Hungarian #language-Armenian #language-Indonesian #language-Icelandic #language-Italian #language-Japanese #language-Khunsari #language-Kazakh #language-Northern Kurdish #language-Korean #language-Komi-Permyak #language-Komi-Zyrian #language-Karelian #language-Latin #language-Lithuanian #language-Latvian #language-Literary Chinese #language-Moksha #language-Marathi #language-Maltese #language-Mundurukú #language-Erzya #language-Dutch #language-Norwegian #language-Nayini #language-Livvi #language-Old Russian #language-Old Turkish #language-Nigerian Pidgin #language-Polish #language-Portuguese #language-Romanian #language-Russian #language-Sanskrit #language-Slovak #language-Slovenian #language-Northern Sami #language-Skolt Sami #language-Soi #language-Albanian #language-Serbian #language-Swedish #language-Swedish Sign Language #language-Tamil #language-Telugu #language-Thai #language-Tagalog #language-Tupinambá #language-Turkish #language-Uighur #language-Ukrainian #language-Urdu #language-Vietnamese #language-Warlpiri #language-Wolof #language-Yoruba #language-Yue Chinese #language-Chinese #license-unknown #constituency-parsing #dependency-parsing #region-us
# Dataset Card for Universal Dependencies Treebank ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: Universal Dependencies - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @patrickvonplaten, @jplu for adding this dataset.
[ "# Dataset Card for Universal Dependencies Treebank", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Universal Dependencies\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @patrickvonplaten, @jplu for adding this dataset." ]
[ "TAGS\n#task_categories-token-classification #task_ids-parsing #annotations_creators-expert-generated #language_creators-crowdsourced #multilinguality-multilingual #size_categories-1K<n<10K #source_datasets-original #language-Afrikaans #language-Assyrian Neo-Aramaic #language-South Levantine Arabic #language-Akkadian #language-Amharic #language-Apurinã #language-Akuntsu #language-Arabic #language-Belarusian #language-Bulgarian #language-Bhojpuri #language-Bambara #language-Breton #language-Russia Buriat #language-Catalan #language-Chukot #language-Coptic #language-Czech #language-Church Slavic #language-Welsh #language-Danish #language-German #language-Modern Greek (1453-) #language-English #language-Spanish #language-Estonian #language-Basque #language-Persian #language-Finnish #language-Faroese #language-French #language-Old French (842-ca. 1400) #language-Irish #language-Scottish Gaelic #language-Galician #language-Gothic #language-Ancient Greek (to 1453) #language-Swiss German #language-Mbyá Guaraní #language-Manx #language-Hebrew #language-Hindi #language-Croatian #language-Upper Sorbian #language-Hungarian #language-Armenian #language-Indonesian #language-Icelandic #language-Italian #language-Japanese #language-Khunsari #language-Kazakh #language-Northern Kurdish #language-Korean #language-Komi-Permyak #language-Komi-Zyrian #language-Karelian #language-Latin #language-Lithuanian #language-Latvian #language-Literary Chinese #language-Moksha #language-Marathi #language-Maltese #language-Mundurukú #language-Erzya #language-Dutch #language-Norwegian #language-Nayini #language-Livvi #language-Old Russian #language-Old Turkish #language-Nigerian Pidgin #language-Polish #language-Portuguese #language-Romanian #language-Russian #language-Sanskrit #language-Slovak #language-Slovenian #language-Northern Sami #language-Skolt Sami #language-Soi #language-Albanian #language-Serbian #language-Swedish #language-Swedish Sign Language #language-Tamil #language-Telugu #language-Thai #language-Tagalog #language-Tupinambá #language-Turkish #language-Uighur #language-Ukrainian #language-Urdu #language-Vietnamese #language-Warlpiri #language-Wolof #language-Yoruba #language-Yue Chinese #language-Chinese #license-unknown #constituency-parsing #dependency-parsing #region-us \n", "# Dataset Card for Universal Dependencies Treebank", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Universal Dependencies\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @patrickvonplaten, @jplu for adding this dataset." ]
[ 729, 11, 120, 28, 6, 10, 4, 6, 6, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 23 ]
[ "passage: " ]
20d4829303a5b853db00a3b82116fcdc1ff3d070
# Dataset Card for [Dataset Name] ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [UniMorph Homepage](https://unimorph.github.io/) - **Repository:** [List of UniMorph repositories](https://github.com/unimorph) - **Paper:** [The Composition and Use of the Universal Morphological Feature Schema (UniMorph Schema)](https://unimorph.github.io/doc/unimorph-schema.pdf) - **Point of Contact:** [Arya McCarthy](mailto:arya@jhu.edu) ### Dataset Summary The Universal Morphology (UniMorph) project is a collaborative effort to improve how NLP handles complex morphology in the world’s languages. The goal of UniMorph is to annotate morphological data in a universal schema that allows an inflected word from any language to be defined by its lexical meaning, typically carried by the lemma, and by a rendering of its inflectional form in terms of a bundle of morphological features from our schema. The specification of the schema is described in Sylak-Glassman (2016). ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The current version of the UniMorph dataset covers 110 languages. ## Dataset Structure ### Data Instances Each data instance comprises of a lemma and a set of possible realizations with morphological and meaning annotations. For example: ``` {'forms': {'Aktionsart': [[], [], [], [], []], 'Animacy': [[], [], [], [], []], ... 'Finiteness': [[], [], [], [1], []], ... 'Number': [[], [], [0], [], []], 'Other': [[], [], [], [], []], 'Part_Of_Speech': [[7], [10], [7], [7], [10]], ... 'Tense': [[1], [1], [0], [], [0]], ... 'word': ['ablated', 'ablated', 'ablates', 'ablate', 'ablating']}, 'lemma': 'ablate'} ``` ### Data Fields Each instance in the dataset has the following fields: - `lemma`: the common lemma for all all_forms - `forms`: all annotated forms for this lemma, with: - `word`: the full word form - [`category`]: a categorical variable denoting one or several tags in a category (several to represent composite tags, originally denoted with `A+B`). The full list of categories and possible tags for each can be found [here](https://github.com/unimorph/unimorph.github.io/blob/master/unimorph-schema-json/dimensions-to-features.json) ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@yjernite](https://github.com/yjernite) for adding this dataset.
universal_morphologies
[ "task_categories:token-classification", "task_categories:text-classification", "task_ids:multi-class-classification", "task_ids:multi-label-classification", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "size_categories:1K<n<10K", "size_categories:n<1K", "source_datasets:original", "language:ady", "language:ang", "language:ar", "language:arn", "language:ast", "language:az", "language:ba", "language:be", "language:bg", "language:bn", "language:bo", "language:br", "language:ca", "language:ckb", "language:crh", "language:cs", "language:csb", "language:cu", "language:cy", "language:da", "language:de", "language:dsb", "language:el", "language:en", "language:es", "language:et", "language:eu", "language:fa", "language:fi", "language:fo", "language:fr", "language:frm", "language:fro", "language:frr", "language:fur", "language:fy", "language:ga", "language:gal", "language:gd", "language:gmh", "language:gml", "language:got", "language:grc", "language:gv", "language:hai", "language:he", "language:hi", "language:hu", "language:hy", "language:is", "language:it", "language:izh", "language:ka", "language:kbd", "language:kjh", "language:kk", "language:kl", "language:klr", "language:kmr", "language:kn", "language:krl", "language:kw", "language:la", "language:liv", "language:lld", "language:lt", "language:lud", "language:lv", "language:mk", "language:mt", "language:mwf", "language:nap", "language:nb", "language:nds", "language:nl", "language:nn", "language:nv", "language:oc", "language:olo", "language:osx", "language:pl", "language:ps", "language:pt", "language:qu", "language:ro", "language:ru", "language:sa", "language:sga", "language:sh", "language:sl", "language:sme", "language:sq", "language:sv", "language:swc", "language:syc", "language:te", "language:tg", "language:tk", "language:tr", "language:tt", "language:uk", "language:ur", "language:uz", "language:vec", "language:vep", "language:vot", "language:xcl", "language:xno", "language:yi", "language:zu", "license:cc-by-sa-3.0", "morphology", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found"], "language": ["ady", "ang", "ar", "arn", "ast", "az", "ba", "be", "bg", "bn", "bo", "br", "ca", "ckb", "crh", "cs", "csb", "cu", "cy", "da", "de", "dsb", "el", "en", "es", "et", "eu", "fa", "fi", "fo", "fr", "frm", "fro", "frr", "fur", "fy", "ga", "gal", "gd", "gmh", "gml", "got", "grc", "gv", "hai", "he", "hi", "hu", "hy", "is", "it", "izh", "ka", "kbd", "kjh", "kk", "kl", "klr", "kmr", "kn", "krl", "kw", "la", "liv", "lld", "lt", "lud", "lv", "mk", "mt", "mwf", "nap", "nb", "nds", "nl", "nn", "nv", "oc", "olo", "osx", "pl", "ps", "pt", "qu", "ro", "ru", "sa", "sga", "sh", "sl", "sme", "sq", "sv", "swc", "syc", "te", "tg", "tk", "tr", "tt", "uk", "ur", "uz", "vec", "vep", "vot", "xcl", "xno", "yi", "zu"], "license": ["cc-by-sa-3.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K", "1K<n<10K", "n<1K"], "source_datasets": ["original"], "task_categories": ["token-classification", "text-classification"], "task_ids": ["multi-class-classification", "multi-label-classification"], "pretty_name": "UniversalMorphologies", "config_names": ["ady", "ang", "ara", "arn", "ast", "aze", "bak", "bel", "ben", "bod", "bre", "bul", "cat", "ces", "chu", "ckb", "cor", "crh", "csb", "cym", "dan", "deu", "dsb", "ell", "eng", "est", "eus", "fao", "fas", "fin", "fra", "frm", "fro", "frr", "fry", "fur", "gal", "gla", "gle", "glv", "gmh", "gml", "got", "grc", "hai", "hbs", "heb", "hin", "hun", "hye", "isl", "ita", "izh", "kal", "kan", "kat", "kaz", "kbd", "kjh", "klr", "kmr", "krl", "lat", "lav", "lit", "liv", "lld", "lud", "mkd", "mlt", "mwf", "nap", "nav", "nds", "nld", "nno", "nob", "oci", "olo", "osx", "pol", "por", "pus", "que", "ron", "rus", "san", "sga", "slv", "sme", "spa", "sqi", "swc", "swe", "syc", "tat", "tel", "tgk", "tuk", "tur", "ukr", "urd", "uzb", "vec", "vep", "vot", "xcl", "xno", "yid", "zul"], "tags": ["morphology"], "dataset_info": [{"config_name": "ady", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 3428235, "num_examples": 1666}], "download_size": 1008487, "dataset_size": 3428235}, {"config_name": "ang", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 6569844, "num_examples": 1867}], "download_size": 1435972, "dataset_size": 6569844}, {"config_name": "ara", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 24388295, "num_examples": 4134}], "download_size": 7155824, "dataset_size": 24388295}, {"config_name": "arn", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 124050, "num_examples": 26}], "download_size": 20823, "dataset_size": 124050}, {"config_name": "ast", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 4913008, "num_examples": 436}], "download_size": 1175901, "dataset_size": 4913008}, {"config_name": "aze", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 1248687, "num_examples": 340}], "download_size": 276306, "dataset_size": 1248687}, {"config_name": "bak", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 1984657, "num_examples": 1084}], "download_size": 494758, "dataset_size": 1984657}, {"config_name": "bel", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 2626405, "num_examples": 1027}], "download_size": 739537, "dataset_size": 2626405}, {"config_name": "ben", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 746181, "num_examples": 136}], "download_size": 251991, "dataset_size": 746181}, {"config_name": "bod", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 880074, "num_examples": 1335}], "download_size": 197523, "dataset_size": 880074}, {"config_name": "bre", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 387583, "num_examples": 44}], "download_size": 82159, "dataset_size": 387583}, {"config_name": "bul", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 9589915, "num_examples": 2468}], "download_size": 3074574, "dataset_size": 9589915}, {"config_name": "cat", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 12988492, "num_examples": 1547}], "download_size": 2902458, "dataset_size": 12988492}, {"config_name": "ces", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 21056640, "num_examples": 5125}], "download_size": 4875288, "dataset_size": 21056640}, {"config_name": "chu", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 628237, "num_examples": 152}], "download_size": 149081, "dataset_size": 628237}, {"config_name": "ckb", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 3843267, "num_examples": 274}], "download_size": 914302, "dataset_size": 3843267}, {"config_name": "cor", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 83434, "num_examples": 9}], "download_size": 17408, "dataset_size": 83434}, {"config_name": "crh", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 1154595, "num_examples": 1230}], "download_size": 186325, "dataset_size": 1154595}, {"config_name": "csb", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 82172, "num_examples": 37}], "download_size": 14259, "dataset_size": 82172}, {"config_name": "cym", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 1748431, "num_examples": 183}], "download_size": 374501, "dataset_size": 1748431}, {"config_name": "dan", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 4204551, "num_examples": 3193}], "download_size": 845939, "dataset_size": 4204551}, {"config_name": "deu", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 28436466, "num_examples": 15060}], "download_size": 5966618, "dataset_size": 28436466}, {"config_name": "dsb", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 2985168, "num_examples": 994}], "download_size": 536096, "dataset_size": 2985168}, {"config_name": "ell", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 34112450, "num_examples": 11906}], "download_size": 11222248, "dataset_size": 34112450}, {"config_name": "eng", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 18455909, "num_examples": 22765}], "download_size": 3285554, "dataset_size": 18455909}, {"config_name": "est", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 6125879, "num_examples": 886}], "download_size": 1397385, "dataset_size": 6125879}, {"config_name": "eus", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 2444247, "num_examples": 26}], "download_size": 876480, "dataset_size": 2444247}, {"config_name": "fao", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 7117926, "num_examples": 3077}], "download_size": 1450065, "dataset_size": 7117926}, {"config_name": "fas", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 6382709, "num_examples": 273}], "download_size": 2104724, "dataset_size": 6382709}, {"config_name": "fin", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "1", "num_bytes": 331855860, "num_examples": 46152}, {"name": "2", "num_bytes": 81091817, "num_examples": 11491}], "download_size": 109324828, "dataset_size": 412947677}, {"config_name": "fra", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 58747699, "num_examples": 7535}], "download_size": 13404983, "dataset_size": 58747699}, {"config_name": "frm", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 6015940, "num_examples": 603}], "download_size": 1441122, "dataset_size": 6015940}, {"config_name": "fro", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 20260793, "num_examples": 1700}], "download_size": 4945582, "dataset_size": 20260793}, {"config_name": "frr", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 526898, "num_examples": 51}], "download_size": 112236, "dataset_size": 526898}, {"config_name": "fry", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 222067, "num_examples": 85}], "download_size": 38227, "dataset_size": 222067}, {"config_name": "fur", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 1282374, "num_examples": 168}], "download_size": 258793, "dataset_size": 1282374}, {"config_name": "gal", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 5844604, "num_examples": 486}], "download_size": 1259120, "dataset_size": 5844604}, {"config_name": "gla", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 126847, "num_examples": 73}], "download_size": 25025, "dataset_size": 126847}, {"config_name": "gle", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 17065939, "num_examples": 7464}], "download_size": 3853188, "dataset_size": 17065939}, {"config_name": "glv", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 7523, "num_examples": 1}], "download_size": 401, "dataset_size": 7523}, {"config_name": "gmh", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 114677, "num_examples": 29}], "download_size": 20851, "dataset_size": 114677}, {"config_name": "gml", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 233831, "num_examples": 52}], "download_size": 47151, "dataset_size": 233831}, {"config_name": "got", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train"}], "download_size": 2, "dataset_size": 0}, {"config_name": "grc", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 6779867, "num_examples": 2431}], "download_size": 2057514, "dataset_size": 6779867}, {"config_name": "hai", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 1166240, "num_examples": 41}], "download_size": 329817, "dataset_size": 1166240}, {"config_name": "hbs", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 132933961, "num_examples": 24419}], "download_size": 32194142, "dataset_size": 132933961}, {"config_name": "heb", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 2211208, "num_examples": 510}], "download_size": 498065, "dataset_size": 2211208}, {"config_name": "hin", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 10083004, "num_examples": 258}], "download_size": 3994359, "dataset_size": 10083004}, {"config_name": "hun", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 83517327, "num_examples": 14892}], "download_size": 19544319, "dataset_size": 83517327}, {"config_name": "hye", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 56537127, "num_examples": 7033}], "download_size": 17810316, "dataset_size": 56537127}, {"config_name": "isl", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 12120572, "num_examples": 4775}], "download_size": 2472980, "dataset_size": 12120572}, {"config_name": "ita", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 81905203, "num_examples": 10009}], "download_size": 19801423, "dataset_size": 81905203}, {"config_name": "izh", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 170094, "num_examples": 50}], "download_size": 28558, "dataset_size": 170094}, {"config_name": "kal", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 60434, "num_examples": 23}], "download_size": 9795, "dataset_size": 60434}, {"config_name": "kan", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 1052294, "num_examples": 159}], "download_size": 318512, "dataset_size": 1052294}, {"config_name": "kat", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 12532540, "num_examples": 3782}], "download_size": 4678979, "dataset_size": 12532540}, {"config_name": "kaz", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 62519, "num_examples": 26}], "download_size": 14228, "dataset_size": 62519}, {"config_name": "kbd", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 511406, "num_examples": 250}], "download_size": 133788, "dataset_size": 511406}, {"config_name": "kjh", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 193741, "num_examples": 75}], "download_size": 44907, "dataset_size": 193741}, {"config_name": "klr", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 28909688, "num_examples": 591}], "download_size": 7561829, "dataset_size": 28909688}, {"config_name": "kmr", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 35504487, "num_examples": 15083}], "download_size": 8592722, "dataset_size": 35504487}, {"config_name": "krl", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 106475, "num_examples": 20}], "download_size": 19024, "dataset_size": 106475}, {"config_name": "lat", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 81932667, "num_examples": 17214}], "download_size": 19567252, "dataset_size": 81932667}, {"config_name": "lav", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 21219584, "num_examples": 7548}], "download_size": 5048680, "dataset_size": 21219584}, {"config_name": "lit", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 5287268, "num_examples": 1458}], "download_size": 1191554, "dataset_size": 5287268}, {"config_name": "liv", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 642166, "num_examples": 203}], "download_size": 141467, "dataset_size": 642166}, {"config_name": "lld", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 1240257, "num_examples": 180}], "download_size": 278592, "dataset_size": 1240257}, {"config_name": "lud", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "mikhailovskoye", "num_bytes": 11361, "num_examples": 2}, {"name": "new_written", "num_bytes": 35132, "num_examples": 94}, {"name": "southern_ludian_svjatozero", "num_bytes": 57276, "num_examples": 71}], "download_size": 14697, "dataset_size": 103769}, {"config_name": "mkd", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 27800390, "num_examples": 10313}], "download_size": 8157589, "dataset_size": 27800390}, {"config_name": "mlt", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 604577, "num_examples": 112}], "download_size": 124584, "dataset_size": 604577}, {"config_name": "mwf", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 172890, "num_examples": 29}], "download_size": 25077, "dataset_size": 172890}, {"config_name": "nap", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 293699, "num_examples": 40}], "download_size": 64163, "dataset_size": 293699}, {"config_name": "nav", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 2051393, "num_examples": 674}], "download_size": 523673, "dataset_size": 2051393}, {"config_name": "nds", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train"}], "download_size": 2, "dataset_size": 0}, {"config_name": "nld", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 8813867, "num_examples": 4993}], "download_size": 1874427, "dataset_size": 8813867}, {"config_name": "nno", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 2704566, "num_examples": 4689}], "download_size": 420695, "dataset_size": 2704566}, {"config_name": "nob", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 3359706, "num_examples": 5527}], "download_size": 544432, "dataset_size": 3359706}, {"config_name": "oci", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 1327716, "num_examples": 174}], "download_size": 276611, "dataset_size": 1327716}, {"config_name": "olo", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "kotkozero", "num_bytes": 7682, "num_examples": 5}, {"name": "new_written", "num_bytes": 11158424, "num_examples": 15293}, {"name": "syamozero", "num_bytes": 6379, "num_examples": 2}, {"name": "vedlozero", "num_bytes": 6120, "num_examples": 1}, {"name": "vidlitsa", "num_bytes": 54363, "num_examples": 3}], "download_size": 2130154, "dataset_size": 11232968}, {"config_name": "osx", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 3500590, "num_examples": 863}], "download_size": 759997, "dataset_size": 3500590}, {"config_name": "pol", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 30855235, "num_examples": 10185}], "download_size": 6666266, "dataset_size": 30855235}, {"config_name": "por", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 48530106, "num_examples": 4001}], "download_size": 10982524, "dataset_size": 48530106}, {"config_name": "pus", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 1176421, "num_examples": 395}], "download_size": 297043, "dataset_size": 1176421}, {"config_name": "que", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 27823298, "num_examples": 1006}], "download_size": 6742890, "dataset_size": 27823298}, {"config_name": "ron", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 13187957, "num_examples": 4405}], "download_size": 2990521, "dataset_size": 13187957}, {"config_name": "rus", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 77484460, "num_examples": 28068}], "download_size": 25151401, "dataset_size": 77484460}, {"config_name": "san", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 5500001, "num_examples": 917}], "download_size": 1788739, "dataset_size": 5500001}, {"config_name": "sga", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 190479, "num_examples": 49}], "download_size": 43469, "dataset_size": 190479}, {"config_name": "slv", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 9071547, "num_examples": 2535}], "download_size": 1911039, "dataset_size": 9071547}, {"config_name": "sme", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 9764653, "num_examples": 2103}], "download_size": 2050015, "dataset_size": 9764653}, {"config_name": "spa", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 61472202, "num_examples": 5460}], "download_size": 14386131, "dataset_size": 61472202}, {"config_name": "sqi", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 5422400, "num_examples": 589}], "download_size": 1261468, "dataset_size": 5422400}, {"config_name": "swc", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 1694529, "num_examples": 100}], "download_size": 414624, "dataset_size": 1694529}, {"config_name": "swe", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 12897827, "num_examples": 10553}], "download_size": 2709960, "dataset_size": 12897827}, {"config_name": "syc", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 553392, "num_examples": 160}], "download_size": 130000, "dataset_size": 553392}, {"config_name": "tat", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 1203356, "num_examples": 1283}], "download_size": 194277, "dataset_size": 1203356}, {"config_name": "tel", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 285769, "num_examples": 127}], "download_size": 95069, "dataset_size": 285769}, {"config_name": "tgk", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 25276, "num_examples": 75}], "download_size": 2366, "dataset_size": 25276}, {"config_name": "tuk", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 127712, "num_examples": 68}], "download_size": 20540, "dataset_size": 127712}, {"config_name": "tur", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 44723850, "num_examples": 3579}], "download_size": 11552946, "dataset_size": 44723850}, {"config_name": "ukr", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 3299187, "num_examples": 1493}], "download_size": 870660, "dataset_size": 3299187}, {"config_name": "urd", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 2197237, "num_examples": 182}], "download_size": 685613, "dataset_size": 2197237}, {"config_name": "uzb", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 196802, "num_examples": 15}], "download_size": 41921, "dataset_size": 196802}, {"config_name": "vec", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 2892987, "num_examples": 368}], "download_size": 615931, "dataset_size": 2892987}, {"config_name": "vep", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "central_eastern", "num_bytes": 500981, "num_examples": 65}, {"name": "central_western", "num_bytes": 2527618, "num_examples": 111}, {"name": "new_written", "num_bytes": 79899484, "num_examples": 9304}, {"name": "northern", "num_bytes": 175242, "num_examples": 21}, {"name": "southern", "num_bytes": 206289, "num_examples": 17}], "download_size": 20131151, "dataset_size": 83309614}, {"config_name": "vot", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 217663, "num_examples": 55}], "download_size": 37179, "dataset_size": 217663}, {"config_name": "xcl", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 16856327, "num_examples": 4300}], "download_size": 4950513, "dataset_size": 16856327}, {"config_name": "xno", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 48938, "num_examples": 5}], "download_size": 9641, "dataset_size": 48938}, {"config_name": "yid", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 1409582, "num_examples": 803}], "download_size": 429391, "dataset_size": 1409582}, {"config_name": "zul", "features": [{"name": "lemma", "dtype": "string"}, {"name": "forms", "sequence": [{"name": "word", "dtype": "string"}, {"name": "Aktionsart", "sequence": {"class_label": {"names": {"0": "STAT", "1": "DYN", "2": "TEL", "3": "ATEL", "4": "PCT", "5": "DUR", "6": "ACH", "7": "ACCMP", "8": "SEMEL", "9": "ACTY"}}}}, {"name": "Animacy", "sequence": {"class_label": {"names": {"0": "ANIM", "1": "INAN", "2": "HUM", "3": "NHUM"}}}}, {"name": "Argument_Marking", "sequence": {"class_label": {"names": {"0": "ARGNO1S", "1": "ARGNO2S", "2": "ARGNO3S", "3": "ARGNO1P", "4": "ARGNO2P", "5": "ARGNO3P", "6": "ARGAC1S", "7": "ARGAC2S", "8": "ARGAC3S", "9": "ARGAC1P", "10": "ARGAC2P", "11": "ARGAC3P", "12": "ARGAB1S", "13": "ARGAB2S", "14": "ARGAB3S", "15": "ARGAB1P", "16": "ARGAB2P", "17": "ARGAB3P", "18": "ARGER1S", "19": "ARGER2S", "20": "ARGER3S", "21": "ARGER1P", "22": "ARGER2P", "23": "ARGER3P", "24": "ARGDA1S", "25": "ARGDA2S", "26": "ARGDA3S", "27": "ARGDA1P", "28": "ARGDA2P", "29": "ARGDA3P", "30": "ARGBE1S", "31": "ARGBE2S", "32": "ARGBE3S", "33": "ARGBE1P", "34": "ARGBE2P", "35": "ARGBE3P"}}}}, {"name": "Aspect", "sequence": {"class_label": {"names": {"0": "IPFV", "1": "PFV", "2": "PRF", "3": "PROG", "4": "PROSP", "5": "ITER", "6": "HAB"}}}}, {"name": "Case", "sequence": {"class_label": {"names": {"0": "NOM", "1": "ACC", "2": "ERG", "3": "ABS", "4": "NOMS", "5": "DAT", "6": "BEN", "7": "PRP", "8": "GEN", "9": "REL", "10": "PRT", "11": "INS", "12": "COM", "13": "VOC", "14": "COMPV", "15": "EQTV", "16": "PRIV", "17": "PROPR", "18": "AVR", "19": "FRML", "20": "TRANS", "21": "BYWAY", "22": "INTER", "23": "AT", "24": "POST", "25": "IN", "26": "CIRC", "27": "ANTE", "28": "APUD", "29": "ON", "30": "ONHR", "31": "ONVR", "32": "SUB", "33": "REM", "34": "PROXM", "35": "ESS", "36": "ALL", "37": "ABL", "38": "APPRX", "39": "TERM"}}}}, {"name": "Comparison", "sequence": {"class_label": {"names": {"0": "CMPR", "1": "SPRL", "2": "AB", "3": "RL", "4": "EQT"}}}}, {"name": "Definiteness", "sequence": {"class_label": {"names": {"0": "DEF", "1": "INDF", "2": "SPEC", "3": "NSPEC"}}}}, {"name": "Deixis", "sequence": {"class_label": {"names": {"0": "PROX", "1": "MED", "2": "REMT", "3": "REF1", "4": "REF2", "5": "NOREF", "6": "PHOR", "7": "VIS", "8": "NVIS", "9": "ABV", "10": "EVEN", "11": "BEL"}}}}, {"name": "Evidentiality", "sequence": {"class_label": {"names": {"0": "FH", "1": "DRCT", "2": "SEN", "3": "VISU", "4": "NVSEN", "5": "AUD", "6": "NFH", "7": "QUOT", "8": "RPRT", "9": "HRSY", "10": "INFER", "11": "ASSUM"}}}}, {"name": "Finiteness", "sequence": {"class_label": {"names": {"0": "FIN", "1": "NFIN"}}}}, {"name": "Gender", "sequence": {"class_label": {"names": {"0": "MASC", "1": "FEM", "2": "NEUT", "3": "NAKH1", "4": "NAKH2", "5": "NAKH3", "6": "NAKH4", "7": "NAKH5", "8": "NAKH6", "9": "NAKH7", "10": "NAKH8", "11": "BANTU1", "12": "BANTU2", "13": "BANTU3", "14": "BANTU4", "15": "BANTU5", "16": "BANTU6", "17": "BANTU7", "18": "BANTU8", "19": "BANTU9", "20": "BANTU10", "21": "BANTU11", "22": "BANTU12", "23": "BANTU13", "24": "BANTU14", "25": "BANTU15", "26": "BANTU16", "27": "BANTU17", "28": "BANTU18", "29": "BANTU19", "30": "BANTU20", "31": "BANTU21", "32": "BANTU22", "33": "BANTU23"}}}}, {"name": "Information_Structure", "sequence": {"class_label": {"names": {"0": "TOP", "1": "FOC"}}}}, {"name": "Interrogativity", "sequence": {"class_label": {"names": {"0": "DECL", "1": "INT"}}}}, {"name": "Language_Specific", "sequence": {"class_label": {"names": {"0": "LGSPEC1", "1": "LGSPEC2", "2": "LGSPEC3", "3": "LGSPEC4", "4": "LGSPEC5", "5": "LGSPEC6", "6": "LGSPEC7", "7": "LGSPEC8", "8": "LGSPEC9", "9": "LGSPEC10"}}}}, {"name": "Mood", "sequence": {"class_label": {"names": {"0": "IND", "1": "SBJV", "2": "REAL", "3": "IRR", "4": "AUPRP", "5": "AUNPRP", "6": "IMP", "7": "COND", "8": "PURP", "9": "INTEN", "10": "POT", "11": "LKLY", "12": "ADM", "13": "OBLIG", "14": "DEB", "15": "PERM", "16": "DED", "17": "SIM", "18": "OPT"}}}}, {"name": "Number", "sequence": {"class_label": {"names": {"0": "SG", "1": "PL", "2": "GRPL", "3": "DU", "4": "TRI", "5": "PAUC", "6": "GRPAUC", "7": "INVN"}}}}, {"name": "Part_Of_Speech", "sequence": {"class_label": {"names": {"0": "N", "1": "PROPN", "2": "ADJ", "3": "PRO", "4": "CLF", "5": "ART", "6": "DET", "7": "V", "8": "ADV", "9": "AUX", "10": "V.PTCP", "11": "V.MSDR", "12": "V.CVB", "13": "ADP", "14": "COMP", "15": "CONJ", "16": "NUM", "17": "PART", "18": "INTJ"}}}}, {"name": "Person", "sequence": {"class_label": {"names": {"0": "0", "1": "1", "2": "2", "3": "3", "4": "4", "5": "INCL", "6": "EXCL", "7": "PRX", "8": "OBV"}}}}, {"name": "Polarity", "sequence": {"class_label": {"names": {"0": "POS", "1": "NEG"}}}}, {"name": "Politeness", "sequence": {"class_label": {"names": {"0": "INFM", "1": "FORM", "2": "ELEV", "3": "HUMB", "4": "POL", "5": "AVOID", "6": "LOW", "7": "HIGH", "8": "STELEV", "9": "STSUPR", "10": "LIT", "11": "FOREG", "12": "COL"}}}}, {"name": "Possession", "sequence": {"class_label": {"names": {"0": "ALN", "1": "NALN", "2": "PSS1S", "3": "PSS2S", "4": "PSS2SF", "5": "PSS2SM", "6": "PSS2SINFM", "7": "PSS2SFORM", "8": "PSS3S", "9": "PSS3SF", "10": "PSS3SM", "11": "PSS1D", "12": "PSS1DI", "13": "PSS1DE", "14": "PSS2D", "15": "PSS2DM", "16": "PSS2DF", "17": "PSS3D", "18": "PSS3DF", "19": "PSS3DM", "20": "PSS1P", "21": "PSS1PI", "22": "PSS1PE", "23": "PSS2P", "24": "PSS2PF", "25": "PSS2PM", "26": "PSS3PF", "27": "PSS3PM"}}}}, {"name": "Switch_Reference", "sequence": {"class_label": {"names": {"0": "SS", "1": "SSADV", "2": "DS", "3": "DSADV", "4": "OR", "5": "SIMMA", "6": "SEQMA", "7": "LOG"}}}}, {"name": "Tense", "sequence": {"class_label": {"names": {"0": "PRS", "1": "PST", "2": "FUT", "3": "IMMED", "4": "HOD", "5": "1DAY", "6": "RCT", "7": "RMT"}}}}, {"name": "Valency", "sequence": {"class_label": {"names": {"0": "IMPRS", "1": "INTR", "2": "TR", "3": "DITR", "4": "REFL", "5": "RECP", "6": "CAUS", "7": "APPL"}}}}, {"name": "Voice", "sequence": {"class_label": {"names": {"0": "ACT", "1": "MID", "2": "PASS", "3": "ANTIP", "4": "DIR", "5": "INV", "6": "AGFOC", "7": "PFOC", "8": "LFOC", "9": "BFOC", "10": "ACFOC", "11": "IFOC", "12": "CFOC"}}}}, {"name": "Other", "sequence": "string"}]}], "splits": [{"name": "train", "num_bytes": 7152507, "num_examples": 566}], "download_size": 1581402, "dataset_size": 7152507}]}
2023-06-08T08:28:28+00:00
[]
[ "ady", "ang", "ar", "arn", "ast", "az", "ba", "be", "bg", "bn", "bo", "br", "ca", "ckb", "crh", "cs", "csb", "cu", "cy", "da", "de", "dsb", "el", "en", "es", "et", "eu", "fa", "fi", "fo", "fr", "frm", "fro", "frr", "fur", "fy", "ga", "gal", "gd", "gmh", "gml", "got", "grc", "gv", "hai", "he", "hi", "hu", "hy", "is", "it", "izh", "ka", "kbd", "kjh", "kk", "kl", "klr", "kmr", "kn", "krl", "kw", "la", "liv", "lld", "lt", "lud", "lv", "mk", "mt", "mwf", "nap", "nb", "nds", "nl", "nn", "nv", "oc", "olo", "osx", "pl", "ps", "pt", "qu", "ro", "ru", "sa", "sga", "sh", "sl", "sme", "sq", "sv", "swc", "syc", "te", "tg", "tk", "tr", "tt", "uk", "ur", "uz", "vec", "vep", "vot", "xcl", "xno", "yi", "zu" ]
TAGS #task_categories-token-classification #task_categories-text-classification #task_ids-multi-class-classification #task_ids-multi-label-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #size_categories-1K<n<10K #size_categories-n<1K #source_datasets-original #language-Adyghe #language-Old English (ca. 450-1100) #language-Arabic #language-Mapudungun #language-Asturian #language-Azerbaijani #language-Bashkir #language-Belarusian #language-Bulgarian #language-Bengali #language-Tibetan #language-Breton #language-Catalan #language-Central Kurdish #language-Crimean Tatar #language-Czech #language-Kashubian #language-Church Slavic #language-Welsh #language-Danish #language-German #language-Lower Sorbian #language-Modern Greek (1453-) #language-English #language-Spanish #language-Estonian #language-Basque #language-Persian #language-Finnish #language-Faroese #language-French #language-Middle French (ca. 1400-1600) #language-Old French (842-ca. 1400) #language-Northern Frisian #language-Friulian #language-Western Frisian #language-Irish #language-Galolen #language-Scottish Gaelic #language-Middle High German (ca. 1050-1500) #language-Middle Low German #language-Gothic #language-Ancient Greek (to 1453) #language-Manx #language-Haida #language-Hebrew #language-Hindi #language-Hungarian #language-Armenian #language-Icelandic #language-Italian #language-Ingrian #language-Georgian #language-Kabardian #language-Khakas #language-Kazakh #language-Kalaallisut #language-Khaling #language-Northern Kurdish #language-Kannada #language-Karelian #language-Cornish #language-Latin #language-Liv #language-Ladin #language-Lithuanian #language-Ludian #language-Latvian #language-Macedonian #language-Maltese #language-Murrinh-Patha #language-Neapolitan #language-Norwegian Bokmål #language-Low German #language-Dutch #language-Norwegian Nynorsk #language-Navajo #language-Occitan (post 1500) #language-Livvi #language-Old Saxon #language-Polish #language-Pushto #language-Portuguese #language-Quechua #language-Romanian #language-Russian #language-Sanskrit #language-Old Irish (to 900) #language-Serbo-Croatian #language-Slovenian #language-Northern Sami #language-Albanian #language-Swedish #language-Congo Swahili #language-Classical Syriac #language-Telugu #language-Tajik #language-Turkmen #language-Turkish #language-Tatar #language-Ukrainian #language-Urdu #language-Uzbek #language-Venetian #language-Veps #language-Votic #language-Classical Armenian #language-Anglo-Norman #language-Yiddish #language-Zulu #license-cc-by-sa-3.0 #morphology #region-us
# Dataset Card for [Dataset Name] ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: UniMorph Homepage - Repository: List of UniMorph repositories - Paper: The Composition and Use of the Universal Morphological Feature Schema (UniMorph Schema) - Point of Contact: Arya McCarthy ### Dataset Summary The Universal Morphology (UniMorph) project is a collaborative effort to improve how NLP handles complex morphology in the world’s languages. The goal of UniMorph is to annotate morphological data in a universal schema that allows an inflected word from any language to be defined by its lexical meaning, typically carried by the lemma, and by a rendering of its inflectional form in terms of a bundle of morphological features from our schema. The specification of the schema is described in Sylak-Glassman (2016). ### Supported Tasks and Leaderboards ### Languages The current version of the UniMorph dataset covers 110 languages. ## Dataset Structure ### Data Instances Each data instance comprises of a lemma and a set of possible realizations with morphological and meaning annotations. For example: ### Data Fields Each instance in the dataset has the following fields: - 'lemma': the common lemma for all all_forms - 'forms': all annotated forms for this lemma, with: - 'word': the full word form - ['category']: a categorical variable denoting one or several tags in a category (several to represent composite tags, originally denoted with 'A+B'). The full list of categories and possible tags for each can be found here ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @yjernite for adding this dataset.
[ "# Dataset Card for [Dataset Name]", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: UniMorph Homepage\n- Repository: List of UniMorph repositories\n- Paper: The Composition and Use of the Universal Morphological Feature Schema (UniMorph Schema)\n- Point of Contact: Arya McCarthy", "### Dataset Summary\n\nThe Universal Morphology (UniMorph) project is a collaborative effort to improve how NLP handles complex morphology in the world’s languages.\nThe goal of UniMorph is to annotate morphological data in a universal schema that allows an inflected word from any language to be defined by its lexical meaning,\ntypically carried by the lemma, and by a rendering of its inflectional form in terms of a bundle of morphological features from our schema.\nThe specification of the schema is described in Sylak-Glassman (2016).", "### Supported Tasks and Leaderboards", "### Languages\n\nThe current version of the UniMorph dataset covers 110 languages.", "## Dataset Structure", "### Data Instances\n\nEach data instance comprises of a lemma and a set of possible realizations with morphological and meaning annotations. For example:", "### Data Fields\n\nEach instance in the dataset has the following fields:\n- 'lemma': the common lemma for all all_forms\n- 'forms': all annotated forms for this lemma, with:\n - 'word': the full word form\n - ['category']: a categorical variable denoting one or several tags in a category (several to represent composite tags, originally denoted with 'A+B'). The full list of categories and possible tags for each can be found here", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @yjernite for adding this dataset." ]
[ "TAGS\n#task_categories-token-classification #task_categories-text-classification #task_ids-multi-class-classification #task_ids-multi-label-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #size_categories-1K<n<10K #size_categories-n<1K #source_datasets-original #language-Adyghe #language-Old English (ca. 450-1100) #language-Arabic #language-Mapudungun #language-Asturian #language-Azerbaijani #language-Bashkir #language-Belarusian #language-Bulgarian #language-Bengali #language-Tibetan #language-Breton #language-Catalan #language-Central Kurdish #language-Crimean Tatar #language-Czech #language-Kashubian #language-Church Slavic #language-Welsh #language-Danish #language-German #language-Lower Sorbian #language-Modern Greek (1453-) #language-English #language-Spanish #language-Estonian #language-Basque #language-Persian #language-Finnish #language-Faroese #language-French #language-Middle French (ca. 1400-1600) #language-Old French (842-ca. 1400) #language-Northern Frisian #language-Friulian #language-Western Frisian #language-Irish #language-Galolen #language-Scottish Gaelic #language-Middle High German (ca. 1050-1500) #language-Middle Low German #language-Gothic #language-Ancient Greek (to 1453) #language-Manx #language-Haida #language-Hebrew #language-Hindi #language-Hungarian #language-Armenian #language-Icelandic #language-Italian #language-Ingrian #language-Georgian #language-Kabardian #language-Khakas #language-Kazakh #language-Kalaallisut #language-Khaling #language-Northern Kurdish #language-Kannada #language-Karelian #language-Cornish #language-Latin #language-Liv #language-Ladin #language-Lithuanian #language-Ludian #language-Latvian #language-Macedonian #language-Maltese #language-Murrinh-Patha #language-Neapolitan #language-Norwegian Bokmål #language-Low German #language-Dutch #language-Norwegian Nynorsk #language-Navajo #language-Occitan (post 1500) #language-Livvi #language-Old Saxon #language-Polish #language-Pushto #language-Portuguese #language-Quechua #language-Romanian #language-Russian #language-Sanskrit #language-Old Irish (to 900) #language-Serbo-Croatian #language-Slovenian #language-Northern Sami #language-Albanian #language-Swedish #language-Congo Swahili #language-Classical Syriac #language-Telugu #language-Tajik #language-Turkmen #language-Turkish #language-Tatar #language-Ukrainian #language-Urdu #language-Uzbek #language-Venetian #language-Veps #language-Votic #language-Classical Armenian #language-Anglo-Norman #language-Yiddish #language-Zulu #license-cc-by-sa-3.0 #morphology #region-us \n", "# Dataset Card for [Dataset Name]", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: UniMorph Homepage\n- Repository: List of UniMorph repositories\n- Paper: The Composition and Use of the Universal Morphological Feature Schema (UniMorph Schema)\n- Point of Contact: Arya McCarthy", "### Dataset Summary\n\nThe Universal Morphology (UniMorph) project is a collaborative effort to improve how NLP handles complex morphology in the world’s languages.\nThe goal of UniMorph is to annotate morphological data in a universal schema that allows an inflected word from any language to be defined by its lexical meaning,\ntypically carried by the lemma, and by a rendering of its inflectional form in terms of a bundle of morphological features from our schema.\nThe specification of the schema is described in Sylak-Glassman (2016).", "### Supported Tasks and Leaderboards", "### Languages\n\nThe current version of the UniMorph dataset covers 110 languages.", "## Dataset Structure", "### Data Instances\n\nEach data instance comprises of a lemma and a set of possible realizations with morphological and meaning annotations. For example:", "### Data Fields\n\nEach instance in the dataset has the following fields:\n- 'lemma': the common lemma for all all_forms\n- 'forms': all annotated forms for this lemma, with:\n - 'word': the full word form\n - ['category']: a categorical variable denoting one or several tags in a category (several to represent composite tags, originally denoted with 'A+B'). The full list of categories and possible tags for each can be found here", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @yjernite for adding this dataset." ]
[ 840, 10, 120, 58, 129, 10, 20, 6, 35, 121, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 17 ]
[ "passage: ", "passage: TAGS\n#task_categories-token-classification #task_categories-text-classification #task_ids-multi-class-classification #task_ids-multi-label-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #size_categories-1K<n<10K #size_categories-n<1K #source_datasets-original #language-Adyghe #language-Old English (ca. 450-1100) #language-Arabic #language-Mapudungun #language-Asturian #language-Azerbaijani #language-Bashkir #language-Belarusian #language-Bulgarian #language-Bengali #language-Tibetan #language-Breton #language-Catalan #language-Central Kurdish #language-Crimean Tatar #language-Czech #language-Kashubian #language-Church Slavic #language-Welsh #language-Danish #language-German #language-Lower Sorbian #language-Modern Greek (1453-) #language-English #language-Spanish #language-Estonian #language-Basque #language-Persian #language-Finnish #language-Faroese #language-French #language-Middle French (ca. 1400-1600) #language-Old French (842-ca. 1400) #language-Northern Frisian #language-Friulian #language-Western Frisian #language-Irish #language-Galolen #language-Scottish Gaelic #language-Middle High German (ca. 1050-1500) #language-Middle Low German #language-Gothic #language-Ancient Greek (to 1453) #language-Manx #language-Haida #language-Hebrew #language-Hindi #language-Hungarian #language-Armenian #language-Icelandic #language-Italian #language-Ingrian #language-Georgian #language-Kabardian #language-Khakas #language-Kazakh #language-Kalaallisut #language-Khaling #language-Northern Kurdish #language-Kannada #language-Karelian #language-Cornish #language-Latin #language-Liv #language-Ladin #language-Lithuanian #language-Ludian #language-Latvian #language-Macedonian #language-Maltese #language-Murrinh-Patha #language-Neapolitan #language-Norwegian Bokmål #language-Low German #language-Dutch #language-Norwegian Nynorsk #language-Navajo #language-Occitan (post 1500) #language-Livvi #language-Old Saxon #language-Polish #language-Pushto #language-Portuguese #language-Quechua #language-Romanian #language-Russian #language-Sanskrit #language-Old Irish (to 900) #language-Serbo-Croatian #language-Slovenian #language-Northern Sami #language-Albanian #language-Swedish #language-Congo Swahili #language-Classical Syriac #language-Telugu #language-Tajik #language-Turkmen #language-Turkish #language-Tatar #language-Ukrainian #language-Urdu #language-Uzbek #language-Venetian #language-Veps #language-Votic #language-Classical Armenian #language-Anglo-Norman #language-Yiddish #language-Zulu #license-cc-by-sa-3.0 #morphology #region-us \n# Dataset Card for [Dataset Name]## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: UniMorph Homepage\n- Repository: List of UniMorph repositories\n- Paper: The Composition and Use of the Universal Morphological Feature Schema (UniMorph Schema)\n- Point of Contact: Arya McCarthy### Dataset Summary\n\nThe Universal Morphology (UniMorph) project is a collaborative effort to improve how NLP handles complex morphology in the world’s languages.\nThe goal of UniMorph is to annotate morphological data in a universal schema that allows an inflected word from any language to be defined by its lexical meaning,\ntypically carried by the lemma, and by a rendering of its inflectional form in terms of a bundle of morphological features from our schema.\nThe specification of the schema is described in Sylak-Glassman (2016).### Supported Tasks and Leaderboards### Languages\n\nThe current version of the UniMorph dataset covers 110 languages.## Dataset Structure### Data Instances\n\nEach data instance comprises of a lemma and a set of possible realizations with morphological and meaning annotations. For example:" ]
27003997aad4f0ad34fb06ab1a0841b7f23f6dbe
# Dataset Card for Bend the Truth (Urdu Fake News) ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Github](https://github.com/MaazAmjad/Datasets-for-Urdu-news/) - **Repository:** [Github](https://github.com/MaazAmjad/Datasets-for-Urdu-news/) - **Paper:** - **Leaderboard:** - **Point of Contact:** [Maaz Amjad](https://github.com/MaazAmjad) ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields - news: a string in urdu - label: the label indicating whethere the provided news is real or fake. - category: The intent of the news being presented. The available 5 classes are Sports, Health, Technology, Entertainment, and Business. ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@chaitnayabasava](https://github.com/chaitnayabasava) for adding this dataset.
urdu_fake_news
[ "task_categories:text-classification", "task_ids:fact-checking", "task_ids:intent-classification", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:n<1K", "source_datasets:original", "language:ur", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["expert-generated"], "language": ["ur"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["n<1K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["fact-checking", "intent-classification"], "pretty_name": "Bend the Truth (Urdu Fake News)", "dataset_info": {"features": [{"name": "news", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "Fake", "1": "Real"}}}}, {"name": "category", "dtype": {"class_label": {"names": {"0": "bus", "1": "hlth", "2": "sp", "3": "tch", "4": "sbz"}}}}], "splits": [{"name": "train", "num_bytes": 1762905, "num_examples": 638}, {"name": "test", "num_bytes": 799587, "num_examples": 262}], "download_size": 1042653, "dataset_size": 2562492}}
2024-01-18T11:17:48+00:00
[]
[ "ur" ]
TAGS #task_categories-text-classification #task_ids-fact-checking #task_ids-intent-classification #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-n<1K #source_datasets-original #language-Urdu #license-unknown #region-us
# Dataset Card for Bend the Truth (Urdu Fake News) ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: Github - Repository: Github - Paper: - Leaderboard: - Point of Contact: Maaz Amjad ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances ### Data Fields - news: a string in urdu - label: the label indicating whethere the provided news is real or fake. - category: The intent of the news being presented. The available 5 classes are Sports, Health, Technology, Entertainment, and Business. ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @chaitnayabasava for adding this dataset.
[ "# Dataset Card for Bend the Truth (Urdu Fake News)", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Github\n- Repository: Github\n- Paper:\n- Leaderboard:\n- Point of Contact: Maaz Amjad", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- news: a string in urdu\n- label: the label indicating whethere the provided news is real or fake.\n- category: The intent of the news being presented. The available 5 classes are Sports, Health, Technology, Entertainment, and Business.", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @chaitnayabasava for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #task_ids-fact-checking #task_ids-intent-classification #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-n<1K #source_datasets-original #language-Urdu #license-unknown #region-us \n", "# Dataset Card for Bend the Truth (Urdu Fake News)", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Github\n- Repository: Github\n- Paper:\n- Leaderboard:\n- Point of Contact: Maaz Amjad", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- news: a string in urdu\n- label: the label indicating whethere the provided news is real or fake.\n- category: The intent of the news being presented. The available 5 classes are Sports, Health, Technology, Entertainment, and Business.", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @chaitnayabasava for adding this dataset." ]
[ 100, 16, 120, 34, 6, 10, 4, 6, 6, 57, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 20 ]
[ "passage: TAGS\n#task_categories-text-classification #task_ids-fact-checking #task_ids-intent-classification #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-n<1K #source_datasets-original #language-Urdu #license-unknown #region-us \n# Dataset Card for Bend the Truth (Urdu Fake News)## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: Github\n- Repository: Github\n- Paper:\n- Leaderboard:\n- Point of Contact: Maaz Amjad### Dataset Summary### Supported Tasks and Leaderboards### Languages## Dataset Structure### Data Instances### Data Fields\n\n- news: a string in urdu\n- label: the label indicating whethere the provided news is real or fake.\n- category: The intent of the news being presented. The available 5 classes are Sports, Health, Technology, Entertainment, and Business.### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators### Licensing Information### Contributions\n\nThanks to @chaitnayabasava for adding this dataset." ]
2ca5cf14c6657a71347a9b75de460b218cf8a1ae
# Dataset Card for Urdu Sentiment Corpus (USC) ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Github](https://github.com/MuhammadYaseenKhan/Urdu-Sentiment-Corpus) - **Repository:** [Github](https://github.com/MuhammadYaseenKhan/Urdu-Sentiment-Corpus) - **Paper:** [IEEE](https://ieeexplore.ieee.org/abstract/document/9080043) - **Leaderboard:** - **Point of Contact:** [Muhammad Yaseen Khan](https://github.com/MuhammadYaseenKhan) ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields - sentences: The Urdu tweet - sentiment: The sentiment that was exhibited in the tweet, which can be Positive(P) or Negative(N) or Objective(O). ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@chaitnayabasava](https://github.com/chaitnayabasava) for adding this dataset.
urdu_sentiment_corpus
[ "task_categories:text-classification", "task_ids:sentiment-classification", "annotations_creators:expert-generated", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:ur", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["crowdsourced"], "language": ["ur"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["sentiment-classification"], "paperswithcode_id": "urdu-sentiment-corpus", "pretty_name": "Urdu Sentiment Corpus (USC)", "dataset_info": {"features": [{"name": "sentence", "dtype": "string"}, {"name": "sentiment", "dtype": {"class_label": {"names": {"0": "P", "1": "N", "2": "O"}}}}], "splits": [{"name": "train", "num_bytes": 161190, "num_examples": 1000}], "download_size": 51583, "dataset_size": 161190}}
2024-01-18T11:17:50+00:00
[]
[ "ur" ]
TAGS #task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #language_creators-crowdsourced #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Urdu #license-unknown #region-us
# Dataset Card for Urdu Sentiment Corpus (USC) ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: Github - Repository: Github - Paper: IEEE - Leaderboard: - Point of Contact: Muhammad Yaseen Khan ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances ### Data Fields - sentences: The Urdu tweet - sentiment: The sentiment that was exhibited in the tweet, which can be Positive(P) or Negative(N) or Objective(O). ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @chaitnayabasava for adding this dataset.
[ "# Dataset Card for Urdu Sentiment Corpus (USC)", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Github\n- Repository: Github\n- Paper: IEEE\n- Leaderboard:\n- Point of Contact: Muhammad Yaseen Khan", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- sentences: The Urdu tweet\n- sentiment: The sentiment that was exhibited in the tweet, which can be Positive(P) or Negative(N) or Objective(O).", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @chaitnayabasava for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #language_creators-crowdsourced #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Urdu #license-unknown #region-us \n", "# Dataset Card for Urdu Sentiment Corpus (USC)", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Github\n- Repository: Github\n- Paper: IEEE\n- Leaderboard:\n- Point of Contact: Muhammad Yaseen Khan", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- sentences: The Urdu tweet\n- sentiment: The sentiment that was exhibited in the tweet, which can be Positive(P) or Negative(N) or Objective(O).", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @chaitnayabasava for adding this dataset." ]
[ 92, 13, 120, 36, 6, 10, 4, 6, 6, 45, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 20 ]
[ "passage: TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #annotations_creators-expert-generated #language_creators-crowdsourced #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Urdu #license-unknown #region-us \n# Dataset Card for Urdu Sentiment Corpus (USC)## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: Github\n- Repository: Github\n- Paper: IEEE\n- Leaderboard:\n- Point of Contact: Muhammad Yaseen Khan### Dataset Summary### Supported Tasks and Leaderboards### Languages## Dataset Structure### Data Instances### Data Fields\n\n- sentences: The Urdu tweet\n- sentiment: The sentiment that was exhibited in the tweet, which can be Positive(P) or Negative(N) or Objective(O).### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators### Licensing Information### Contributions\n\nThanks to @chaitnayabasava for adding this dataset." ]
5d5366174451f77f6f3b432f00ba3104d6d03627
# Dataset Card for VCTK ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Edinburg DataShare](https://doi.org/10.7488/ds/2645) - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary This CSTR VCTK Corpus includes around 44-hours of speech data uttered by 110 English speakers with various accents. Each speaker reads out about 400 sentences, which were selected from a newspaper, the rainbow passage and an elicitation paragraph used for the speech accent archive. ### Supported Tasks - `automatic-speech-recognition`, `speaker-identification`: The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). - `text-to-speech`, `text-to-audio`: The dataset can also be used to train a model for Text-To-Speech (TTS). ### Languages [More Information Needed] ## Dataset Structure ### Data Instances A data point comprises the path to the audio file, called `file` and its transcription, called `text`. ``` { 'speaker_id': 'p225', 'text_id': '001', 'text': 'Please call Stella.', 'age': '23', 'gender': 'F', 'accent': 'English', 'region': 'Southern England', 'file': '/datasets/downloads/extracted/8ed7dad05dfffdb552a3699777442af8e8ed11e656feb277f35bf9aea448f49e/wav48_silence_trimmed/p225/p225_001_mic1.flac', 'audio': { 'path': '/datasets/downloads/extracted/8ed7dad05dfffdb552a3699777442af8e8ed11e656feb277f35bf9aea448f49e/wav48_silence_trimmed/p225/p225_001_mic1.flac', 'array': array([0.00485229, 0.00689697, 0.00619507, ..., 0.00811768, 0.00836182, 0.00854492], dtype=float32), 'sampling_rate': 48000 }, 'comment': '' } ``` Each audio file is a single-channel FLAC with a sample rate of 48000 Hz. ### Data Fields Each row consists of the following fields: - `speaker_id`: Speaker ID - `audio`: Audio recording - `file`: Path to audio file - `text`: Text transcription of corresponding audio - `text_id`: Text ID - `age`: Speaker's age - `gender`: Speaker's gender - `accent`: Speaker's accent - `region`: Speaker's region, if annotation exists - `comment`: Miscellaneous comments, if any ### Data Splits The dataset has no predefined splits. ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in this dataset. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Public Domain, Creative Commons Attribution 4.0 International Public License ([CC-BY-4.0](https://creativecommons.org/licenses/by/4.0/legalcode)) ### Citation Information ```bibtex @inproceedings{Veaux2017CSTRVC, title = {CSTR VCTK Corpus: English Multi-speaker Corpus for CSTR Voice Cloning Toolkit}, author = {Christophe Veaux and Junichi Yamagishi and Kirsten MacDonald}, year = 2017 } ``` ### Contributions Thanks to [@jaketae](https://github.com/jaketae) for adding this dataset.
vctk
[ "task_categories:automatic-speech-recognition", "task_categories:text-to-speech", "task_categories:text-to-audio", "annotations_creators:expert-generated", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:cc-by-4.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["automatic-speech-recognition", "text-to-speech", "text-to-audio"], "task_ids": [], "paperswithcode_id": "vctk", "pretty_name": "VCTK", "dataset_info": {"features": [{"name": "speaker_id", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 48000}}}, {"name": "file", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "text_id", "dtype": "string"}, {"name": "age", "dtype": "string"}, {"name": "gender", "dtype": "string"}, {"name": "accent", "dtype": "string"}, {"name": "region", "dtype": "string"}, {"name": "comment", "dtype": "string"}], "config_name": "main", "splits": [{"name": "train", "num_bytes": 40103111, "num_examples": 88156}], "download_size": 11747302977, "dataset_size": 40103111}, "train-eval-index": [{"config": "main", "task": "automatic-speech-recognition", "task_id": "speech_recognition", "splits": {"train_split": "train"}, "col_mapping": {"file": "path", "text": "text"}, "metrics": [{"type": "wer", "name": "WER"}, {"type": "cer", "name": "CER"}]}]}
2024-01-18T11:17:51+00:00
[]
[ "en" ]
TAGS #task_categories-automatic-speech-recognition #task_categories-text-to-speech #task_categories-text-to-audio #annotations_creators-expert-generated #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-4.0 #region-us
# Dataset Card for VCTK ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: Edinburg DataShare - Repository: - Paper: - Leaderboard: - Point of Contact: ### Dataset Summary This CSTR VCTK Corpus includes around 44-hours of speech data uttered by 110 English speakers with various accents. Each speaker reads out about 400 sentences, which were selected from a newspaper, the rainbow passage and an elicitation paragraph used for the speech accent archive. ### Supported Tasks - 'automatic-speech-recognition', 'speaker-identification': The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). - 'text-to-speech', 'text-to-audio': The dataset can also be used to train a model for Text-To-Speech (TTS). ### Languages ## Dataset Structure ### Data Instances A data point comprises the path to the audio file, called 'file' and its transcription, called 'text'. Each audio file is a single-channel FLAC with a sample rate of 48000 Hz. ### Data Fields Each row consists of the following fields: - 'speaker_id': Speaker ID - 'audio': Audio recording - 'file': Path to audio file - 'text': Text transcription of corresponding audio - 'text_id': Text ID - 'age': Speaker's age - 'gender': Speaker's gender - 'accent': Speaker's accent - 'region': Speaker's region, if annotation exists - 'comment': Miscellaneous comments, if any ### Data Splits The dataset has no predefined splits. ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in this dataset. ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information Public Domain, Creative Commons Attribution 4.0 International Public License (CC-BY-4.0) ### Contributions Thanks to @jaketae for adding this dataset.
[ "# Dataset Card for VCTK", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Edinburg DataShare\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nThis CSTR VCTK Corpus includes around 44-hours of speech data uttered by 110 English speakers with various accents. Each speaker reads out about 400 sentences, which were selected from a newspaper, the rainbow passage and an elicitation paragraph used for the speech accent archive.", "### Supported Tasks\n\n- 'automatic-speech-recognition', 'speaker-identification': The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER).\n- 'text-to-speech', 'text-to-audio': The dataset can also be used to train a model for Text-To-Speech (TTS).", "### Languages", "## Dataset Structure", "### Data Instances\n\nA data point comprises the path to the audio file, called 'file' and its transcription, called 'text'.\n\n\n\nEach audio file is a single-channel FLAC with a sample rate of 48000 Hz.", "### Data Fields\n\nEach row consists of the following fields:\n\n- 'speaker_id': Speaker ID\n- 'audio': Audio recording\n- 'file': Path to audio file\n- 'text': Text transcription of corresponding audio\n- 'text_id': Text ID\n- 'age': Speaker's age\n- 'gender': Speaker's gender\n- 'accent': Speaker's accent\n- 'region': Speaker's region, if annotation exists\n- 'comment': Miscellaneous comments, if any", "### Data Splits\n\nThe dataset has no predefined splits.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\nThe dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in this dataset.", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nPublic Domain, Creative Commons Attribution 4.0 International Public License (CC-BY-4.0)", "### Contributions\n\nThanks to @jaketae for adding this dataset." ]
[ "TAGS\n#task_categories-automatic-speech-recognition #task_categories-text-to-speech #task_categories-text-to-audio #annotations_creators-expert-generated #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-4.0 #region-us \n", "# Dataset Card for VCTK", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Edinburg DataShare\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nThis CSTR VCTK Corpus includes around 44-hours of speech data uttered by 110 English speakers with various accents. Each speaker reads out about 400 sentences, which were selected from a newspaper, the rainbow passage and an elicitation paragraph used for the speech accent archive.", "### Supported Tasks\n\n- 'automatic-speech-recognition', 'speaker-identification': The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER).\n- 'text-to-speech', 'text-to-audio': The dataset can also be used to train a model for Text-To-Speech (TTS).", "### Languages", "## Dataset Structure", "### Data Instances\n\nA data point comprises the path to the audio file, called 'file' and its transcription, called 'text'.\n\n\n\nEach audio file is a single-channel FLAC with a sample rate of 48000 Hz.", "### Data Fields\n\nEach row consists of the following fields:\n\n- 'speaker_id': Speaker ID\n- 'audio': Audio recording\n- 'file': Path to audio file\n- 'text': Text transcription of corresponding audio\n- 'text_id': Text ID\n- 'age': Speaker's age\n- 'gender': Speaker's gender\n- 'accent': Speaker's accent\n- 'region': Speaker's region, if annotation exists\n- 'comment': Miscellaneous comments, if any", "### Data Splits\n\nThe dataset has no predefined splits.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\nThe dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in this dataset.", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nPublic Domain, Creative Commons Attribution 4.0 International Public License (CC-BY-4.0)", "### Contributions\n\nThanks to @jaketae for adding this dataset." ]
[ 113, 8, 120, 28, 70, 125, 4, 6, 52, 123, 17, 5, 7, 4, 10, 10, 5, 5, 9, 40, 8, 7, 8, 7, 5, 6, 23, 17 ]
[ "passage: TAGS\n#task_categories-automatic-speech-recognition #task_categories-text-to-speech #task_categories-text-to-audio #annotations_creators-expert-generated #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-4.0 #region-us \n# Dataset Card for VCTK## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: Edinburg DataShare\n- Repository:\n- Paper:\n- Leaderboard:\n- Point of Contact:### Dataset Summary\n\nThis CSTR VCTK Corpus includes around 44-hours of speech data uttered by 110 English speakers with various accents. Each speaker reads out about 400 sentences, which were selected from a newspaper, the rainbow passage and an elicitation paragraph used for the speech accent archive.### Supported Tasks\n\n- 'automatic-speech-recognition', 'speaker-identification': The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER).\n- 'text-to-speech', 'text-to-audio': The dataset can also be used to train a model for Text-To-Speech (TTS).### Languages## Dataset Structure" ]
3cbfb2502e5e84776b4b778b020a09759f723f52
# Dataset Card for VIVOS ## Table of Contents - [Dataset Card for VIVOS](#dataset-card-for-vivos) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Initial Data Collection and Normalization](#initial-data-collection-and-normalization) - [Who are the source language producers?](#who-are-the-source-language-producers) - [Annotations](#annotations) - [Annotation process](#annotation-process) - [Who are the annotators?](#who-are-the-annotators) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://doi.org/10.5281/zenodo.7068130 - **Repository:** [Needs More Information] - **Paper:** [A non-expert Kaldi recipe for Vietnamese Speech Recognition System](https://aclanthology.org/W16-5207/) - **Leaderboard:** [Needs More Information] - **Point of Contact:** [AILAB](mailto:ailab@hcmus.edu.vn) ### Dataset Summary VIVOS is a free Vietnamese speech corpus consisting of 15 hours of recording speech prepared for Vietnamese Automatic Speech Recognition task. The corpus was prepared by AILAB, a computer science lab of VNUHCM - University of Science, with Prof. Vu Hai Quan is the head of. We publish this corpus in hope to attract more scientists to solve Vietnamese speech recognition problems. ### Supported Tasks and Leaderboards [Needs More Information] ### Languages Vietnamese ## Dataset Structure ### Data Instances A typical data point comprises the path to the audio file, called `path` and its transcription, called `sentence`. Some additional information about the speaker and the passage which contains the transcription is provided. ``` {'speaker_id': 'VIVOSSPK01', 'path': '/home/admin/.cache/huggingface/datasets/downloads/extracted/b7ded9969e09942ab65313e691e6fc2e12066192ee8527e21d634aca128afbe2/vivos/train/waves/VIVOSSPK01/VIVOSSPK01_R001.wav', 'audio': {'path': '/home/admin/.cache/huggingface/datasets/downloads/extracted/b7ded9969e09942ab65313e691e6fc2e12066192ee8527e21d634aca128afbe2/vivos/train/waves/VIVOSSPK01/VIVOSSPK01_R001.wav', 'array': array([-0.00048828, -0.00018311, -0.00137329, ..., 0.00079346, 0.00091553, 0.00085449], dtype=float32), 'sampling_rate': 16000}, 'sentence': 'KHÁCH SẠN'} ``` ### Data Fields - speaker_id: An id for which speaker (voice) made the recording - path: The path to the audio file - audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: `dataset[0]["audio"]` the audio file is automatically decoded and resampled to `dataset.features["audio"].sampling_rate`. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the `"audio"` column, *i.e.* `dataset[0]["audio"]` should **always** be preferred over `dataset["audio"][0]`. - sentence: The sentence the user was prompted to speak ### Data Splits The speech material has been subdivided into portions for train and test. Speech was recorded in a quiet environment with high quality microphone, speakers were asked to read one sentence at a time. | | Train | Test | | ---------------- | ----- | ----- | | Speakers | 46 | 19 | | Utterances | 11660 | 760 | | Duration | 14:55 | 00:45 | | Unique Syllables | 4617 | 1692 | ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in this dataset. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations Dataset provided for research purposes only. Please check dataset license for additional information. ## Additional Information ### Dataset Curators The dataset was initially prepared by AILAB, a computer science lab of VNUHCM - University of Science. ### Licensing Information Public Domain, Creative Commons Attribution NonCommercial ShareAlike v4.0 ([CC BY-NC-SA 4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/legalcode)) ### Citation Information ``` @inproceedings{luong-vu-2016-non, title = "A non-expert {K}aldi recipe for {V}ietnamese Speech Recognition System", author = "Luong, Hieu-Thi and Vu, Hai-Quan", booktitle = "Proceedings of the Third International Workshop on Worldwide Language Service Infrastructure and Second Workshop on Open Infrastructures and Analysis Frameworks for Human Language Technologies ({WLSI}/{OIAF}4{HLT}2016)", month = dec, year = "2016", address = "Osaka, Japan", publisher = "The COLING 2016 Organizing Committee", url = "https://aclanthology.org/W16-5207", pages = "51--55", } ``` ### Contributions Thanks to [@binh234](https://github.com/binh234) for adding this dataset.
vivos
[ "task_categories:automatic-speech-recognition", "annotations_creators:expert-generated", "language_creators:crowdsourced", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:vi", "license:cc-by-nc-sa-4.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["crowdsourced", "expert-generated"], "language": ["vi"], "license": ["cc-by-nc-sa-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["automatic-speech-recognition"], "task_ids": [], "pretty_name": "VIVOS", "dataset_info": {"features": [{"name": "speaker_id", "dtype": "string"}, {"name": "path", "dtype": "string"}, {"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1722002133, "num_examples": 11660}, {"name": "test", "num_bytes": 86120227, "num_examples": 760}], "download_size": 1475540500, "dataset_size": 1808122360}}
2023-06-14T07:29:21+00:00
[]
[ "vi" ]
TAGS #task_categories-automatic-speech-recognition #annotations_creators-expert-generated #language_creators-crowdsourced #language_creators-expert-generated #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Vietnamese #license-cc-by-nc-sa-4.0 #region-us
Dataset Card for VIVOS ====================== Table of Contents ----------------- * Dataset Card for VIVOS + Table of Contents + Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages + Dataset Structure - Data Instances - Data Fields - Data Splits + Dataset Creation - Curation Rationale - Source Data * Initial Data Collection and Normalization * Who are the source language producers? - Annotations * Annotation process * Who are the annotators? - Personal and Sensitive Information + Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations + Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: A non-expert Kaldi recipe for Vietnamese Speech Recognition System * Leaderboard: * Point of Contact: AILAB ### Dataset Summary VIVOS is a free Vietnamese speech corpus consisting of 15 hours of recording speech prepared for Vietnamese Automatic Speech Recognition task. The corpus was prepared by AILAB, a computer science lab of VNUHCM - University of Science, with Prof. Vu Hai Quan is the head of. We publish this corpus in hope to attract more scientists to solve Vietnamese speech recognition problems. ### Supported Tasks and Leaderboards ### Languages Vietnamese Dataset Structure ----------------- ### Data Instances A typical data point comprises the path to the audio file, called 'path' and its transcription, called 'sentence'. Some additional information about the speaker and the passage which contains the transcription is provided. ### Data Fields * speaker\_id: An id for which speaker (voice) made the recording * path: The path to the audio file * audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: 'dataset[0]["audio"]' the audio file is automatically decoded and resampled to 'dataset.features["audio"].sampling\_rate'. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the '"audio"' column, *i.e.* 'dataset[0]["audio"]' should always be preferred over 'dataset["audio"][0]'. * sentence: The sentence the user was prompted to speak ### Data Splits The speech material has been subdivided into portions for train and test. Speech was recorded in a quiet environment with high quality microphone, speakers were asked to read one sentence at a time. Train: Speakers, Test: 46 Train: Utterances, Test: 11660 Train: Duration, Test: 14:55 Train: Unique Syllables, Test: 4617 Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in this dataset. Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Dataset provided for research purposes only. Please check dataset license for additional information. Additional Information ---------------------- ### Dataset Curators The dataset was initially prepared by AILAB, a computer science lab of VNUHCM - University of Science. ### Licensing Information Public Domain, Creative Commons Attribution NonCommercial ShareAlike v4.0 (CC BY-NC-SA 4.0) ### Contributions Thanks to @binh234 for adding this dataset.
[ "### Dataset Summary\n\n\nVIVOS is a free Vietnamese speech corpus consisting of 15 hours of recording speech prepared for Vietnamese Automatic Speech Recognition task.\n\n\nThe corpus was prepared by AILAB, a computer science lab of VNUHCM - University of Science, with Prof. Vu Hai Quan is the head of.\n\n\nWe publish this corpus in hope to attract more scientists to solve Vietnamese speech recognition problems.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nVietnamese\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nA typical data point comprises the path to the audio file, called 'path' and its transcription, called 'sentence'. Some additional information about the speaker and the passage which contains the transcription is provided.", "### Data Fields\n\n\n* speaker\\_id: An id for which speaker (voice) made the recording\n* path: The path to the audio file\n* audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: 'dataset[0][\"audio\"]' the audio file is automatically decoded and resampled to 'dataset.features[\"audio\"].sampling\\_rate'. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the '\"audio\"' column, *i.e.* 'dataset[0][\"audio\"]' should always be preferred over 'dataset[\"audio\"][0]'.\n* sentence: The sentence the user was prompted to speak", "### Data Splits\n\n\nThe speech material has been subdivided into portions for train and test.\n\n\nSpeech was recorded in a quiet environment with high quality microphone, speakers were asked to read one sentence at a time.\n\n\nTrain: Speakers, Test: 46\nTrain: Utterances, Test: 11660\nTrain: Duration, Test: 14:55\nTrain: Unique Syllables, Test: 4617\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nThe dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in this dataset.\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nDataset provided for research purposes only. Please check dataset license for additional information.\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThe dataset was initially prepared by AILAB, a computer science lab of VNUHCM - University of Science.", "### Licensing Information\n\n\nPublic Domain, Creative Commons Attribution NonCommercial ShareAlike v4.0 (CC BY-NC-SA 4.0)", "### Contributions\n\n\nThanks to @binh234 for adding this dataset." ]
[ "TAGS\n#task_categories-automatic-speech-recognition #annotations_creators-expert-generated #language_creators-crowdsourced #language_creators-expert-generated #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Vietnamese #license-cc-by-nc-sa-4.0 #region-us \n", "### Dataset Summary\n\n\nVIVOS is a free Vietnamese speech corpus consisting of 15 hours of recording speech prepared for Vietnamese Automatic Speech Recognition task.\n\n\nThe corpus was prepared by AILAB, a computer science lab of VNUHCM - University of Science, with Prof. Vu Hai Quan is the head of.\n\n\nWe publish this corpus in hope to attract more scientists to solve Vietnamese speech recognition problems.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nVietnamese\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nA typical data point comprises the path to the audio file, called 'path' and its transcription, called 'sentence'. Some additional information about the speaker and the passage which contains the transcription is provided.", "### Data Fields\n\n\n* speaker\\_id: An id for which speaker (voice) made the recording\n* path: The path to the audio file\n* audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: 'dataset[0][\"audio\"]' the audio file is automatically decoded and resampled to 'dataset.features[\"audio\"].sampling\\_rate'. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the '\"audio\"' column, *i.e.* 'dataset[0][\"audio\"]' should always be preferred over 'dataset[\"audio\"][0]'.\n* sentence: The sentence the user was prompted to speak", "### Data Splits\n\n\nThe speech material has been subdivided into portions for train and test.\n\n\nSpeech was recorded in a quiet environment with high quality microphone, speakers were asked to read one sentence at a time.\n\n\nTrain: Speakers, Test: 46\nTrain: Utterances, Test: 11660\nTrain: Duration, Test: 14:55\nTrain: Unique Syllables, Test: 4617\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nThe dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in this dataset.\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nDataset provided for research purposes only. Please check dataset license for additional information.\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThe dataset was initially prepared by AILAB, a computer science lab of VNUHCM - University of Science.", "### Licensing Information\n\n\nPublic Domain, Creative Commons Attribution NonCommercial ShareAlike v4.0 (CC BY-NC-SA 4.0)", "### Contributions\n\n\nThanks to @binh234 for adding this dataset." ]
[ 105, 87, 10, 13, 52, 217, 91, 7, 4, 10, 10, 5, 5, 9, 50, 7, 8, 32, 30, 29, 18 ]
[ "passage: TAGS\n#task_categories-automatic-speech-recognition #annotations_creators-expert-generated #language_creators-crowdsourced #language_creators-expert-generated #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Vietnamese #license-cc-by-nc-sa-4.0 #region-us \n### Dataset Summary\n\n\nVIVOS is a free Vietnamese speech corpus consisting of 15 hours of recording speech prepared for Vietnamese Automatic Speech Recognition task.\n\n\nThe corpus was prepared by AILAB, a computer science lab of VNUHCM - University of Science, with Prof. Vu Hai Quan is the head of.\n\n\nWe publish this corpus in hope to attract more scientists to solve Vietnamese speech recognition problems.### Supported Tasks and Leaderboards### Languages\n\n\nVietnamese\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nA typical data point comprises the path to the audio file, called 'path' and its transcription, called 'sentence'. Some additional information about the speaker and the passage which contains the transcription is provided.### Data Fields\n\n\n* speaker\\_id: An id for which speaker (voice) made the recording\n* path: The path to the audio file\n* audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: 'dataset[0][\"audio\"]' the audio file is automatically decoded and resampled to 'dataset.features[\"audio\"].sampling\\_rate'. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the '\"audio\"' column, *i.e.* 'dataset[0][\"audio\"]' should always be preferred over 'dataset[\"audio\"][0]'.\n* sentence: The sentence the user was prompted to speak" ]
d0454a3e88ede56554bf1c34a98c0ca0f9442983
# Dataset Card for WebNLG ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [WebNLG challenge website](https://webnlg-challenge.loria.fr/) - **Repository:** [WebNLG GitLab repository](https://gitlab.com/shimorina/webnlg-dataset/-/tree/master/) - **Paper:** [Creating Training Corpora for NLG Micro-Planning](https://www.aclweb.org/anthology/P17-1017.pdf) - **Leaderboard:** [WebNLG leaderboards](https://gerbil-nlg.dice-research.org/gerbil/webnlg2020results) - **Point of Contact:** [anastasia.shimorina@loria.fr](anastasia.shimorina@loria.fr) ### Dataset Summary The WebNLG challenge consists in mapping data to text. The training data consists of Data/Text pairs where the data is a set of triples extracted from DBpedia and the text is a verbalisation of these triples. For instance, given the 3 DBpedia triples shown in (a), the aim is to generate a text such as (b). ``` a. (John_E_Blaha birthDate 1942_08_26) (John_E_Blaha birthPlace San_Antonio) (John_E_Blaha occupation Fighter_pilot) b. John E Blaha, born in San Antonio on 1942-08-26, worked as a fighter pilot ``` As the example illustrates, the task involves specific NLG subtasks such as sentence segmentation (how to chunk the input data into sentences), lexicalisation (of the DBpedia properties), aggregation (how to avoid repetitions) and surface realisation (how to build a syntactically correct and natural sounding text). ### Supported Tasks and Leaderboards The dataset supports a Structured to Text task which requires a model takes a set of RDF (Resource Description Format) triples from a database (DBpedia) of the form (subject, property, object) as input and write out a natural language sentence expressing the information contained in the triples. The dataset has supportd two challenges: the [WebNLG2017](https://www.aclweb.org/anthology/W17-3518/) and [WebNLG2020](https://gerbil-nlg.dice-research.org/gerbil/webnlg2020results) challenge. Results were ordered by their [METEOR](https://huggingface.co/metrics/meteor) to the reference, but the leaderboards report a range of other metrics including [BLEU](https://huggingface.co/metrics/bleu), [BERTscore](https://huggingface.co/metrics/bertscore), and [BLEURT](https://huggingface.co/metrics/bleurt). The v3 release (`release_v3.0_en`, `release_v3.0_ru`) for the WebNLG2020 challenge also supports a semantic `parsing` task. ### Languages All releases contain English (`en`) data. The v3 release (`release_v3.0_ru`) also contains Russian (`ru`) examples. ## Dataset Structure ### Data Instances A typical example contains the original RDF triples in the set, a modified version which presented to crowd workers, and a set of possible verbalizations for this set of triples: ``` {'2017_test_category': '', 'category': 'Politician', 'eid': 'Id10', 'lex': {'comment': ['good', 'good', 'good'], 'lid': ['Id1', 'Id2', 'Id3'], 'text': ['World War II had Chiang Kai-shek as a commander and United States Army soldier Abner W. Sibal.', 'Abner W. Sibal served in the United States Army during the Second World War and during that war Chiang Kai-shek was one of the commanders.', 'Abner W. Sibal, served in the United States Army and fought in World War II, one of the commanders of which, was Chiang Kai-shek.']}, 'modified_triple_sets': {'mtriple_set': [['Abner_W._Sibal | battle | World_War_II', 'World_War_II | commander | Chiang_Kai-shek', 'Abner_W._Sibal | militaryBranch | United_States_Army']]}, 'original_triple_sets': {'otriple_set': [['Abner_W._Sibal | battles | World_War_II', 'World_War_II | commander | Chiang_Kai-shek', 'Abner_W._Sibal | branch | United_States_Army'], ['Abner_W._Sibal | militaryBranch | United_States_Army', 'Abner_W._Sibal | battles | World_War_II', 'World_War_II | commander | Chiang_Kai-shek']]}, 'shape': '(X (X) (X (X)))', 'shape_type': 'mixed', 'size': 3} ``` ### Data Fields The following fields can be found in the instances: - `category`: the category of the DBpedia entities present in the RDF triples. - `eid`: an example ID, only unique per split per category. - `size`: number of RDF triples in the set. - `shape`: (since v2) Each set of RDF-triples is a tree, which is characterised by its shape and shape type. `shape` is a string representation of the tree with nested parentheses where X is a node (see [Newick tree format](https://en.wikipedia.org/wiki/Newick_format)) - `shape_type`: (since v2) is a type of the tree shape, which can be: `chain` (the object of one triple is the subject of the other); `sibling` (triples with a shared subject); `mixed` (both chain and sibling types present). - `test_category`: (for `webnlg_challenge_2017` and `v3`) tells whether the set of RDF triples was present in the training set or not. Several splits of the test set are available: with and without references, and for RDF-to-text generation / for semantic parsing. - `lex`: the lexicalizations, with: - `text`: the text to be predicted. - `lid`: a lexicalization ID, unique per example. - `comment`: the lexicalizations were rated by crowd workers are either `good` or `bad` - `lang`: (for `release_v3.0_ru`) the language used because original English texts were kept in the Russian version. Russian data has additional optional fields comparing to English: - `dbpedialinks`: RDF triples extracted from DBpedia between English and Russian entities by means of the property `sameAs`. - `links`: RDF triples created manually for some entities to serve as pointers to translators. There are two types of them: * with `sameAs` (`Spaniards | sameAs | испанцы`) * with `includes` (`Tomatoes, guanciale, cheese, olive oil | includes | гуанчиале`). Those were mostly created for string literals to translate some parts of them. ### Data Splits For `v3.0` releases: | English (v3.0) | Train | Dev | Test (data-to-text) | |-----------------|--------|-------|-------| | **triple sets** | 13,211 | 1,667 | 1,779 | | **texts** | 35,426 | 4,464 | 5,150 | |**properties** | 372 | 290 | 220 | | Russian (v3.0) | Train | Dev | Test (data-to-text) | |-----------------|--------|-------|---------------------| | **triple sets** | 5,573 | 790 | 1,102 | | **texts** | 14,239 | 2,026 | 2,780 | |**properties** | 226 | 115 | 192 | ## Dataset Creation ### Curation Rationale The WebNLG dataset was created to promote the development _(i)_ of RDF verbalisers and _(ii)_ of microplanners able to handle a wide range of linguistic constructions. The dataset aims at covering knowledge in different domains ("categories"). The same properties and entities can appear in several categories. ### Source Data The data was compiled from raw DBpedia triples. [This paper](https://www.aclweb.org/anthology/C16-1141/) explains how the triples were selected. #### Initial Data Collection and Normalization Initial triples extracted from DBpedia were modified in several ways. See [official documentation](https://webnlg-challenge.loria.fr/docs/) for the most frequent changes that have been made. An original tripleset and a modified tripleset usually represent a one-to-one mapping. However, there are cases with many-to-one mappings when several original triplesets are mapped to one modified tripleset. Entities that served as roots of RDF trees are listed in [this file](https://gitlab.com/shimorina/webnlg-dataset/-/blob/master/supplementary/entities_dict.json). The English WebNLG 2020 dataset (v3.0) for training comprises data-text pairs for 16 distinct DBpedia categories: - The 10 seen categories used in the 2017 version: Airport, Astronaut, Building, City, ComicsCharacter, Food, Monument, SportsTeam, University, and WrittenWork. - The 5 unseen categories of 2017, which are now part of the seen data: Athlete, Artist, CelestialBody, MeanOfTransportation, Politician. - 1 new category: Company. The Russian dataset (v3.0) comprises data-text pairs for 9 distinct categories: Airport, Astronaut, Building, CelestialBody, ComicsCharacter, Food, Monument, SportsTeam, and University. #### Who are the source language producers? There are no source texts, all textual material was compiled during the annotation process. ### Annotations #### Annotation process Annotators were first asked to create sentences that verbalise single triples. In a second round, annotators were asked to combine single-triple sentences together into sentences that cover 2 triples. And so on until 7 triples. Quality checks were performed to ensure the quality of the annotations. See Section 3.3 in [the dataset paper](https://www.aclweb.org/anthology/P17-1017.pdf). Russian data was translated from English with an MT system and then was post-edited by crowdworkers. See Section 2.2 of [this paper](https://webnlg-challenge.loria.fr/files/2020.webnlg-papers.7.pdf). #### Who are the annotators? All references were collected through crowdsourcing platforms (CrowdFlower/Figure 8 and Amazon Mechanical Turk). For Russian, post-editing was done using the Yandex.Toloka crowdsourcing platform. ### Personal and Sensitive Information Neither the dataset as published or the annotation process involves the collection or sharing of any kind of personal / demographic information. ## Considerations for Using the Data ### Social Impact of Dataset We do not foresee any negative social impact in particular from this dataset or task. Positive outlooks: Being able to generate good quality text from RDF data would permit, e.g., making this data more accessible to lay users, enriching existing text with information drawn from knowledge bases such as DBpedia or describing, comparing and relating entities present in these knowledge bases. ### Discussion of Biases This dataset is created using DBpedia RDF triples which naturally exhibit biases that have been found to exist in Wikipedia such as some forms of, e.g., gender bias. The choice of [entities](https://gitlab.com/shimorina/webnlg-dataset/-/blob/master/supplementary/entities_dict.json), described by RDF trees, was not controlled. As such, they may contain gender biases; for instance, all the astronauts described by RDF triples are male. Hence, in texts, pronouns _he/him/his_ occur more often. Similarly, entities can be related to the Western culture more often than to other cultures. ### Other Known Limitations The quality of the crowdsourced references is limited, in particular in terms of fluency/naturalness of the collected texts. Russian data was machine-translated and then post-edited by crowdworkers, so some examples may still exhibit issues related to bad translations. ## Additional Information ### Dataset Curators The principle curator of the dataset is Anastasia Shimorina (Université de Lorraine / LORIA, France). Throughout the WebNLG releases, several people contributed to their construction: Claire Gardent (CNRS / LORIA, France), Shashi Narayan (Google, UK), Laura Perez-Beltrachini (University of Edinburgh, UK), Elena Khasanova, and Thiago Castro Ferreira (Federal University of Minas Gerais, Brazil). The dataset construction was funded by the French National Research Agency (ANR). ### Licensing Information The dataset uses the `cc-by-nc-sa-4.0` license. The source DBpedia project uses the `cc-by-sa-3.0` and `gfdl-1.1` licenses. ### Citation Information - If you use the WebNLG corpus, cite: ``` @inproceedings{web_nlg, author = {Claire Gardent and Anastasia Shimorina and Shashi Narayan and Laura Perez{-}Beltrachini}, editor = {Regina Barzilay and Min{-}Yen Kan}, title = {Creating Training Corpora for {NLG} Micro-Planners}, booktitle = {Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, {ACL} 2017, Vancouver, Canada, July 30 - August 4, Volume 1: Long Papers}, pages = {179--188}, publisher = {Association for Computational Linguistics}, year = {2017}, url = {https://doi.org/10.18653/v1/P17-1017}, doi = {10.18653/v1/P17-1017} } ``` - If you use `release_v2_constrained` in particular, cite: ``` @InProceedings{shimorina2018handling, author = "Shimorina, Anastasia and Gardent, Claire", title = "Handling Rare Items in Data-to-Text Generation", booktitle = "Proceedings of the 11th International Conference on Natural Language Generation", year = "2018", publisher = "Association for Computational Linguistics", pages = "360--370", location = "Tilburg University, The Netherlands", url = "http://aclweb.org/anthology/W18-6543" } ``` ### Contributions Thanks to [@Shimorina](https://github.com/Shimorina), [@yjernite](https://github.com/yjernite) for adding this dataset.
web_nlg
[ "task_categories:tabular-to-text", "task_ids:rdf-to-text", "annotations_creators:found", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:extended|other-db_pedia", "source_datasets:original", "language:en", "language:ru", "license:cc-by-sa-3.0", "license:cc-by-nc-sa-4.0", "license:gfdl", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["crowdsourced"], "language": ["en", "ru"], "license": ["cc-by-sa-3.0", "cc-by-nc-sa-4.0", "gfdl"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["extended|other-db_pedia", "original"], "task_categories": ["tabular-to-text"], "task_ids": ["rdf-to-text"], "paperswithcode_id": "webnlg", "pretty_name": "WebNLG", "config_names": ["release_v1", "release_v2", "release_v2.1", "release_v2.1_constrained", "release_v2_constrained", "release_v3.0_en", "release_v3.0_ru", "webnlg_challenge_2017"], "dataset_info": [{"config_name": "webnlg_challenge_2017", "features": [{"name": "category", "dtype": "string"}, {"name": "size", "dtype": "int32"}, {"name": "eid", "dtype": "string"}, {"name": "original_triple_sets", "sequence": [{"name": "otriple_set", "sequence": "string"}]}, {"name": "modified_triple_sets", "sequence": [{"name": "mtriple_set", "sequence": "string"}]}, {"name": "shape", "dtype": "string"}, {"name": "shape_type", "dtype": "string"}, {"name": "lex", "sequence": [{"name": "comment", "dtype": "string"}, {"name": "lid", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "lang", "dtype": "string"}]}, {"name": "test_category", "dtype": "string"}, {"name": "dbpedia_links", "sequence": "string"}, {"name": "links", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 5594812, "num_examples": 6940}, {"name": "dev", "num_bytes": 706653, "num_examples": 872}, {"name": "test", "num_bytes": 3122533, "num_examples": 4615}], "download_size": 25499351, "dataset_size": 9423998}, {"config_name": "release_v1", "features": [{"name": "category", "dtype": "string"}, {"name": "size", "dtype": "int32"}, {"name": "eid", "dtype": "string"}, {"name": "original_triple_sets", "sequence": [{"name": "otriple_set", "sequence": "string"}]}, {"name": "modified_triple_sets", "sequence": [{"name": "mtriple_set", "sequence": "string"}]}, {"name": "shape", "dtype": "string"}, {"name": "shape_type", "dtype": "string"}, {"name": "lex", "sequence": [{"name": "comment", "dtype": "string"}, {"name": "lid", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "lang", "dtype": "string"}]}, {"name": "test_category", "dtype": "string"}, {"name": "dbpedia_links", "sequence": "string"}, {"name": "links", "sequence": "string"}], "splits": [{"name": "full", "num_bytes": 11684308, "num_examples": 14237}], "download_size": 25499351, "dataset_size": 11684308}, {"config_name": "release_v2", "features": [{"name": "category", "dtype": "string"}, {"name": "size", "dtype": "int32"}, {"name": "eid", "dtype": "string"}, {"name": "original_triple_sets", "sequence": [{"name": "otriple_set", "sequence": "string"}]}, {"name": "modified_triple_sets", "sequence": [{"name": "mtriple_set", "sequence": "string"}]}, {"name": "shape", "dtype": "string"}, {"name": "shape_type", "dtype": "string"}, {"name": "lex", "sequence": [{"name": "comment", "dtype": "string"}, {"name": "lid", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "lang", "dtype": "string"}]}, {"name": "test_category", "dtype": "string"}, {"name": "dbpedia_links", "sequence": "string"}, {"name": "links", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 10830413, "num_examples": 12876}, {"name": "dev", "num_bytes": 1360033, "num_examples": 1619}, {"name": "test", "num_bytes": 1324934, "num_examples": 1600}], "download_size": 25499351, "dataset_size": 13515380}, {"config_name": "release_v2_constrained", "features": [{"name": "category", "dtype": "string"}, {"name": "size", "dtype": "int32"}, {"name": "eid", "dtype": "string"}, {"name": "original_triple_sets", "sequence": [{"name": "otriple_set", "sequence": "string"}]}, {"name": "modified_triple_sets", "sequence": [{"name": "mtriple_set", "sequence": "string"}]}, {"name": "shape", "dtype": "string"}, {"name": "shape_type", "dtype": "string"}, {"name": "lex", "sequence": [{"name": "comment", "dtype": "string"}, {"name": "lid", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "lang", "dtype": "string"}]}, {"name": "test_category", "dtype": "string"}, {"name": "dbpedia_links", "sequence": "string"}, {"name": "links", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 10853434, "num_examples": 12895}, {"name": "dev", "num_bytes": 1421590, "num_examples": 1594}, {"name": "test", "num_bytes": 1243182, "num_examples": 1606}], "download_size": 25499351, "dataset_size": 13518206}, {"config_name": "release_v2.1", "features": [{"name": "category", "dtype": "string"}, {"name": "size", "dtype": "int32"}, {"name": "eid", "dtype": "string"}, {"name": "original_triple_sets", "sequence": [{"name": "otriple_set", "sequence": "string"}]}, {"name": "modified_triple_sets", "sequence": [{"name": "mtriple_set", "sequence": "string"}]}, {"name": "shape", "dtype": "string"}, {"name": "shape_type", "dtype": "string"}, {"name": "lex", "sequence": [{"name": "comment", "dtype": "string"}, {"name": "lid", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "lang", "dtype": "string"}]}, {"name": "test_category", "dtype": "string"}, {"name": "dbpedia_links", "sequence": "string"}, {"name": "links", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 10848793, "num_examples": 12876}, {"name": "dev", "num_bytes": 1362072, "num_examples": 1619}, {"name": "test", "num_bytes": 1325860, "num_examples": 1600}], "download_size": 25499351, "dataset_size": 13536725}, {"config_name": "release_v2.1_constrained", "features": [{"name": "category", "dtype": "string"}, {"name": "size", "dtype": "int32"}, {"name": "eid", "dtype": "string"}, {"name": "original_triple_sets", "sequence": [{"name": "otriple_set", "sequence": "string"}]}, {"name": "modified_triple_sets", "sequence": [{"name": "mtriple_set", "sequence": "string"}]}, {"name": "shape", "dtype": "string"}, {"name": "shape_type", "dtype": "string"}, {"name": "lex", "sequence": [{"name": "comment", "dtype": "string"}, {"name": "lid", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "lang", "dtype": "string"}]}, {"name": "test_category", "dtype": "string"}, {"name": "dbpedia_links", "sequence": "string"}, {"name": "links", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 11040016, "num_examples": 12895}, {"name": "dev", "num_bytes": 1284044, "num_examples": 1594}, {"name": "test", "num_bytes": 1212665, "num_examples": 1606}], "download_size": 25499351, "dataset_size": 13536725}, {"config_name": "release_v3.0_en", "features": [{"name": "category", "dtype": "string"}, {"name": "size", "dtype": "int32"}, {"name": "eid", "dtype": "string"}, {"name": "original_triple_sets", "sequence": [{"name": "otriple_set", "sequence": "string"}]}, {"name": "modified_triple_sets", "sequence": [{"name": "mtriple_set", "sequence": "string"}]}, {"name": "shape", "dtype": "string"}, {"name": "shape_type", "dtype": "string"}, {"name": "lex", "sequence": [{"name": "comment", "dtype": "string"}, {"name": "lid", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "lang", "dtype": "string"}]}, {"name": "test_category", "dtype": "string"}, {"name": "dbpedia_links", "sequence": "string"}, {"name": "links", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 11084860, "num_examples": 13211}, {"name": "dev", "num_bytes": 1394243, "num_examples": 1667}, {"name": "test", "num_bytes": 4039282, "num_examples": 5713}], "download_size": 25499351, "dataset_size": 16518385}, {"config_name": "release_v3.0_ru", "features": [{"name": "category", "dtype": "string"}, {"name": "size", "dtype": "int32"}, {"name": "eid", "dtype": "string"}, {"name": "original_triple_sets", "sequence": [{"name": "otriple_set", "sequence": "string"}]}, {"name": "modified_triple_sets", "sequence": [{"name": "mtriple_set", "sequence": "string"}]}, {"name": "shape", "dtype": "string"}, {"name": "shape_type", "dtype": "string"}, {"name": "lex", "sequence": [{"name": "comment", "dtype": "string"}, {"name": "lid", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "lang", "dtype": "string"}]}, {"name": "test_category", "dtype": "string"}, {"name": "dbpedia_links", "sequence": "string"}, {"name": "links", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 9550340, "num_examples": 5573}, {"name": "dev", "num_bytes": 1314226, "num_examples": 790}, {"name": "test", "num_bytes": 3656501, "num_examples": 3410}], "download_size": 25499351, "dataset_size": 14521067}]}
2024-01-18T11:17:52+00:00
[]
[ "en", "ru" ]
TAGS #task_categories-tabular-to-text #task_ids-rdf-to-text #annotations_creators-found #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-extended|other-db_pedia #source_datasets-original #language-English #language-Russian #license-cc-by-sa-3.0 #license-cc-by-nc-sa-4.0 #license-gfdl #region-us
Dataset Card for WebNLG ======================= Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: WebNLG challenge website * Repository: WebNLG GitLab repository * Paper: Creating Training Corpora for NLG Micro-Planning * Leaderboard: WebNLG leaderboards * Point of Contact: anastasia.shimorina@URL ### Dataset Summary The WebNLG challenge consists in mapping data to text. The training data consists of Data/Text pairs where the data is a set of triples extracted from DBpedia and the text is a verbalisation of these triples. For instance, given the 3 DBpedia triples shown in (a), the aim is to generate a text such as (b). As the example illustrates, the task involves specific NLG subtasks such as sentence segmentation (how to chunk the input data into sentences), lexicalisation (of the DBpedia properties), aggregation (how to avoid repetitions) and surface realisation (how to build a syntactically correct and natural sounding text). ### Supported Tasks and Leaderboards The dataset supports a Structured to Text task which requires a model takes a set of RDF (Resource Description Format) triples from a database (DBpedia) of the form (subject, property, object) as input and write out a natural language sentence expressing the information contained in the triples. The dataset has supportd two challenges: the WebNLG2017 and WebNLG2020 challenge. Results were ordered by their METEOR to the reference, but the leaderboards report a range of other metrics including BLEU, BERTscore, and BLEURT. The v3 release ('release\_v3.0\_en', 'release\_v3.0\_ru') for the WebNLG2020 challenge also supports a semantic 'parsing' task. ### Languages All releases contain English ('en') data. The v3 release ('release\_v3.0\_ru') also contains Russian ('ru') examples. Dataset Structure ----------------- ### Data Instances A typical example contains the original RDF triples in the set, a modified version which presented to crowd workers, and a set of possible verbalizations for this set of triples: ### Data Fields The following fields can be found in the instances: * 'category': the category of the DBpedia entities present in the RDF triples. * 'eid': an example ID, only unique per split per category. * 'size': number of RDF triples in the set. * 'shape': (since v2) Each set of RDF-triples is a tree, which is characterised by its shape and shape type. 'shape' is a string representation of the tree with nested parentheses where X is a node (see Newick tree format) * 'shape\_type': (since v2) is a type of the tree shape, which can be: 'chain' (the object of one triple is the subject of the other); 'sibling' (triples with a shared subject); 'mixed' (both chain and sibling types present). * 'test\_category': (for 'webnlg\_challenge\_2017' and 'v3') tells whether the set of RDF triples was present in the training set or not. Several splits of the test set are available: with and without references, and for RDF-to-text generation / for semantic parsing. * 'lex': the lexicalizations, with: + 'text': the text to be predicted. + 'lid': a lexicalization ID, unique per example. + 'comment': the lexicalizations were rated by crowd workers are either 'good' or 'bad' + 'lang': (for 'release\_v3.0\_ru') the language used because original English texts were kept in the Russian version. Russian data has additional optional fields comparing to English: * 'dbpedialinks': RDF triples extracted from DBpedia between English and Russian entities by means of the property 'sameAs'. * 'links': RDF triples created manually for some entities to serve as pointers to translators. There are two types of them: + with 'sameAs' ('Spaniards | sameAs | испанцы') + with 'includes' ('Tomatoes, guanciale, cheese, olive oil | includes | гуанчиале'). Those were mostly created for string literals to translate some parts of them. ### Data Splits For 'v3.0' releases: Dataset Creation ---------------- ### Curation Rationale The WebNLG dataset was created to promote the development *(i)* of RDF verbalisers and *(ii)* of microplanners able to handle a wide range of linguistic constructions. The dataset aims at covering knowledge in different domains ("categories"). The same properties and entities can appear in several categories. ### Source Data The data was compiled from raw DBpedia triples. This paper explains how the triples were selected. #### Initial Data Collection and Normalization Initial triples extracted from DBpedia were modified in several ways. See official documentation for the most frequent changes that have been made. An original tripleset and a modified tripleset usually represent a one-to-one mapping. However, there are cases with many-to-one mappings when several original triplesets are mapped to one modified tripleset. Entities that served as roots of RDF trees are listed in this file. The English WebNLG 2020 dataset (v3.0) for training comprises data-text pairs for 16 distinct DBpedia categories: * The 10 seen categories used in the 2017 version: Airport, Astronaut, Building, City, ComicsCharacter, Food, Monument, SportsTeam, University, and WrittenWork. * The 5 unseen categories of 2017, which are now part of the seen data: Athlete, Artist, CelestialBody, MeanOfTransportation, Politician. * 1 new category: Company. The Russian dataset (v3.0) comprises data-text pairs for 9 distinct categories: Airport, Astronaut, Building, CelestialBody, ComicsCharacter, Food, Monument, SportsTeam, and University. #### Who are the source language producers? There are no source texts, all textual material was compiled during the annotation process. ### Annotations #### Annotation process Annotators were first asked to create sentences that verbalise single triples. In a second round, annotators were asked to combine single-triple sentences together into sentences that cover 2 triples. And so on until 7 triples. Quality checks were performed to ensure the quality of the annotations. See Section 3.3 in the dataset paper. Russian data was translated from English with an MT system and then was post-edited by crowdworkers. See Section 2.2 of this paper. #### Who are the annotators? All references were collected through crowdsourcing platforms (CrowdFlower/Figure 8 and Amazon Mechanical Turk). For Russian, post-editing was done using the Yandex.Toloka crowdsourcing platform. ### Personal and Sensitive Information Neither the dataset as published or the annotation process involves the collection or sharing of any kind of personal / demographic information. Considerations for Using the Data --------------------------------- ### Social Impact of Dataset We do not foresee any negative social impact in particular from this dataset or task. Positive outlooks: Being able to generate good quality text from RDF data would permit, e.g., making this data more accessible to lay users, enriching existing text with information drawn from knowledge bases such as DBpedia or describing, comparing and relating entities present in these knowledge bases. ### Discussion of Biases This dataset is created using DBpedia RDF triples which naturally exhibit biases that have been found to exist in Wikipedia such as some forms of, e.g., gender bias. The choice of entities, described by RDF trees, was not controlled. As such, they may contain gender biases; for instance, all the astronauts described by RDF triples are male. Hence, in texts, pronouns *he/him/his* occur more often. Similarly, entities can be related to the Western culture more often than to other cultures. ### Other Known Limitations The quality of the crowdsourced references is limited, in particular in terms of fluency/naturalness of the collected texts. Russian data was machine-translated and then post-edited by crowdworkers, so some examples may still exhibit issues related to bad translations. Additional Information ---------------------- ### Dataset Curators The principle curator of the dataset is Anastasia Shimorina (Université de Lorraine / LORIA, France). Throughout the WebNLG releases, several people contributed to their construction: Claire Gardent (CNRS / LORIA, France), Shashi Narayan (Google, UK), Laura Perez-Beltrachini (University of Edinburgh, UK), Elena Khasanova, and Thiago Castro Ferreira (Federal University of Minas Gerais, Brazil). The dataset construction was funded by the French National Research Agency (ANR). ### Licensing Information The dataset uses the 'cc-by-nc-sa-4.0' license. The source DBpedia project uses the 'cc-by-sa-3.0' and 'gfdl-1.1' licenses. * If you use the WebNLG corpus, cite: * If you use 'release\_v2\_constrained' in particular, cite: ### Contributions Thanks to @Shimorina, @yjernite for adding this dataset.
[ "### Dataset Summary\n\n\nThe WebNLG challenge consists in mapping data to text. The training data consists\nof Data/Text pairs where the data is a set of triples extracted from DBpedia and the text is a verbalisation\nof these triples. For instance, given the 3 DBpedia triples shown in (a), the aim is to generate a text such as (b).\n\n\nAs the example illustrates, the task involves specific NLG subtasks such as sentence segmentation\n(how to chunk the input data into sentences), lexicalisation (of the DBpedia properties),\naggregation (how to avoid repetitions) and surface realisation\n(how to build a syntactically correct and natural sounding text).", "### Supported Tasks and Leaderboards\n\n\nThe dataset supports a Structured to Text task which requires a model takes a set of RDF (Resource Description Format) triples from a database (DBpedia) of the form (subject, property, object) as input and write out a natural language sentence expressing the information contained in the triples. The dataset has supportd two challenges: the WebNLG2017 and WebNLG2020 challenge. Results were ordered by their METEOR to the reference, but the leaderboards report a range of other metrics including BLEU, BERTscore, and BLEURT. The v3 release ('release\\_v3.0\\_en', 'release\\_v3.0\\_ru') for the WebNLG2020 challenge also supports a semantic 'parsing' task.", "### Languages\n\n\nAll releases contain English ('en') data. The v3 release ('release\\_v3.0\\_ru') also contains Russian ('ru') examples.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nA typical example contains the original RDF triples in the set, a modified version which presented to crowd workers, and a set of possible verbalizations for this set of triples:", "### Data Fields\n\n\nThe following fields can be found in the instances:\n\n\n* 'category': the category of the DBpedia entities present in the RDF triples.\n* 'eid': an example ID, only unique per split per category.\n* 'size': number of RDF triples in the set.\n* 'shape': (since v2) Each set of RDF-triples is a tree, which is characterised by its shape and shape type. 'shape' is a string representation of the tree with nested parentheses where X is a node (see Newick tree format)\n* 'shape\\_type': (since v2) is a type of the tree shape, which can be: 'chain' (the object of one triple is the subject of the other); 'sibling' (triples with a shared subject); 'mixed' (both chain and sibling types present).\n* 'test\\_category': (for 'webnlg\\_challenge\\_2017' and 'v3') tells whether the set of RDF triples was present in the training set or not. Several splits of the test set are available: with and without references, and for RDF-to-text generation / for semantic parsing.\n* 'lex': the lexicalizations, with:\n\t+ 'text': the text to be predicted.\n\t+ 'lid': a lexicalization ID, unique per example.\n\t+ 'comment': the lexicalizations were rated by crowd workers are either 'good' or 'bad'\n\t+ 'lang': (for 'release\\_v3.0\\_ru') the language used because original English texts were kept in the Russian version.\n\n\nRussian data has additional optional fields comparing to English:\n\n\n* 'dbpedialinks': RDF triples extracted from DBpedia between English and Russian entities by means of the property 'sameAs'.\n* 'links': RDF triples created manually for some entities to serve as pointers to translators. There are two types of them:\n\t+ with 'sameAs' ('Spaniards | sameAs | испанцы')\n\t+ with 'includes' ('Tomatoes, guanciale, cheese, olive oil | includes | гуанчиале'). Those were mostly created for string literals to translate some parts of them.", "### Data Splits\n\n\nFor 'v3.0' releases:\n\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nThe WebNLG dataset was created to promote the development *(i)* of RDF verbalisers and *(ii)* of microplanners able to handle a wide range of linguistic constructions. The dataset aims at covering knowledge in different domains (\"categories\"). The same properties and entities can appear in several categories.", "### Source Data\n\n\nThe data was compiled from raw DBpedia triples. This paper explains how the triples were selected.", "#### Initial Data Collection and Normalization\n\n\nInitial triples extracted from DBpedia were modified in several ways. See official documentation for the most frequent changes that have been made. An original tripleset and a modified tripleset usually represent a one-to-one mapping. However, there are cases with many-to-one mappings when several original triplesets are mapped to one modified tripleset.\n\n\nEntities that served as roots of RDF trees are listed in this file.\n\n\nThe English WebNLG 2020 dataset (v3.0) for training comprises data-text pairs for 16 distinct DBpedia categories:\n\n\n* The 10 seen categories used in the 2017 version: Airport, Astronaut, Building, City, ComicsCharacter, Food, Monument, SportsTeam, University, and WrittenWork.\n* The 5 unseen categories of 2017, which are now part of the seen data: Athlete, Artist, CelestialBody, MeanOfTransportation, Politician.\n* 1 new category: Company.\n\n\nThe Russian dataset (v3.0) comprises data-text pairs for 9 distinct categories: Airport, Astronaut, Building, CelestialBody, ComicsCharacter, Food, Monument, SportsTeam, and University.", "#### Who are the source language producers?\n\n\nThere are no source texts, all textual material was compiled during the annotation process.", "### Annotations", "#### Annotation process\n\n\nAnnotators were first asked to create sentences that verbalise single triples. In a second round, annotators were asked to combine single-triple sentences together into sentences that cover 2 triples. And so on until 7 triples. Quality checks were performed to ensure the quality of the annotations. See Section 3.3 in the dataset paper.\n\n\nRussian data was translated from English with an MT system and then was post-edited by crowdworkers. See Section 2.2 of this paper.", "#### Who are the annotators?\n\n\nAll references were collected through crowdsourcing platforms (CrowdFlower/Figure 8 and Amazon Mechanical Turk). For Russian, post-editing was done using the Yandex.Toloka crowdsourcing platform.", "### Personal and Sensitive Information\n\n\nNeither the dataset as published or the annotation process involves the collection or sharing of any kind of personal / demographic information.\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset\n\n\nWe do not foresee any negative social impact in particular from this dataset or task.\n\n\nPositive outlooks: Being able to generate good quality text from RDF data would permit, e.g., making this data more accessible to lay users, enriching existing text with information drawn from knowledge bases such as DBpedia or describing, comparing and relating entities present in these knowledge bases.", "### Discussion of Biases\n\n\nThis dataset is created using DBpedia RDF triples which naturally exhibit biases that have been found to exist in Wikipedia such as some forms of, e.g., gender bias.\n\n\nThe choice of entities, described by RDF trees, was not controlled. As such, they may contain gender biases; for instance, all the astronauts described by RDF triples are male. Hence, in texts, pronouns *he/him/his* occur more often. Similarly, entities can be related to the Western culture more often than to other cultures.", "### Other Known Limitations\n\n\nThe quality of the crowdsourced references is limited, in particular in terms of fluency/naturalness of the collected texts.\n\n\nRussian data was machine-translated and then post-edited by crowdworkers, so some examples may still exhibit issues related to bad translations.\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThe principle curator of the dataset is Anastasia Shimorina (Université de Lorraine / LORIA, France). Throughout the WebNLG releases, several people contributed to their construction: Claire Gardent (CNRS / LORIA, France), Shashi Narayan (Google, UK), Laura Perez-Beltrachini (University of Edinburgh, UK), Elena Khasanova, and Thiago Castro Ferreira (Federal University of Minas Gerais, Brazil).\nThe dataset construction was funded by the French National Research Agency (ANR).", "### Licensing Information\n\n\nThe dataset uses the 'cc-by-nc-sa-4.0' license. The source DBpedia project uses the 'cc-by-sa-3.0' and 'gfdl-1.1' licenses.\n\n\n* If you use the WebNLG corpus, cite:\n* If you use 'release\\_v2\\_constrained' in particular, cite:", "### Contributions\n\n\nThanks to @Shimorina, @yjernite for adding this dataset." ]
[ "TAGS\n#task_categories-tabular-to-text #task_ids-rdf-to-text #annotations_creators-found #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-extended|other-db_pedia #source_datasets-original #language-English #language-Russian #license-cc-by-sa-3.0 #license-cc-by-nc-sa-4.0 #license-gfdl #region-us \n", "### Dataset Summary\n\n\nThe WebNLG challenge consists in mapping data to text. The training data consists\nof Data/Text pairs where the data is a set of triples extracted from DBpedia and the text is a verbalisation\nof these triples. For instance, given the 3 DBpedia triples shown in (a), the aim is to generate a text such as (b).\n\n\nAs the example illustrates, the task involves specific NLG subtasks such as sentence segmentation\n(how to chunk the input data into sentences), lexicalisation (of the DBpedia properties),\naggregation (how to avoid repetitions) and surface realisation\n(how to build a syntactically correct and natural sounding text).", "### Supported Tasks and Leaderboards\n\n\nThe dataset supports a Structured to Text task which requires a model takes a set of RDF (Resource Description Format) triples from a database (DBpedia) of the form (subject, property, object) as input and write out a natural language sentence expressing the information contained in the triples. The dataset has supportd two challenges: the WebNLG2017 and WebNLG2020 challenge. Results were ordered by their METEOR to the reference, but the leaderboards report a range of other metrics including BLEU, BERTscore, and BLEURT. The v3 release ('release\\_v3.0\\_en', 'release\\_v3.0\\_ru') for the WebNLG2020 challenge also supports a semantic 'parsing' task.", "### Languages\n\n\nAll releases contain English ('en') data. The v3 release ('release\\_v3.0\\_ru') also contains Russian ('ru') examples.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nA typical example contains the original RDF triples in the set, a modified version which presented to crowd workers, and a set of possible verbalizations for this set of triples:", "### Data Fields\n\n\nThe following fields can be found in the instances:\n\n\n* 'category': the category of the DBpedia entities present in the RDF triples.\n* 'eid': an example ID, only unique per split per category.\n* 'size': number of RDF triples in the set.\n* 'shape': (since v2) Each set of RDF-triples is a tree, which is characterised by its shape and shape type. 'shape' is a string representation of the tree with nested parentheses where X is a node (see Newick tree format)\n* 'shape\\_type': (since v2) is a type of the tree shape, which can be: 'chain' (the object of one triple is the subject of the other); 'sibling' (triples with a shared subject); 'mixed' (both chain and sibling types present).\n* 'test\\_category': (for 'webnlg\\_challenge\\_2017' and 'v3') tells whether the set of RDF triples was present in the training set or not. Several splits of the test set are available: with and without references, and for RDF-to-text generation / for semantic parsing.\n* 'lex': the lexicalizations, with:\n\t+ 'text': the text to be predicted.\n\t+ 'lid': a lexicalization ID, unique per example.\n\t+ 'comment': the lexicalizations were rated by crowd workers are either 'good' or 'bad'\n\t+ 'lang': (for 'release\\_v3.0\\_ru') the language used because original English texts were kept in the Russian version.\n\n\nRussian data has additional optional fields comparing to English:\n\n\n* 'dbpedialinks': RDF triples extracted from DBpedia between English and Russian entities by means of the property 'sameAs'.\n* 'links': RDF triples created manually for some entities to serve as pointers to translators. There are two types of them:\n\t+ with 'sameAs' ('Spaniards | sameAs | испанцы')\n\t+ with 'includes' ('Tomatoes, guanciale, cheese, olive oil | includes | гуанчиале'). Those were mostly created for string literals to translate some parts of them.", "### Data Splits\n\n\nFor 'v3.0' releases:\n\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nThe WebNLG dataset was created to promote the development *(i)* of RDF verbalisers and *(ii)* of microplanners able to handle a wide range of linguistic constructions. The dataset aims at covering knowledge in different domains (\"categories\"). The same properties and entities can appear in several categories.", "### Source Data\n\n\nThe data was compiled from raw DBpedia triples. This paper explains how the triples were selected.", "#### Initial Data Collection and Normalization\n\n\nInitial triples extracted from DBpedia were modified in several ways. See official documentation for the most frequent changes that have been made. An original tripleset and a modified tripleset usually represent a one-to-one mapping. However, there are cases with many-to-one mappings when several original triplesets are mapped to one modified tripleset.\n\n\nEntities that served as roots of RDF trees are listed in this file.\n\n\nThe English WebNLG 2020 dataset (v3.0) for training comprises data-text pairs for 16 distinct DBpedia categories:\n\n\n* The 10 seen categories used in the 2017 version: Airport, Astronaut, Building, City, ComicsCharacter, Food, Monument, SportsTeam, University, and WrittenWork.\n* The 5 unseen categories of 2017, which are now part of the seen data: Athlete, Artist, CelestialBody, MeanOfTransportation, Politician.\n* 1 new category: Company.\n\n\nThe Russian dataset (v3.0) comprises data-text pairs for 9 distinct categories: Airport, Astronaut, Building, CelestialBody, ComicsCharacter, Food, Monument, SportsTeam, and University.", "#### Who are the source language producers?\n\n\nThere are no source texts, all textual material was compiled during the annotation process.", "### Annotations", "#### Annotation process\n\n\nAnnotators were first asked to create sentences that verbalise single triples. In a second round, annotators were asked to combine single-triple sentences together into sentences that cover 2 triples. And so on until 7 triples. Quality checks were performed to ensure the quality of the annotations. See Section 3.3 in the dataset paper.\n\n\nRussian data was translated from English with an MT system and then was post-edited by crowdworkers. See Section 2.2 of this paper.", "#### Who are the annotators?\n\n\nAll references were collected through crowdsourcing platforms (CrowdFlower/Figure 8 and Amazon Mechanical Turk). For Russian, post-editing was done using the Yandex.Toloka crowdsourcing platform.", "### Personal and Sensitive Information\n\n\nNeither the dataset as published or the annotation process involves the collection or sharing of any kind of personal / demographic information.\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset\n\n\nWe do not foresee any negative social impact in particular from this dataset or task.\n\n\nPositive outlooks: Being able to generate good quality text from RDF data would permit, e.g., making this data more accessible to lay users, enriching existing text with information drawn from knowledge bases such as DBpedia or describing, comparing and relating entities present in these knowledge bases.", "### Discussion of Biases\n\n\nThis dataset is created using DBpedia RDF triples which naturally exhibit biases that have been found to exist in Wikipedia such as some forms of, e.g., gender bias.\n\n\nThe choice of entities, described by RDF trees, was not controlled. As such, they may contain gender biases; for instance, all the astronauts described by RDF triples are male. Hence, in texts, pronouns *he/him/his* occur more often. Similarly, entities can be related to the Western culture more often than to other cultures.", "### Other Known Limitations\n\n\nThe quality of the crowdsourced references is limited, in particular in terms of fluency/naturalness of the collected texts.\n\n\nRussian data was machine-translated and then post-edited by crowdworkers, so some examples may still exhibit issues related to bad translations.\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThe principle curator of the dataset is Anastasia Shimorina (Université de Lorraine / LORIA, France). Throughout the WebNLG releases, several people contributed to their construction: Claire Gardent (CNRS / LORIA, France), Shashi Narayan (Google, UK), Laura Perez-Beltrachini (University of Edinburgh, UK), Elena Khasanova, and Thiago Castro Ferreira (Federal University of Minas Gerais, Brazil).\nThe dataset construction was funded by the French National Research Agency (ANR).", "### Licensing Information\n\n\nThe dataset uses the 'cc-by-nc-sa-4.0' license. The source DBpedia project uses the 'cc-by-sa-3.0' and 'gfdl-1.1' licenses.\n\n\n* If you use the WebNLG corpus, cite:\n* If you use 'release\\_v2\\_constrained' in particular, cite:", "### Contributions\n\n\nThanks to @Shimorina, @yjernite for adding this dataset." ]
[ 136, 158, 184, 51, 47, 535, 19, 81, 28, 277, 31, 5, 114, 58, 46, 95, 138, 78, 123, 86, 22 ]
[ "passage: TAGS\n#task_categories-tabular-to-text #task_ids-rdf-to-text #annotations_creators-found #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-extended|other-db_pedia #source_datasets-original #language-English #language-Russian #license-cc-by-sa-3.0 #license-cc-by-nc-sa-4.0 #license-gfdl #region-us \n### Dataset Summary\n\n\nThe WebNLG challenge consists in mapping data to text. The training data consists\nof Data/Text pairs where the data is a set of triples extracted from DBpedia and the text is a verbalisation\nof these triples. For instance, given the 3 DBpedia triples shown in (a), the aim is to generate a text such as (b).\n\n\nAs the example illustrates, the task involves specific NLG subtasks such as sentence segmentation\n(how to chunk the input data into sentences), lexicalisation (of the DBpedia properties),\naggregation (how to avoid repetitions) and surface realisation\n(how to build a syntactically correct and natural sounding text).### Supported Tasks and Leaderboards\n\n\nThe dataset supports a Structured to Text task which requires a model takes a set of RDF (Resource Description Format) triples from a database (DBpedia) of the form (subject, property, object) as input and write out a natural language sentence expressing the information contained in the triples. The dataset has supportd two challenges: the WebNLG2017 and WebNLG2020 challenge. Results were ordered by their METEOR to the reference, but the leaderboards report a range of other metrics including BLEU, BERTscore, and BLEURT. The v3 release ('release\\_v3.0\\_en', 'release\\_v3.0\\_ru') for the WebNLG2020 challenge also supports a semantic 'parsing' task.", "passage: ### Languages\n\n\nAll releases contain English ('en') data. The v3 release ('release\\_v3.0\\_ru') also contains Russian ('ru') examples.\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nA typical example contains the original RDF triples in the set, a modified version which presented to crowd workers, and a set of possible verbalizations for this set of triples:", "passage: ### Data Fields\n\n\nThe following fields can be found in the instances:\n\n\n* 'category': the category of the DBpedia entities present in the RDF triples.\n* 'eid': an example ID, only unique per split per category.\n* 'size': number of RDF triples in the set.\n* 'shape': (since v2) Each set of RDF-triples is a tree, which is characterised by its shape and shape type. 'shape' is a string representation of the tree with nested parentheses where X is a node (see Newick tree format)\n* 'shape\\_type': (since v2) is a type of the tree shape, which can be: 'chain' (the object of one triple is the subject of the other); 'sibling' (triples with a shared subject); 'mixed' (both chain and sibling types present).\n* 'test\\_category': (for 'webnlg\\_challenge\\_2017' and 'v3') tells whether the set of RDF triples was present in the training set or not. Several splits of the test set are available: with and without references, and for RDF-to-text generation / for semantic parsing.\n* 'lex': the lexicalizations, with:\n\t+ 'text': the text to be predicted.\n\t+ 'lid': a lexicalization ID, unique per example.\n\t+ 'comment': the lexicalizations were rated by crowd workers are either 'good' or 'bad'\n\t+ 'lang': (for 'release\\_v3.0\\_ru') the language used because original English texts were kept in the Russian version.\n\n\nRussian data has additional optional fields comparing to English:\n\n\n* 'dbpedialinks': RDF triples extracted from DBpedia between English and Russian entities by means of the property 'sameAs'.\n* 'links': RDF triples created manually for some entities to serve as pointers to translators. There are two types of them:\n\t+ with 'sameAs' ('Spaniards | sameAs | испанцы')\n\t+ with 'includes' ('Tomatoes, guanciale, cheese, olive oil | includes | гуанчиале'). Those were mostly created for string literals to translate some parts of them.### Data Splits\n\n\nFor 'v3.0' releases:\n\n\n\n\nDataset Creation\n----------------### Curation Rationale\n\n\nThe WebNLG dataset was created to promote the development *(i)* of RDF verbalisers and *(ii)* of microplanners able to handle a wide range of linguistic constructions. The dataset aims at covering knowledge in different domains (\"categories\"). The same properties and entities can appear in several categories.### Source Data\n\n\nThe data was compiled from raw DBpedia triples. This paper explains how the triples were selected.#### Initial Data Collection and Normalization\n\n\nInitial triples extracted from DBpedia were modified in several ways. See official documentation for the most frequent changes that have been made. An original tripleset and a modified tripleset usually represent a one-to-one mapping. However, there are cases with many-to-one mappings when several original triplesets are mapped to one modified tripleset.\n\n\nEntities that served as roots of RDF trees are listed in this file.\n\n\nThe English WebNLG 2020 dataset (v3.0) for training comprises data-text pairs for 16 distinct DBpedia categories:\n\n\n* The 10 seen categories used in the 2017 version: Airport, Astronaut, Building, City, ComicsCharacter, Food, Monument, SportsTeam, University, and WrittenWork.\n* The 5 unseen categories of 2017, which are now part of the seen data: Athlete, Artist, CelestialBody, MeanOfTransportation, Politician.\n* 1 new category: Company.\n\n\nThe Russian dataset (v3.0) comprises data-text pairs for 9 distinct categories: Airport, Astronaut, Building, CelestialBody, ComicsCharacter, Food, Monument, SportsTeam, and University.#### Who are the source language producers?\n\n\nThere are no source texts, all textual material was compiled during the annotation process.### Annotations", "passage: #### Annotation process\n\n\nAnnotators were first asked to create sentences that verbalise single triples. In a second round, annotators were asked to combine single-triple sentences together into sentences that cover 2 triples. And so on until 7 triples. Quality checks were performed to ensure the quality of the annotations. See Section 3.3 in the dataset paper.\n\n\nRussian data was translated from English with an MT system and then was post-edited by crowdworkers. See Section 2.2 of this paper.#### Who are the annotators?\n\n\nAll references were collected through crowdsourcing platforms (CrowdFlower/Figure 8 and Amazon Mechanical Turk). For Russian, post-editing was done using the Yandex.Toloka crowdsourcing platform.### Personal and Sensitive Information\n\n\nNeither the dataset as published or the annotation process involves the collection or sharing of any kind of personal / demographic information.\n\n\nConsiderations for Using the Data\n---------------------------------### Social Impact of Dataset\n\n\nWe do not foresee any negative social impact in particular from this dataset or task.\n\n\nPositive outlooks: Being able to generate good quality text from RDF data would permit, e.g., making this data more accessible to lay users, enriching existing text with information drawn from knowledge bases such as DBpedia or describing, comparing and relating entities present in these knowledge bases.### Discussion of Biases\n\n\nThis dataset is created using DBpedia RDF triples which naturally exhibit biases that have been found to exist in Wikipedia such as some forms of, e.g., gender bias.\n\n\nThe choice of entities, described by RDF trees, was not controlled. As such, they may contain gender biases; for instance, all the astronauts described by RDF triples are male. Hence, in texts, pronouns *he/him/his* occur more often. Similarly, entities can be related to the Western culture more often than to other cultures.### Other Known Limitations\n\n\nThe quality of the crowdsourced references is limited, in particular in terms of fluency/naturalness of the collected texts.\n\n\nRussian data was machine-translated and then post-edited by crowdworkers, so some examples may still exhibit issues related to bad translations.\n\n\nAdditional Information\n----------------------" ]
26ae29c300edec4acafbda030653640241c9f9a8
# Dataset Card for "web_of_science" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://data.mendeley.com/datasets/9rw3vkcfy4/6](https://data.mendeley.com/datasets/9rw3vkcfy4/6) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 180.67 MB - **Size of the generated dataset:** 89.81 MB - **Total amount of disk used:** 270.48 MB ### Dataset Summary Copyright (c) 2017 Kamran Kowsari Permission is hereby granted, free of charge, to any person obtaining a copy of this dataset and associated documentation files (the "Dataset"), to deal in the dataset without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Dataset, and to permit persons to whom the dataset is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Dataset. If you use this dataset please cite: Referenced paper: HDLTex: Hierarchical Deep Learning for Text Classification Description of Dataset: Here is three datasets which include WOS-11967 , WOS-46985, and WOS-5736 Each folder contains: -X.txt -Y.txt -YL1.txt -YL2.txt X is input data that include text sequences Y is target value YL1 is target value of level one (parent label) YL2 is target value of level one (child label) Web of Science Dataset WOS-5736 -This dataset contains 5,736 documents with 11 categories which include 3 parents categories. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### WOS11967 - **Size of downloaded dataset files:** 60.22 MB - **Size of the generated dataset:** 16.25 MB - **Total amount of disk used:** 76.48 MB An example of 'train' looks as follows. ``` ``` #### WOS46985 - **Size of downloaded dataset files:** 60.22 MB - **Size of the generated dataset:** 65.50 MB - **Total amount of disk used:** 125.72 MB An example of 'train' looks as follows. ``` ``` #### WOS5736 - **Size of downloaded dataset files:** 60.22 MB - **Size of the generated dataset:** 8.05 MB - **Total amount of disk used:** 68.27 MB An example of 'train' looks as follows. ``` ``` ### Data Fields The data fields are the same among all splits. #### WOS11967 - `input_data`: a `string` feature. - `label`: a `int32` feature. - `label_level_1`: a `int32` feature. - `label_level_2`: a `int32` feature. #### WOS46985 - `input_data`: a `string` feature. - `label`: a `int32` feature. - `label_level_1`: a `int32` feature. - `label_level_2`: a `int32` feature. #### WOS5736 - `input_data`: a `string` feature. - `label`: a `int32` feature. - `label_level_1`: a `int32` feature. - `label_level_2`: a `int32` feature. ### Data Splits | name |train| |--------|----:| |WOS11967|11967| |WOS46985|46985| |WOS5736 | 5736| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @inproceedings{kowsari2017HDLTex, title={HDLTex: Hierarchical Deep Learning for Text Classification}, author={Kowsari, Kamran and Brown, Donald E and Heidarysafa, Mojtaba and Jafari Meimandi, Kiana and and Gerber, Matthew S and Barnes, Laura E}, booktitle={Machine Learning and Applications (ICMLA), 2017 16th IEEE International Conference on}, year={2017}, organization={IEEE} } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@lhoestq](https://github.com/lhoestq), [@mariamabarham](https://github.com/mariamabarham), [@lewtun](https://github.com/lewtun) for adding this dataset.
web_of_science
[ "language:en", "region:us" ]
2022-03-02T23:29:22+00:00
{"language": ["en"], "paperswithcode_id": "web-of-science-dataset", "pretty_name": "Web of Science Dataset", "dataset_info": [{"config_name": "WOS5736", "features": [{"name": "input_data", "dtype": "string"}, {"name": "label", "dtype": "int32"}, {"name": "label_level_1", "dtype": "int32"}, {"name": "label_level_2", "dtype": "int32"}], "splits": [{"name": "train", "num_bytes": 8051533, "num_examples": 5736}], "download_size": 60222421, "dataset_size": 8051533}, {"config_name": "WOS11967", "features": [{"name": "input_data", "dtype": "string"}, {"name": "label", "dtype": "int32"}, {"name": "label_level_1", "dtype": "int32"}, {"name": "label_level_2", "dtype": "int32"}], "splits": [{"name": "train", "num_bytes": 16248391, "num_examples": 11967}], "download_size": 60222421, "dataset_size": 16248391}, {"config_name": "WOS46985", "features": [{"name": "input_data", "dtype": "string"}, {"name": "label", "dtype": "int32"}, {"name": "label_level_1", "dtype": "int32"}, {"name": "label_level_2", "dtype": "int32"}], "splits": [{"name": "train", "num_bytes": 65471726, "num_examples": 46985}], "download_size": 60222421, "dataset_size": 65471726}]}
2024-01-18T11:17:53+00:00
[]
[ "en" ]
TAGS #language-English #region-us
Dataset Card for "web\_of\_science" =================================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: * Point of Contact: * Size of downloaded dataset files: 180.67 MB * Size of the generated dataset: 89.81 MB * Total amount of disk used: 270.48 MB ### Dataset Summary Copyright (c) 2017 Kamran Kowsari Permission is hereby granted, free of charge, to any person obtaining a copy of this dataset and associated documentation files (the "Dataset"), to deal in the dataset without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Dataset, and to permit persons to whom the dataset is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Dataset. If you use this dataset please cite: Referenced paper: HDLTex: Hierarchical Deep Learning for Text Classification Description of Dataset: Here is three datasets which include WOS-11967 , WOS-46985, and WOS-5736 Each folder contains: -X.txt -Y.txt -URL -URL X is input data that include text sequences Y is target value YL1 is target value of level one (parent label) YL2 is target value of level one (child label) Web of Science Dataset WOS-5736 -This dataset contains 5,736 documents with 11 categories which include 3 parents categories. ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- ### Data Instances #### WOS11967 * Size of downloaded dataset files: 60.22 MB * Size of the generated dataset: 16.25 MB * Total amount of disk used: 76.48 MB An example of 'train' looks as follows. #### WOS46985 * Size of downloaded dataset files: 60.22 MB * Size of the generated dataset: 65.50 MB * Total amount of disk used: 125.72 MB An example of 'train' looks as follows. #### WOS5736 * Size of downloaded dataset files: 60.22 MB * Size of the generated dataset: 8.05 MB * Total amount of disk used: 68.27 MB An example of 'train' looks as follows. ### Data Fields The data fields are the same among all splits. #### WOS11967 * 'input\_data': a 'string' feature. * 'label': a 'int32' feature. * 'label\_level\_1': a 'int32' feature. * 'label\_level\_2': a 'int32' feature. #### WOS46985 * 'input\_data': a 'string' feature. * 'label': a 'int32' feature. * 'label\_level\_1': a 'int32' feature. * 'label\_level\_2': a 'int32' feature. #### WOS5736 * 'input\_data': a 'string' feature. * 'label': a 'int32' feature. * 'label\_level\_1': a 'int32' feature. * 'label\_level\_2': a 'int32' feature. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @thomwolf, @lhoestq, @mariamabarham, @lewtun for adding this dataset.
[ "### Dataset Summary\n\n\nCopyright (c) 2017 Kamran Kowsari\n\n\nPermission is hereby granted, free of charge, to any person obtaining a copy of this dataset and associated documentation files (the \"Dataset\"), to deal\nin the dataset without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\ncopies of the Dataset, and to permit persons to whom the dataset is furnished to do so, subject to the following conditions:\n\n\nThe above copyright notice and this permission notice shall be included in all copies or substantial portions of the Dataset.\n\n\nIf you use this dataset please cite: Referenced paper: HDLTex: Hierarchical Deep Learning for Text Classification\n\n\nDescription of Dataset:\n\n\nHere is three datasets which include WOS-11967 , WOS-46985, and WOS-5736\nEach folder contains:\n-X.txt\n-Y.txt\n-URL\n-URL\n\n\nX is input data that include text sequences\nY is target value\nYL1 is target value of level one (parent label)\nYL2 is target value of level one (child label)\nWeb of Science Dataset WOS-5736\n-This dataset contains 5,736 documents with 11 categories which include 3 parents categories.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### WOS11967\n\n\n* Size of downloaded dataset files: 60.22 MB\n* Size of the generated dataset: 16.25 MB\n* Total amount of disk used: 76.48 MB\n\n\nAn example of 'train' looks as follows.", "#### WOS46985\n\n\n* Size of downloaded dataset files: 60.22 MB\n* Size of the generated dataset: 65.50 MB\n* Total amount of disk used: 125.72 MB\n\n\nAn example of 'train' looks as follows.", "#### WOS5736\n\n\n* Size of downloaded dataset files: 60.22 MB\n* Size of the generated dataset: 8.05 MB\n* Total amount of disk used: 68.27 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### WOS11967\n\n\n* 'input\\_data': a 'string' feature.\n* 'label': a 'int32' feature.\n* 'label\\_level\\_1': a 'int32' feature.\n* 'label\\_level\\_2': a 'int32' feature.", "#### WOS46985\n\n\n* 'input\\_data': a 'string' feature.\n* 'label': a 'int32' feature.\n* 'label\\_level\\_1': a 'int32' feature.\n* 'label\\_level\\_2': a 'int32' feature.", "#### WOS5736\n\n\n* 'input\\_data': a 'string' feature.\n* 'label': a 'int32' feature.\n* 'label\\_level\\_1': a 'int32' feature.\n* 'label\\_level\\_2': a 'int32' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @thomwolf, @lhoestq, @mariamabarham, @lewtun for adding this dataset." ]
[ "TAGS\n#language-English #region-us \n", "### Dataset Summary\n\n\nCopyright (c) 2017 Kamran Kowsari\n\n\nPermission is hereby granted, free of charge, to any person obtaining a copy of this dataset and associated documentation files (the \"Dataset\"), to deal\nin the dataset without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\ncopies of the Dataset, and to permit persons to whom the dataset is furnished to do so, subject to the following conditions:\n\n\nThe above copyright notice and this permission notice shall be included in all copies or substantial portions of the Dataset.\n\n\nIf you use this dataset please cite: Referenced paper: HDLTex: Hierarchical Deep Learning for Text Classification\n\n\nDescription of Dataset:\n\n\nHere is three datasets which include WOS-11967 , WOS-46985, and WOS-5736\nEach folder contains:\n-X.txt\n-Y.txt\n-URL\n-URL\n\n\nX is input data that include text sequences\nY is target value\nYL1 is target value of level one (parent label)\nYL2 is target value of level one (child label)\nWeb of Science Dataset WOS-5736\n-This dataset contains 5,736 documents with 11 categories which include 3 parents categories.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### WOS11967\n\n\n* Size of downloaded dataset files: 60.22 MB\n* Size of the generated dataset: 16.25 MB\n* Total amount of disk used: 76.48 MB\n\n\nAn example of 'train' looks as follows.", "#### WOS46985\n\n\n* Size of downloaded dataset files: 60.22 MB\n* Size of the generated dataset: 65.50 MB\n* Total amount of disk used: 125.72 MB\n\n\nAn example of 'train' looks as follows.", "#### WOS5736\n\n\n* Size of downloaded dataset files: 60.22 MB\n* Size of the generated dataset: 8.05 MB\n* Total amount of disk used: 68.27 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### WOS11967\n\n\n* 'input\\_data': a 'string' feature.\n* 'label': a 'int32' feature.\n* 'label\\_level\\_1': a 'int32' feature.\n* 'label\\_level\\_2': a 'int32' feature.", "#### WOS46985\n\n\n* 'input\\_data': a 'string' feature.\n* 'label': a 'int32' feature.\n* 'label\\_level\\_1': a 'int32' feature.\n* 'label\\_level\\_2': a 'int32' feature.", "#### WOS5736\n\n\n* 'input\\_data': a 'string' feature.\n* 'label': a 'int32' feature.\n* 'label\\_level\\_1': a 'int32' feature.\n* 'label\\_level\\_2': a 'int32' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @thomwolf, @lhoestq, @mariamabarham, @lewtun for adding this dataset." ]
[ 10, 284, 10, 11, 6, 54, 55, 54, 17, 68, 68, 68, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 6, 32 ]
[ "passage: TAGS\n#language-English #region-us \n### Dataset Summary\n\n\nCopyright (c) 2017 Kamran Kowsari\n\n\nPermission is hereby granted, free of charge, to any person obtaining a copy of this dataset and associated documentation files (the \"Dataset\"), to deal\nin the dataset without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\ncopies of the Dataset, and to permit persons to whom the dataset is furnished to do so, subject to the following conditions:\n\n\nThe above copyright notice and this permission notice shall be included in all copies or substantial portions of the Dataset.\n\n\nIf you use this dataset please cite: Referenced paper: HDLTex: Hierarchical Deep Learning for Text Classification\n\n\nDescription of Dataset:\n\n\nHere is three datasets which include WOS-11967 , WOS-46985, and WOS-5736\nEach folder contains:\n-X.txt\n-Y.txt\n-URL\n-URL\n\n\nX is input data that include text sequences\nY is target value\nYL1 is target value of level one (parent label)\nYL2 is target value of level one (child label)\nWeb of Science Dataset WOS-5736\n-This dataset contains 5,736 documents with 11 categories which include 3 parents categories.### Supported Tasks and Leaderboards### Languages\n\n\nDataset Structure\n-----------------### Data Instances#### WOS11967\n\n\n* Size of downloaded dataset files: 60.22 MB\n* Size of the generated dataset: 16.25 MB\n* Total amount of disk used: 76.48 MB\n\n\nAn example of 'train' looks as follows.#### WOS46985\n\n\n* Size of downloaded dataset files: 60.22 MB\n* Size of the generated dataset: 65.50 MB\n* Total amount of disk used: 125.72 MB\n\n\nAn example of 'train' looks as follows.#### WOS5736\n\n\n* Size of downloaded dataset files: 60.22 MB\n* Size of the generated dataset: 8.05 MB\n* Total amount of disk used: 68.27 MB\n\n\nAn example of 'train' looks as follows.### Data Fields\n\n\nThe data fields are the same among all splits." ]
0e473cbe21d1e91ec18da343644498be6a3f5454
# Dataset Card for "web_questions" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://worksheets.codalab.org/worksheets/0xba659fe363cb46e7a505c5b6a774dc8a](https://worksheets.codalab.org/worksheets/0xba659fe363cb46e7a505c5b6a774dc8a) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [Semantic Parsing on Freebase from Question-Answer Pairs](https://aclanthology.org/D13-1160/) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 1.27 MB - **Size of the generated dataset:** 0.83 MB - **Total amount of disk used:** 2.10 MB ### Dataset Summary This dataset consists of 6,642 question/answer pairs. The questions are supposed to be answerable by Freebase, a large knowledge graph. The questions are mostly centered around a single named entity. The questions are popular ones asked on the web (at least in 2013). ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### default - **Size of downloaded dataset files:** 1.27 MB - **Size of the generated dataset:** 0.83 MB - **Total amount of disk used:** 2.10 MB An example of 'train' looks as follows. ``` { "answers": ["Jamaican Creole English Language", "Jamaican English"], "question": "what does jamaican people speak?", "url": "http://www.freebase.com/view/en/jamaica" } ``` ### Data Fields The data fields are the same among all splits. #### default - `url`: a `string` feature. - `question`: a `string` feature. - `answers`: a `list` of `string` features. ### Data Splits | name |train|test| |-------|----:|---:| |default| 3778|2032| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @inproceedings{berant-etal-2013-semantic, title = "Semantic Parsing on {F}reebase from Question-Answer Pairs", author = "Berant, Jonathan and Chou, Andrew and Frostig, Roy and Liang, Percy", booktitle = "Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing", month = oct, year = "2013", address = "Seattle, Washington, USA", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/D13-1160", pages = "1533--1544", } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@mariamabarham](https://github.com/mariamabarham), [@lewtun](https://github.com/lewtun) for adding this dataset.
web_questions
[ "task_categories:question-answering", "task_ids:open-domain-qa", "annotations_creators:crowdsourced", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["found"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["question-answering"], "task_ids": ["open-domain-qa"], "paperswithcode_id": "webquestions", "pretty_name": "WebQuestions", "dataset_info": {"features": [{"name": "url", "dtype": "string"}, {"name": "question", "dtype": "string"}, {"name": "answers", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 530711, "num_examples": 3778}, {"name": "test", "num_bytes": 288184, "num_examples": 2032}], "download_size": 402395, "dataset_size": 818895}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "test", "path": "data/test-*"}]}]}
2024-01-04T16:41:06+00:00
[]
[ "en" ]
TAGS #task_categories-question-answering #task_ids-open-domain-qa #annotations_creators-crowdsourced #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-unknown #region-us
Dataset Card for "web\_questions" ================================= Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: Semantic Parsing on Freebase from Question-Answer Pairs * Point of Contact: * Size of downloaded dataset files: 1.27 MB * Size of the generated dataset: 0.83 MB * Total amount of disk used: 2.10 MB ### Dataset Summary This dataset consists of 6,642 question/answer pairs. The questions are supposed to be answerable by Freebase, a large knowledge graph. The questions are mostly centered around a single named entity. The questions are popular ones asked on the web (at least in 2013). ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- ### Data Instances #### default * Size of downloaded dataset files: 1.27 MB * Size of the generated dataset: 0.83 MB * Total amount of disk used: 2.10 MB An example of 'train' looks as follows. ### Data Fields The data fields are the same among all splits. #### default * 'url': a 'string' feature. * 'question': a 'string' feature. * 'answers': a 'list' of 'string' features. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @thomwolf, @mariamabarham, @lewtun for adding this dataset.
[ "### Dataset Summary\n\n\nThis dataset consists of 6,642 question/answer pairs.\nThe questions are supposed to be answerable by Freebase, a large knowledge graph.\nThe questions are mostly centered around a single named entity.\nThe questions are popular ones asked on the web (at least in 2013).", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### default\n\n\n* Size of downloaded dataset files: 1.27 MB\n* Size of the generated dataset: 0.83 MB\n* Total amount of disk used: 2.10 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### default\n\n\n* 'url': a 'string' feature.\n* 'question': a 'string' feature.\n* 'answers': a 'list' of 'string' features.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @thomwolf, @mariamabarham, @lewtun for adding this dataset." ]
[ "TAGS\n#task_categories-question-answering #task_ids-open-domain-qa #annotations_creators-crowdsourced #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-unknown #region-us \n", "### Dataset Summary\n\n\nThis dataset consists of 6,642 question/answer pairs.\nThe questions are supposed to be answerable by Freebase, a large knowledge graph.\nThe questions are mostly centered around a single named entity.\nThe questions are popular ones asked on the web (at least in 2013).", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### default\n\n\n* Size of downloaded dataset files: 1.27 MB\n* Size of the generated dataset: 0.83 MB\n* Total amount of disk used: 2.10 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### default\n\n\n* 'url': a 'string' feature.\n* 'question': a 'string' feature.\n* 'answers': a 'list' of 'string' features.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @thomwolf, @mariamabarham, @lewtun for adding this dataset." ]
[ 90, 69, 10, 11, 6, 50, 17, 42, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 6, 27 ]
[ "passage: TAGS\n#task_categories-question-answering #task_ids-open-domain-qa #annotations_creators-crowdsourced #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-unknown #region-us \n### Dataset Summary\n\n\nThis dataset consists of 6,642 question/answer pairs.\nThe questions are supposed to be answerable by Freebase, a large knowledge graph.\nThe questions are mostly centered around a single named entity.\nThe questions are popular ones asked on the web (at least in 2013).### Supported Tasks and Leaderboards### Languages\n\n\nDataset Structure\n-----------------### Data Instances#### default\n\n\n* Size of downloaded dataset files: 1.27 MB\n* Size of the generated dataset: 0.83 MB\n* Total amount of disk used: 2.10 MB\n\n\nAn example of 'train' looks as follows.### Data Fields\n\n\nThe data fields are the same among all splits.#### default\n\n\n* 'url': a 'string' feature.\n* 'question': a 'string' feature.\n* 'answers': a 'list' of 'string' features.### Data Splits\n\n\n\nDataset Creation\n----------------### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------### Social Impact of Dataset### Discussion of Biases### Other Known Limitations\n\n\nAdditional Information\n----------------------### Dataset Curators### Licensing Information### Contributions\n\n\nThanks to @thomwolf, @mariamabarham, @lewtun for adding this dataset." ]
18858d111abe9cc866303a7790124f46a65726d4
# Dataset Card for "Weibo NER" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** None - **Repository:** https://github.com/OYE93/Chinese-NLP-Corpus/tree/master/NER/Weibo - **Paper:** [More Information Needed] - **Leaderboard:** [If the dataset supports an active leaderboard, add link here]() - **Point of Contact:** [More Information Needed] ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@abhishekkrthakur](https://github.com/abhishekkrthakur) for adding this dataset.
weibo_ner
[ "task_categories:token-classification", "task_ids:named-entity-recognition", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:zh", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found"], "language": ["zh"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["token-classification"], "task_ids": ["named-entity-recognition"], "paperswithcode_id": "weibo-ner", "pretty_name": "Weibo NER", "dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "B-GPE.NAM", "1": "B-GPE.NOM", "2": "B-LOC.NAM", "3": "B-LOC.NOM", "4": "B-ORG.NAM", "5": "B-ORG.NOM", "6": "B-PER.NAM", "7": "B-PER.NOM", "8": "I-GPE.NAM", "9": "I-GPE.NOM", "10": "I-LOC.NAM", "11": "I-LOC.NOM", "12": "I-ORG.NAM", "13": "I-ORG.NOM", "14": "I-PER.NAM", "15": "I-PER.NOM", "16": "O"}}}}], "splits": [{"name": "train", "num_bytes": 1179589, "num_examples": 1350}, {"name": "validation", "num_bytes": 232380, "num_examples": 270}, {"name": "test", "num_bytes": 237407, "num_examples": 270}], "download_size": 750687, "dataset_size": 1649376}, "train-eval-index": [{"config": "default", "task": "token-classification", "task_id": "entity_extraction", "splits": {"train_split": "train", "eval_split": "test"}, "col_mapping": {"tokens": "tokens", "ner_tags": "tags"}, "metrics": [{"type": "seqeval", "name": "seqeval"}]}]}
2024-01-18T11:17:54+00:00
[]
[ "zh" ]
TAGS #task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Chinese #license-unknown #region-us
# Dataset Card for "Weibo NER" ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: None - Repository: URL - Paper: - Leaderboard: [If the dataset supports an active leaderboard, add link here]() - Point of Contact: ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @abhishekkrthakur for adding this dataset.
[ "# Dataset Card for \"Weibo NER\"", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: None\n- Repository: URL\n- Paper: \n- Leaderboard: [If the dataset supports an active leaderboard, add link here]()\n- Point of Contact:", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ "TAGS\n#task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Chinese #license-unknown #region-us \n", "# Dataset Card for \"Weibo NER\"", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: None\n- Repository: URL\n- Paper: \n- Leaderboard: [If the dataset supports an active leaderboard, add link here]()\n- Point of Contact:", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ 94, 11, 120, 45, 6, 10, 4, 6, 6, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 20 ]
[ "passage: TAGS\n#task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-Chinese #license-unknown #region-us \n# Dataset Card for \"Weibo NER\"## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: None\n- Repository: URL\n- Paper: \n- Leaderboard: [If the dataset supports an active leaderboard, add link here]()\n- Point of Contact:### Dataset Summary### Supported Tasks and Leaderboards### Languages## Dataset Structure### Data Instances### Data Fields### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators### Licensing Information### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
18844ae1ccd070b05345fa2d487adb8497462c4c
# Dataset Card for Cambridge English Write & Improve + LOCNESS Dataset ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.cl.cam.ac.uk/research/nl/bea2019st/#data - **Repository:** - **Paper:** https://www.aclweb.org/anthology/W19-4406/ - **Leaderboard:** https://competitions.codalab.org/competitions/20228#results - **Point of Contact:** ### Dataset Summary Write & Improve (Yannakoudakis et al., 2018) is an online web platform that assists non-native English students with their writing. Specifically, students from around the world submit letters, stories, articles and essays in response to various prompts, and the W&I system provides instant feedback. Since W&I went live in 2014, W&I annotators have manually annotated some of these submissions and assigned them a CEFR level. The LOCNESS corpus (Granger, 1998) consists of essays written by native English students. It was originally compiled by researchers at the Centre for English Corpus Linguistics at the University of Louvain. Since native English students also sometimes make mistakes, we asked the W&I annotators to annotate a subsection of LOCNESS so researchers can test the effectiveness of their systems on the full range of English levels and abilities. ### Supported Tasks and Leaderboards Grammatical error correction (GEC) is the task of automatically correcting grammatical errors in text; e.g. [I follows his advices -> I followed his advice]. It can be used to not only help language learners improve their writing skills, but also alert native speakers to accidental mistakes or typos. The aim of the task of this dataset is to correct all types of errors in written text. This includes grammatical, lexical and orthographical errors. The following Codalab competition contains the latest leaderboard, along with information on how to submit to the withheld W&I+LOCNESS test set: https://competitions.codalab.org/competitions/20228 ### Languages The dataset is in English. ## Dataset Structure ### Data Instances An example from the `wi` configuration: ``` { 'id': '1-140178', 'userid': '21251', 'cefr': 'A2.i', 'text': 'My town is a medium size city with eighty thousand inhabitants. It has a high density population because its small territory. Despite of it is an industrial city, there are many shops and department stores. I recommend visiting the artificial lake in the certer of the city which is surrounded by a park. Pasteries are very common and most of them offer the special dessert from the city. There are a comercial zone along the widest street of the city where you can find all kind of establishments: banks, bars, chemists, cinemas, pet shops, restaurants, fast food restaurants, groceries, travel agencies, supermarkets and others. Most of the shops have sales and offers at least three months of the year: January, June and August. The quality of the products and services are quite good, because there are a huge competition, however I suggest you taking care about some fakes or cheats.', 'edits': { 'start': [13, 77, 104, 126, 134, 256, 306, 375, 396, 402, 476, 484, 579, 671, 774, 804, 808, 826, 838, 850, 857, 862, 868], 'end': [24, 78, 104, 133, 136, 262, 315, 379, 399, 411, 480, 498, 588, 671, 777, 807, 810, 835, 845, 856, 861, 867, 873], 'text': ['medium-sized', '-', ' of', 'Although', '', 'center', None, 'of', 'is', 'commercial', 'kinds', 'businesses', 'grocers', ' in', 'is', 'is', '', '. However,', 'recommend', 'be', 'careful', 'of', ''] } } ``` An example from the `locness` configuration: ``` { 'id': '7-5819177', 'cefr': 'N', 'text': 'Boxing is a common, well known and well loved sport amongst most countries in the world however it is also punishing, dangerous and disliked to the extent that many people want it banned, possibly with good reason.\nBoxing is a dangerous sport, there are relatively common deaths, tragic injuries and even disease. All professional boxers are at risk from being killed in his next fight. If not killed then more likely paralysed. There have been a number of cases in the last ten years of the top few boxers having tragic losses throughout their ranks. This is just from the elite few, and theres more from those below them.\nMore deaths would occur through boxing if it were banned. The sport would go underground, there would be no safety measures like gloves, a doctor, paramedics or early stopping of the fight if someone looked unable to continue. With this going on the people taking part will be dangerous, and on the streets. Dangerous dogs who were trained to kill and maim in similar underound dog fights have already proved deadly to innocent people, the new boxers could be even more at risk.\nOnce boxing is banned and no-one grows up knowing it as acceptable there will be no interest in boxing and hopefully less all round interest in violence making towns and cities much safer places to live in, there will be less fighting outside pubs and clubs and less violent attacks with little or no reason.\nchange the rules of boxing slightly would much improve the safety risks of the sport and not detract form the entertainment. There are all sorts of proposals, lighter and more cushioning gloves could be worn, ban punches to the head, headguards worn or make fights shorter, as most of the serious injuries occur in the latter rounds, these would all show off the boxers skill and tallent and still be entertaining to watch.\nEven if a boxer is a success and manages not to be seriously hurt he still faces serious consequences in later life diseases that attack the brains have been known to set in as a direct result of boxing, even Muhamed Ali, who was infamous(?) both for his boxing and his quick-witted intelligence now has Alzheimer disease and can no longer do many everyday acts.\nMany other sports are more dangerous than boxing, motor sports and even mountaineering has risks that are real. Boxers chose to box, just as racing drivers drive.', 'edits': { 'start': [24, 39, 52, 87, 242, 371, 400, 528, 589, 713, 869, 992, 1058, 1169, 1209, 1219, 1255, 1308, 1386, 1412, 1513, 1569, 1661, 1731, 1744, 1781, 1792, 1901, 1951, 2038, 2131, 2149, 2247, 2286], 'end': [25, 40, 59, 95, 249, 374, 400, 538, 595, 713, 869, 1001, 1063, 1169, 1209, 1219, 1255, 1315, 1390, 1418, 1517, 1570, 1661, 1737, 1751, 1781, 1799, 1901, 1960, 2044, 2131, 2149, 2248, 2289], 'text': ['-', '-', 'in', '. However,', '. There', 'their', ',', 'among', "there's", ' and', ',', 'underground', '. The', ',', ',', ',', ',', '. There', 'for', 'Changing', 'from', ';', ',', 'later', '. These', "'", 'talent', ',', '. Diseases', '. Even', ',', "'s", ';', 'have'] } } ``` ### Data Fields The fields of the dataset are: - `id`: the id of the text as a string - `cefr`: the [CEFR level](https://www.cambridgeenglish.org/exams-and-tests/cefr/) of the text as a string - `userid`: id of the user - `text`: the text of the submission as a string - `edits`: the edits from W&I: - `start`: start indexes of each edit as a list of integers - `end`: end indexes of each edit as a list of integers - `text`: the text content of each edit as a list of strings - `from`: the original text of each edit as a list of strings ### Data Splits | name |train|validation| |----------|----:|---------:| | wi | 3000| 300| | locness | N/A| 50| ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Write & Improve License: ``` Cambridge English Write & Improve (CEWI) Dataset Licence Agreement 1. By downloading this dataset and licence, this licence agreement is entered into, effective this date, between you, the Licensee, and the University of Cambridge, the Licensor. 2. Copyright of the entire licensed dataset is held by the Licensor. No ownership or interest in the dataset is transferred to the Licensee. 3. The Licensor hereby grants the Licensee a non-exclusive non-transferable right to use the licensed dataset for non-commercial research and educational purposes. 4. Non-commercial purposes exclude without limitation any use of the licensed dataset or information derived from the dataset for or as part of a product or service which is sold, offered for sale, licensed, leased or rented. 5. The Licensee shall acknowledge use of the licensed dataset in all publications of research based on it, in whole or in part, through citation of the following publication: Helen Yannakoudakis, Øistein E. Andersen, Ardeshir Geranpayeh, Ted Briscoe and Diane Nicholls. 2018. Developing an automated writing placement system for ESL learners. Applied Measurement in Education. 6. The Licensee may publish excerpts of less than 100 words from the licensed dataset pursuant to clause 3. 7. The Licensor grants the Licensee this right to use the licensed dataset "as is". Licensor does not make, and expressly disclaims, any express or implied warranties, representations or endorsements of any kind whatsoever. 8. This Agreement shall be governed by and construed in accordance with the laws of England and the English courts shall have exclusive jurisdiction. ``` LOCNESS License: ``` LOCNESS Dataset Licence Agreement 1. The corpus is to be used for non-commercial purposes only 2. All publications on research partly or wholly based on the corpus should give credit to the Centre for English Corpus Linguistics (CECL), Université catholique de Louvain, Belgium. A scanned copy or offprint of the publication should also be sent to <sylviane.granger@uclouvain.be>. 3. No part of the corpus is to be distributed to a third party without specific authorization from CECL. The corpus can only be used by the person agreeing to the licence terms and researchers working in close collaboration with him/her or students under his/her supervision, attached to the same institution, within the framework of the research project. ``` ### Citation Information ``` @inproceedings{bryant-etal-2019-bea, title = "The {BEA}-2019 Shared Task on Grammatical Error Correction", author = "Bryant, Christopher and Felice, Mariano and Andersen, {\O}istein E. and Briscoe, Ted", booktitle = "Proceedings of the Fourteenth Workshop on Innovative Use of NLP for Building Educational Applications", month = aug, year = "2019", address = "Florence, Italy", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W19-4406", doi = "10.18653/v1/W19-4406", pages = "52--75", abstract = "This paper reports on the BEA-2019 Shared Task on Grammatical Error Correction (GEC). As with the CoNLL-2014 shared task, participants are required to correct all types of errors in test data. One of the main contributions of the BEA-2019 shared task is the introduction of a new dataset, the Write{\&}Improve+LOCNESS corpus, which represents a wider range of native and learner English levels and abilities. Another contribution is the introduction of tracks, which control the amount of annotated data available to participants. Systems are evaluated in terms of ERRANT F{\_}0.5, which allows us to report a much wider range of performance statistics. The competition was hosted on Codalab and remains open for further submissions on the blind test set.", } ``` ### Contributions Thanks to [@aseifert](https://github.com/aseifert) for adding this dataset.
wi_locness
[ "task_categories:text2text-generation", "annotations_creators:expert-generated", "language_creators:crowdsourced", "multilinguality:monolingual", "multilinguality:other-language-learner", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:other", "grammatical-error-correction", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["other"], "multilinguality": ["monolingual", "other-language-learner"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text2text-generation"], "task_ids": [], "paperswithcode_id": "locness-corpus", "pretty_name": "Cambridge English Write & Improve + LOCNESS", "config_names": ["locness", "wi"], "tags": ["grammatical-error-correction"], "dataset_info": [{"config_name": "default", "features": [{"name": "id", "dtype": "string"}, {"name": "userid", "dtype": "string"}, {"name": "cefr", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "edits", "sequence": [{"name": "start", "dtype": "int32"}, {"name": "end", "dtype": "int32"}, {"name": "text", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 4375795, "num_examples": 3000}, {"name": "validation", "num_bytes": 447055, "num_examples": 300}], "download_size": 6120469, "dataset_size": 4822850}, {"config_name": "wi", "features": [{"name": "id", "dtype": "string"}, {"name": "userid", "dtype": "string"}, {"name": "cefr", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "edits", "sequence": [{"name": "start", "dtype": "int32"}, {"name": "end", "dtype": "int32"}, {"name": "text", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 4375795, "num_examples": 3000}, {"name": "validation", "num_bytes": 447055, "num_examples": 300}], "download_size": 6120469, "dataset_size": 4822850}, {"config_name": "locness", "features": [{"name": "id", "dtype": "string"}, {"name": "cefr", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "edits", "sequence": [{"name": "start", "dtype": "int32"}, {"name": "end", "dtype": "int32"}, {"name": "text", "dtype": "string"}]}], "splits": [{"name": "validation", "num_bytes": 138176, "num_examples": 50}], "download_size": 6120469, "dataset_size": 138176}]}
2024-01-18T11:17:55+00:00
[]
[ "en" ]
TAGS #task_categories-text2text-generation #annotations_creators-expert-generated #language_creators-crowdsourced #multilinguality-monolingual #multilinguality-other-language-learner #size_categories-1K<n<10K #source_datasets-original #language-English #license-other #grammatical-error-correction #region-us
Dataset Card for Cambridge English Write & Improve + LOCNESS Dataset ==================================================================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: URL * Leaderboard: URL * Point of Contact: ### Dataset Summary Write & Improve (Yannakoudakis et al., 2018) is an online web platform that assists non-native English students with their writing. Specifically, students from around the world submit letters, stories, articles and essays in response to various prompts, and the W&I system provides instant feedback. Since W&I went live in 2014, W&I annotators have manually annotated some of these submissions and assigned them a CEFR level. The LOCNESS corpus (Granger, 1998) consists of essays written by native English students. It was originally compiled by researchers at the Centre for English Corpus Linguistics at the University of Louvain. Since native English students also sometimes make mistakes, we asked the W&I annotators to annotate a subsection of LOCNESS so researchers can test the effectiveness of their systems on the full range of English levels and abilities. ### Supported Tasks and Leaderboards Grammatical error correction (GEC) is the task of automatically correcting grammatical errors in text; e.g. [I follows his advices -> I followed his advice]. It can be used to not only help language learners improve their writing skills, but also alert native speakers to accidental mistakes or typos. The aim of the task of this dataset is to correct all types of errors in written text. This includes grammatical, lexical and orthographical errors. The following Codalab competition contains the latest leaderboard, along with information on how to submit to the withheld W&I+LOCNESS test set: URL ### Languages The dataset is in English. Dataset Structure ----------------- ### Data Instances An example from the 'wi' configuration: An example from the 'locness' configuration: ### Data Fields The fields of the dataset are: * 'id': the id of the text as a string * 'cefr': the CEFR level of the text as a string * 'userid': id of the user * 'text': the text of the submission as a string * 'edits': the edits from W&I: + 'start': start indexes of each edit as a list of integers + 'end': end indexes of each edit as a list of integers + 'text': the text content of each edit as a list of strings + 'from': the original text of each edit as a list of strings ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information Write & Improve License: LOCNESS License: ### Contributions Thanks to @aseifert for adding this dataset.
[ "### Dataset Summary\n\n\nWrite & Improve (Yannakoudakis et al., 2018) is an online web platform that assists non-native English students with their writing. Specifically, students from around the world submit letters, stories, articles and essays in response to various prompts, and the W&I system provides instant feedback. Since W&I went live in 2014, W&I annotators have manually annotated some of these submissions and assigned them a CEFR level.\n\n\nThe LOCNESS corpus (Granger, 1998) consists of essays written by native English students. It was originally compiled by researchers at the Centre for English Corpus Linguistics at the University of Louvain. Since native English students also sometimes make mistakes, we asked the W&I annotators to annotate a subsection of LOCNESS so researchers can test the effectiveness of their systems on the full range of English levels and abilities.", "### Supported Tasks and Leaderboards\n\n\nGrammatical error correction (GEC) is the task of automatically correcting grammatical errors in text; e.g. [I follows his advices -> I followed his advice]. It can be used to not only help language learners improve their writing skills, but also alert native speakers to accidental mistakes or typos.\n\n\nThe aim of the task of this dataset is to correct all types of errors in written text. This includes grammatical, lexical and orthographical errors.\n\n\nThe following Codalab competition contains the latest leaderboard, along with information on how to submit to the withheld W&I+LOCNESS test set: URL", "### Languages\n\n\nThe dataset is in English.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nAn example from the 'wi' configuration:\n\n\nAn example from the 'locness' configuration:", "### Data Fields\n\n\nThe fields of the dataset are:\n\n\n* 'id': the id of the text as a string\n* 'cefr': the CEFR level of the text as a string\n* 'userid': id of the user\n* 'text': the text of the submission as a string\n* 'edits': the edits from W&I:\n\t+ 'start': start indexes of each edit as a list of integers\n\t+ 'end': end indexes of each edit as a list of integers\n\t+ 'text': the text content of each edit as a list of strings\n\t+ 'from': the original text of each edit as a list of strings", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nWrite & Improve License:\n\n\nLOCNESS License:", "### Contributions\n\n\nThanks to @aseifert for adding this dataset." ]
[ "TAGS\n#task_categories-text2text-generation #annotations_creators-expert-generated #language_creators-crowdsourced #multilinguality-monolingual #multilinguality-other-language-learner #size_categories-1K<n<10K #source_datasets-original #language-English #license-other #grammatical-error-correction #region-us \n", "### Dataset Summary\n\n\nWrite & Improve (Yannakoudakis et al., 2018) is an online web platform that assists non-native English students with their writing. Specifically, students from around the world submit letters, stories, articles and essays in response to various prompts, and the W&I system provides instant feedback. Since W&I went live in 2014, W&I annotators have manually annotated some of these submissions and assigned them a CEFR level.\n\n\nThe LOCNESS corpus (Granger, 1998) consists of essays written by native English students. It was originally compiled by researchers at the Centre for English Corpus Linguistics at the University of Louvain. Since native English students also sometimes make mistakes, we asked the W&I annotators to annotate a subsection of LOCNESS so researchers can test the effectiveness of their systems on the full range of English levels and abilities.", "### Supported Tasks and Leaderboards\n\n\nGrammatical error correction (GEC) is the task of automatically correcting grammatical errors in text; e.g. [I follows his advices -> I followed his advice]. It can be used to not only help language learners improve their writing skills, but also alert native speakers to accidental mistakes or typos.\n\n\nThe aim of the task of this dataset is to correct all types of errors in written text. This includes grammatical, lexical and orthographical errors.\n\n\nThe following Codalab competition contains the latest leaderboard, along with information on how to submit to the withheld W&I+LOCNESS test set: URL", "### Languages\n\n\nThe dataset is in English.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nAn example from the 'wi' configuration:\n\n\nAn example from the 'locness' configuration:", "### Data Fields\n\n\nThe fields of the dataset are:\n\n\n* 'id': the id of the text as a string\n* 'cefr': the CEFR level of the text as a string\n* 'userid': id of the user\n* 'text': the text of the submission as a string\n* 'edits': the edits from W&I:\n\t+ 'start': start indexes of each edit as a list of integers\n\t+ 'end': end indexes of each edit as a list of integers\n\t+ 'text': the text content of each edit as a list of strings\n\t+ 'from': the original text of each edit as a list of strings", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nWrite & Improve License:\n\n\nLOCNESS License:", "### Contributions\n\n\nThanks to @aseifert for adding this dataset." ]
[ 101, 212, 158, 18, 25, 149, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 18, 17 ]
[ "passage: TAGS\n#task_categories-text2text-generation #annotations_creators-expert-generated #language_creators-crowdsourced #multilinguality-monolingual #multilinguality-other-language-learner #size_categories-1K<n<10K #source_datasets-original #language-English #license-other #grammatical-error-correction #region-us \n### Dataset Summary\n\n\nWrite & Improve (Yannakoudakis et al., 2018) is an online web platform that assists non-native English students with their writing. Specifically, students from around the world submit letters, stories, articles and essays in response to various prompts, and the W&I system provides instant feedback. Since W&I went live in 2014, W&I annotators have manually annotated some of these submissions and assigned them a CEFR level.\n\n\nThe LOCNESS corpus (Granger, 1998) consists of essays written by native English students. It was originally compiled by researchers at the Centre for English Corpus Linguistics at the University of Louvain. Since native English students also sometimes make mistakes, we asked the W&I annotators to annotate a subsection of LOCNESS so researchers can test the effectiveness of their systems on the full range of English levels and abilities.### Supported Tasks and Leaderboards\n\n\nGrammatical error correction (GEC) is the task of automatically correcting grammatical errors in text; e.g. [I follows his advices -> I followed his advice]. It can be used to not only help language learners improve their writing skills, but also alert native speakers to accidental mistakes or typos.\n\n\nThe aim of the task of this dataset is to correct all types of errors in written text. This includes grammatical, lexical and orthographical errors.\n\n\nThe following Codalab competition contains the latest leaderboard, along with information on how to submit to the withheld W&I+LOCNESS test set: URL### Languages\n\n\nThe dataset is in English.\n\n\nDataset Structure\n-----------------" ]
db171f1b7fedf4d3453e81297ff02f9915356d19
# Dataset Card for WIDER FACE ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** http://shuoyang1213.me/WIDERFACE/index.html - **Repository:** - **Paper:** [WIDER FACE: A Face Detection Benchmark](https://arxiv.org/abs/1511.06523) - **Leaderboard:** http://shuoyang1213.me/WIDERFACE/WiderFace_Results.html - **Point of Contact:** shuoyang.1213@gmail.com ### Dataset Summary WIDER FACE dataset is a face detection benchmark dataset, of which images are selected from the publicly available WIDER dataset. We choose 32,203 images and label 393,703 faces with a high degree of variability in scale, pose and occlusion as depicted in the sample images. WIDER FACE dataset is organized based on 61 event classes. For each event class, we randomly select 40%/10%/50% data as training, validation and testing sets. We adopt the same evaluation metric employed in the PASCAL VOC dataset. Similar to MALF and Caltech datasets, we do not release bounding box ground truth for the test images. Users are required to submit final prediction files, which we shall proceed to evaluate. ### Supported Tasks and Leaderboards - `face-detection`: The dataset can be used to train a model for Face Detection. More information on evaluating the model's performance can be found [here](http://shuoyang1213.me/WIDERFACE/WiderFace_Results.html). ### Languages English ## Dataset Structure ### Data Instances A data point comprises an image and its face annotations. ``` { 'image': <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=1024x755 at 0x19FA12186D8>, 'faces': { 'bbox': [ [178.0, 238.0, 55.0, 73.0], [248.0, 235.0, 59.0, 73.0], [363.0, 157.0, 59.0, 73.0], [468.0, 153.0, 53.0, 72.0], [629.0, 110.0, 56.0, 81.0], [745.0, 138.0, 55.0, 77.0] ], 'blur': [2, 2, 2, 2, 2, 2], 'expression': [0, 0, 0, 0, 0, 0], 'illumination': [0, 0, 0, 0, 0, 0], 'occlusion': [1, 2, 1, 2, 1, 2], 'pose': [0, 0, 0, 0, 0, 0], 'invalid': [False, False, False, False, False, False] } } ``` ### Data Fields - `image`: A `PIL.Image.Image` object containing the image. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]` - `faces`: a dictionary of face attributes for the faces present on the image - `bbox`: the bounding box of each face (in the [coco](https://albumentations.ai/docs/getting_started/bounding_boxes_augmentation/#coco) format) - `blur`: the blur level of each face, with possible values including `clear` (0), `normal` (1) and `heavy` - `expression`: the facial expression of each face, with possible values including `typical` (0) and `exaggerate` (1) - `illumination`: the lightning condition of each face, with possible values including `normal` (0) and `exaggerate` (1) - `occlusion`: the level of occlusion of each face, with possible values including `no` (0), `partial` (1) and `heavy` (2) - `pose`: the pose of each face, with possible values including `typical` (0) and `atypical` (1) - `invalid`: whether the image is valid or invalid. ### Data Splits The data is split into training, validation and testing set. WIDER FACE dataset is organized based on 61 event classes. For each event class, 40%/10%/50% data is randomly selected as training, validation and testing sets. The training set contains 12880 images, the validation set 3226 images and test set 16097 images. ## Dataset Creation ### Curation Rationale The curators state that the current face detection datasets typically contain a few thousand faces, with limited variations in pose, scale, facial expression, occlusion, and background clutters, making it difficult to assess for real world performance. They argue that the limitations of datasets have partially contributed to the failure of some algorithms in coping with heavy occlusion, small scale, and atypical pose. ### Source Data #### Initial Data Collection and Normalization WIDER FACE dataset is a subset of the WIDER dataset. The images in WIDER were collected in the following three steps: 1) Event categories were defined and chosen following the Large Scale Ontology for Multimedia (LSCOM) [22], which provides around 1000 concepts relevant to video event analysis. 2) Images are retrieved using search engines like Google and Bing. For each category, 1000-3000 images were collected. 3) The data were cleaned by manually examining all the images and filtering out images without human face. Then, similar images in each event category were removed to ensure large diversity in face appearance. A total of 32203 images are eventually included in the WIDER FACE dataset. #### Who are the source language producers? The images are selected from publicly available WIDER dataset. ### Annotations #### Annotation process The curators label the bounding boxes for all the recognizable faces in the WIDER FACE dataset. The bounding box is required to tightly contain the forehead, chin, and cheek.. If a face is occluded, they still label it with a bounding box but with an estimation on the scale of occlusion. Similar to the PASCAL VOC dataset [6], they assign an ’Ignore’ flag to the face which is very difficult to be recognized due to low resolution and small scale (10 pixels or less). After annotating the face bounding boxes, they further annotate the following attributes: pose (typical, atypical) and occlusion level (partial, heavy). Each annotation is labeled by one annotator and cross-checked by two different people. #### Who are the annotators? Shuo Yang, Ping Luo, Chen Change Loy and Xiaoou Tang. ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators Shuo Yang, Ping Luo, Chen Change Loy and Xiaoou Tang ### Licensing Information [Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0)](https://creativecommons.org/licenses/by-nc-nd/4.0/). ### Citation Information ``` @inproceedings{yang2016wider, Author = {Yang, Shuo and Luo, Ping and Loy, Chen Change and Tang, Xiaoou}, Booktitle = {IEEE Conference on Computer Vision and Pattern Recognition (CVPR)}, Title = {WIDER FACE: A Face Detection Benchmark}, Year = {2016}} ``` ### Contributions Thanks to [@mariosasko](https://github.com/mariosasko) for adding this dataset.
wider_face
[ "task_categories:object-detection", "task_ids:face-detection", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:extended|other-wider", "language:en", "license:cc-by-nc-nd-4.0", "arxiv:1511.06523", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found"], "language": ["en"], "license": ["cc-by-nc-nd-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["extended|other-wider"], "task_categories": ["object-detection"], "task_ids": ["face-detection"], "paperswithcode_id": "wider-face-1", "pretty_name": "WIDER FACE", "dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "faces", "sequence": [{"name": "bbox", "sequence": "float32", "length": 4}, {"name": "blur", "dtype": {"class_label": {"names": {"0": "clear", "1": "normal", "2": "heavy"}}}}, {"name": "expression", "dtype": {"class_label": {"names": {"0": "typical", "1": "exaggerate"}}}}, {"name": "illumination", "dtype": {"class_label": {"names": {"0": "normal", "1": "exaggerate "}}}}, {"name": "occlusion", "dtype": {"class_label": {"names": {"0": "no", "1": "partial", "2": "heavy"}}}}, {"name": "pose", "dtype": {"class_label": {"names": {"0": "typical", "1": "atypical"}}}}, {"name": "invalid", "dtype": "bool"}]}], "splits": [{"name": "train", "num_bytes": 12049881, "num_examples": 12880}, {"name": "test", "num_bytes": 3761103, "num_examples": 16097}, {"name": "validation", "num_bytes": 2998735, "num_examples": 3226}], "download_size": 3676086479, "dataset_size": 18809719}}
2024-01-18T11:17:56+00:00
[ "1511.06523" ]
[ "en" ]
TAGS #task_categories-object-detection #task_ids-face-detection #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-extended|other-wider #language-English #license-cc-by-nc-nd-4.0 #arxiv-1511.06523 #region-us
# Dataset Card for WIDER FACE ## Table of Contents - Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: WIDER FACE: A Face Detection Benchmark - Leaderboard: URL - Point of Contact: shuoyang.1213@URL ### Dataset Summary WIDER FACE dataset is a face detection benchmark dataset, of which images are selected from the publicly available WIDER dataset. We choose 32,203 images and label 393,703 faces with a high degree of variability in scale, pose and occlusion as depicted in the sample images. WIDER FACE dataset is organized based on 61 event classes. For each event class, we randomly select 40%/10%/50% data as training, validation and testing sets. We adopt the same evaluation metric employed in the PASCAL VOC dataset. Similar to MALF and Caltech datasets, we do not release bounding box ground truth for the test images. Users are required to submit final prediction files, which we shall proceed to evaluate. ### Supported Tasks and Leaderboards - 'face-detection': The dataset can be used to train a model for Face Detection. More information on evaluating the model's performance can be found here. ### Languages English ## Dataset Structure ### Data Instances A data point comprises an image and its face annotations. ### Data Fields - 'image': A 'PIL.Image.Image' object containing the image. Note that when accessing the image column: 'dataset[0]["image"]' the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the '"image"' column, *i.e.* 'dataset[0]["image"]' should always be preferred over 'dataset["image"][0]' - 'faces': a dictionary of face attributes for the faces present on the image - 'bbox': the bounding box of each face (in the coco format) - 'blur': the blur level of each face, with possible values including 'clear' (0), 'normal' (1) and 'heavy' - 'expression': the facial expression of each face, with possible values including 'typical' (0) and 'exaggerate' (1) - 'illumination': the lightning condition of each face, with possible values including 'normal' (0) and 'exaggerate' (1) - 'occlusion': the level of occlusion of each face, with possible values including 'no' (0), 'partial' (1) and 'heavy' (2) - 'pose': the pose of each face, with possible values including 'typical' (0) and 'atypical' (1) - 'invalid': whether the image is valid or invalid. ### Data Splits The data is split into training, validation and testing set. WIDER FACE dataset is organized based on 61 event classes. For each event class, 40%/10%/50% data is randomly selected as training, validation and testing sets. The training set contains 12880 images, the validation set 3226 images and test set 16097 images. ## Dataset Creation ### Curation Rationale The curators state that the current face detection datasets typically contain a few thousand faces, with limited variations in pose, scale, facial expression, occlusion, and background clutters, making it difficult to assess for real world performance. They argue that the limitations of datasets have partially contributed to the failure of some algorithms in coping with heavy occlusion, small scale, and atypical pose. ### Source Data #### Initial Data Collection and Normalization WIDER FACE dataset is a subset of the WIDER dataset. The images in WIDER were collected in the following three steps: 1) Event categories were defined and chosen following the Large Scale Ontology for Multimedia (LSCOM) [22], which provides around 1000 concepts relevant to video event analysis. 2) Images are retrieved using search engines like Google and Bing. For each category, 1000-3000 images were collected. 3) The data were cleaned by manually examining all the images and filtering out images without human face. Then, similar images in each event category were removed to ensure large diversity in face appearance. A total of 32203 images are eventually included in the WIDER FACE dataset. #### Who are the source language producers? The images are selected from publicly available WIDER dataset. ### Annotations #### Annotation process The curators label the bounding boxes for all the recognizable faces in the WIDER FACE dataset. The bounding box is required to tightly contain the forehead, chin, and cheek.. If a face is occluded, they still label it with a bounding box but with an estimation on the scale of occlusion. Similar to the PASCAL VOC dataset [6], they assign an ’Ignore’ flag to the face which is very difficult to be recognized due to low resolution and small scale (10 pixels or less). After annotating the face bounding boxes, they further annotate the following attributes: pose (typical, atypical) and occlusion level (partial, heavy). Each annotation is labeled by one annotator and cross-checked by two different people. #### Who are the annotators? Shuo Yang, Ping Luo, Chen Change Loy and Xiaoou Tang. ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators Shuo Yang, Ping Luo, Chen Change Loy and Xiaoou Tang ### Licensing Information Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0). ### Contributions Thanks to @mariosasko for adding this dataset.
[ "# Dataset Card for WIDER FACE", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper: WIDER FACE: A Face Detection Benchmark\n- Leaderboard: URL\n- Point of Contact: shuoyang.1213@URL", "### Dataset Summary\n\nWIDER FACE dataset is a face detection benchmark dataset, of which images are\nselected from the publicly available WIDER dataset. We choose 32,203 images and\nlabel 393,703 faces with a high degree of variability in scale, pose and\nocclusion as depicted in the sample images. WIDER FACE dataset is organized\nbased on 61 event classes. For each event class, we randomly select 40%/10%/50%\ndata as training, validation and testing sets. We adopt the same evaluation\nmetric employed in the PASCAL VOC dataset. Similar to MALF and Caltech datasets,\nwe do not release bounding box ground truth for the test images. Users are\nrequired to submit final prediction files, which we shall proceed to evaluate.", "### Supported Tasks and Leaderboards\n\n- 'face-detection': The dataset can be used to train a model for Face Detection. More information on evaluating the model's performance can be found here.", "### Languages\n\nEnglish", "## Dataset Structure", "### Data Instances\n\nA data point comprises an image and its face annotations.", "### Data Fields\n\n- 'image': A 'PIL.Image.Image' object containing the image. Note that when accessing the image column: 'dataset[0][\"image\"]' the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the '\"image\"' column, *i.e.* 'dataset[0][\"image\"]' should always be preferred over 'dataset[\"image\"][0]'\n- 'faces': a dictionary of face attributes for the faces present on the image\n - 'bbox': the bounding box of each face (in the coco format)\n - 'blur': the blur level of each face, with possible values including 'clear' (0), 'normal' (1) and 'heavy'\n - 'expression': the facial expression of each face, with possible values including 'typical' (0) and 'exaggerate' (1)\n - 'illumination': the lightning condition of each face, with possible values including 'normal' (0) and 'exaggerate' (1)\n - 'occlusion': the level of occlusion of each face, with possible values including 'no' (0), 'partial' (1) and 'heavy' (2)\n - 'pose': the pose of each face, with possible values including 'typical' (0) and 'atypical' (1)\n - 'invalid': whether the image is valid or invalid.", "### Data Splits\n\nThe data is split into training, validation and testing set. WIDER FACE dataset is organized\nbased on 61 event classes. For each event class, 40%/10%/50%\ndata is randomly selected as training, validation and testing sets. The training set contains 12880 images, the validation set 3226 images and test set 16097 images.", "## Dataset Creation", "### Curation Rationale\n\nThe curators state that the current face detection datasets typically contain a few thousand faces, with limited variations in pose, scale, facial expression, occlusion, and background clutters,\nmaking it difficult to assess for real world performance. They argue that the limitations of datasets have partially contributed to the failure of some algorithms in coping\nwith heavy occlusion, small scale, and atypical pose.", "### Source Data", "#### Initial Data Collection and Normalization\n\nWIDER FACE dataset is a subset of the WIDER dataset.\nThe images in WIDER were collected in the following three steps: 1) Event categories\nwere defined and chosen following the Large Scale Ontology for Multimedia (LSCOM) [22], which provides around 1000 concepts relevant to video event analysis. 2) Images\nare retrieved using search engines like Google and Bing. For\neach category, 1000-3000 images were collected. 3) The\ndata were cleaned by manually examining all the images\nand filtering out images without human face. Then, similar\nimages in each event category were removed to ensure large\ndiversity in face appearance. A total of 32203 images are\neventually included in the WIDER FACE dataset.", "#### Who are the source language producers?\n\nThe images are selected from publicly available WIDER dataset.", "### Annotations", "#### Annotation process\n\nThe curators label the bounding boxes for all\nthe recognizable faces in the WIDER FACE dataset. The\nbounding box is required to tightly contain the forehead,\nchin, and cheek.. If a face is occluded, they still label it with a bounding box but with an estimation on the scale of occlusion. Similar to the PASCAL VOC dataset [6], they assign an ’Ignore’ flag to the face\nwhich is very difficult to be recognized due to low resolution and small scale (10 pixels or less). After annotating\nthe face bounding boxes, they further annotate the following\nattributes: pose (typical, atypical) and occlusion level (partial, heavy). Each annotation is labeled by one annotator\nand cross-checked by two different people.", "#### Who are the annotators?\n\nShuo Yang, Ping Luo, Chen Change Loy and Xiaoou Tang.", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nShuo Yang, Ping Luo, Chen Change Loy and Xiaoou Tang", "### Licensing Information\n\nCreative Commons Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0).", "### Contributions\n\nThanks to @mariosasko for adding this dataset." ]
[ "TAGS\n#task_categories-object-detection #task_ids-face-detection #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-extended|other-wider #language-English #license-cc-by-nc-nd-4.0 #arxiv-1511.06523 #region-us \n", "# Dataset Card for WIDER FACE", "## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper: WIDER FACE: A Face Detection Benchmark\n- Leaderboard: URL\n- Point of Contact: shuoyang.1213@URL", "### Dataset Summary\n\nWIDER FACE dataset is a face detection benchmark dataset, of which images are\nselected from the publicly available WIDER dataset. We choose 32,203 images and\nlabel 393,703 faces with a high degree of variability in scale, pose and\nocclusion as depicted in the sample images. WIDER FACE dataset is organized\nbased on 61 event classes. For each event class, we randomly select 40%/10%/50%\ndata as training, validation and testing sets. We adopt the same evaluation\nmetric employed in the PASCAL VOC dataset. Similar to MALF and Caltech datasets,\nwe do not release bounding box ground truth for the test images. Users are\nrequired to submit final prediction files, which we shall proceed to evaluate.", "### Supported Tasks and Leaderboards\n\n- 'face-detection': The dataset can be used to train a model for Face Detection. More information on evaluating the model's performance can be found here.", "### Languages\n\nEnglish", "## Dataset Structure", "### Data Instances\n\nA data point comprises an image and its face annotations.", "### Data Fields\n\n- 'image': A 'PIL.Image.Image' object containing the image. Note that when accessing the image column: 'dataset[0][\"image\"]' the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the '\"image\"' column, *i.e.* 'dataset[0][\"image\"]' should always be preferred over 'dataset[\"image\"][0]'\n- 'faces': a dictionary of face attributes for the faces present on the image\n - 'bbox': the bounding box of each face (in the coco format)\n - 'blur': the blur level of each face, with possible values including 'clear' (0), 'normal' (1) and 'heavy'\n - 'expression': the facial expression of each face, with possible values including 'typical' (0) and 'exaggerate' (1)\n - 'illumination': the lightning condition of each face, with possible values including 'normal' (0) and 'exaggerate' (1)\n - 'occlusion': the level of occlusion of each face, with possible values including 'no' (0), 'partial' (1) and 'heavy' (2)\n - 'pose': the pose of each face, with possible values including 'typical' (0) and 'atypical' (1)\n - 'invalid': whether the image is valid or invalid.", "### Data Splits\n\nThe data is split into training, validation and testing set. WIDER FACE dataset is organized\nbased on 61 event classes. For each event class, 40%/10%/50%\ndata is randomly selected as training, validation and testing sets. The training set contains 12880 images, the validation set 3226 images and test set 16097 images.", "## Dataset Creation", "### Curation Rationale\n\nThe curators state that the current face detection datasets typically contain a few thousand faces, with limited variations in pose, scale, facial expression, occlusion, and background clutters,\nmaking it difficult to assess for real world performance. They argue that the limitations of datasets have partially contributed to the failure of some algorithms in coping\nwith heavy occlusion, small scale, and atypical pose.", "### Source Data", "#### Initial Data Collection and Normalization\n\nWIDER FACE dataset is a subset of the WIDER dataset.\nThe images in WIDER were collected in the following three steps: 1) Event categories\nwere defined and chosen following the Large Scale Ontology for Multimedia (LSCOM) [22], which provides around 1000 concepts relevant to video event analysis. 2) Images\nare retrieved using search engines like Google and Bing. For\neach category, 1000-3000 images were collected. 3) The\ndata were cleaned by manually examining all the images\nand filtering out images without human face. Then, similar\nimages in each event category were removed to ensure large\ndiversity in face appearance. A total of 32203 images are\neventually included in the WIDER FACE dataset.", "#### Who are the source language producers?\n\nThe images are selected from publicly available WIDER dataset.", "### Annotations", "#### Annotation process\n\nThe curators label the bounding boxes for all\nthe recognizable faces in the WIDER FACE dataset. The\nbounding box is required to tightly contain the forehead,\nchin, and cheek.. If a face is occluded, they still label it with a bounding box but with an estimation on the scale of occlusion. Similar to the PASCAL VOC dataset [6], they assign an ’Ignore’ flag to the face\nwhich is very difficult to be recognized due to low resolution and small scale (10 pixels or less). After annotating\nthe face bounding boxes, they further annotate the following\nattributes: pose (typical, atypical) and occlusion level (partial, heavy). Each annotation is labeled by one annotator\nand cross-checked by two different people.", "#### Who are the annotators?\n\nShuo Yang, Ping Luo, Chen Change Loy and Xiaoou Tang.", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nShuo Yang, Ping Luo, Chen Change Loy and Xiaoou Tang", "### Licensing Information\n\nCreative Commons Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0).", "### Contributions\n\nThanks to @mariosasko for adding this dataset." ]
[ 109, 9, 125, 46, 176, 48, 5, 6, 20, 342, 81, 5, 102, 4, 167, 23, 5, 189, 25, 8, 8, 7, 8, 7, 5, 21, 30, 17 ]
[ "passage: TAGS\n#task_categories-object-detection #task_ids-face-detection #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-extended|other-wider #language-English #license-cc-by-nc-nd-4.0 #arxiv-1511.06523 #region-us \n# Dataset Card for WIDER FACE## Table of Contents\n- Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper: WIDER FACE: A Face Detection Benchmark\n- Leaderboard: URL\n- Point of Contact: shuoyang.1213@URL### Dataset Summary\n\nWIDER FACE dataset is a face detection benchmark dataset, of which images are\nselected from the publicly available WIDER dataset. We choose 32,203 images and\nlabel 393,703 faces with a high degree of variability in scale, pose and\nocclusion as depicted in the sample images. WIDER FACE dataset is organized\nbased on 61 event classes. For each event class, we randomly select 40%/10%/50%\ndata as training, validation and testing sets. We adopt the same evaluation\nmetric employed in the PASCAL VOC dataset. Similar to MALF and Caltech datasets,\nwe do not release bounding box ground truth for the test images. Users are\nrequired to submit final prediction files, which we shall proceed to evaluate.", "passage: ### Supported Tasks and Leaderboards\n\n- 'face-detection': The dataset can be used to train a model for Face Detection. More information on evaluating the model's performance can be found here.### Languages\n\nEnglish## Dataset Structure### Data Instances\n\nA data point comprises an image and its face annotations.### Data Fields\n\n- 'image': A 'PIL.Image.Image' object containing the image. Note that when accessing the image column: 'dataset[0][\"image\"]' the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the '\"image\"' column, *i.e.* 'dataset[0][\"image\"]' should always be preferred over 'dataset[\"image\"][0]'\n- 'faces': a dictionary of face attributes for the faces present on the image\n - 'bbox': the bounding box of each face (in the coco format)\n - 'blur': the blur level of each face, with possible values including 'clear' (0), 'normal' (1) and 'heavy'\n - 'expression': the facial expression of each face, with possible values including 'typical' (0) and 'exaggerate' (1)\n - 'illumination': the lightning condition of each face, with possible values including 'normal' (0) and 'exaggerate' (1)\n - 'occlusion': the level of occlusion of each face, with possible values including 'no' (0), 'partial' (1) and 'heavy' (2)\n - 'pose': the pose of each face, with possible values including 'typical' (0) and 'atypical' (1)\n - 'invalid': whether the image is valid or invalid.### Data Splits\n\nThe data is split into training, validation and testing set. WIDER FACE dataset is organized\nbased on 61 event classes. For each event class, 40%/10%/50%\ndata is randomly selected as training, validation and testing sets. The training set contains 12880 images, the validation set 3226 images and test set 16097 images.## Dataset Creation", "passage: ### Curation Rationale\n\nThe curators state that the current face detection datasets typically contain a few thousand faces, with limited variations in pose, scale, facial expression, occlusion, and background clutters,\nmaking it difficult to assess for real world performance. They argue that the limitations of datasets have partially contributed to the failure of some algorithms in coping\nwith heavy occlusion, small scale, and atypical pose.### Source Data#### Initial Data Collection and Normalization\n\nWIDER FACE dataset is a subset of the WIDER dataset.\nThe images in WIDER were collected in the following three steps: 1) Event categories\nwere defined and chosen following the Large Scale Ontology for Multimedia (LSCOM) [22], which provides around 1000 concepts relevant to video event analysis. 2) Images\nare retrieved using search engines like Google and Bing. For\neach category, 1000-3000 images were collected. 3) The\ndata were cleaned by manually examining all the images\nand filtering out images without human face. Then, similar\nimages in each event category were removed to ensure large\ndiversity in face appearance. A total of 32203 images are\neventually included in the WIDER FACE dataset.#### Who are the source language producers?\n\nThe images are selected from publicly available WIDER dataset.### Annotations#### Annotation process\n\nThe curators label the bounding boxes for all\nthe recognizable faces in the WIDER FACE dataset. The\nbounding box is required to tightly contain the forehead,\nchin, and cheek.. If a face is occluded, they still label it with a bounding box but with an estimation on the scale of occlusion. Similar to the PASCAL VOC dataset [6], they assign an ’Ignore’ flag to the face\nwhich is very difficult to be recognized due to low resolution and small scale (10 pixels or less). After annotating\nthe face bounding boxes, they further annotate the following\nattributes: pose (typical, atypical) and occlusion level (partial, heavy). Each annotation is labeled by one annotator\nand cross-checked by two different people.#### Who are the annotators?\n\nShuo Yang, Ping Luo, Chen Change Loy and Xiaoou Tang.### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators\n\nShuo Yang, Ping Luo, Chen Change Loy and Xiaoou Tang### Licensing Information\n\nCreative Commons Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0)." ]
7b21a2e64b90323b2d3d1b81aa349bb4bc76d9bf
# Dataset Card for "wiki40b" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://research.google/pubs/pub49029/](https://research.google/pubs/pub49029/) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 0.00 MB - **Size of the generated dataset:** 10.47 GB - **Total amount of disk used:** 10.47 GB ### Dataset Summary Clean-up text for 40+ Wikipedia languages editions of pages correspond to entities. The datasets have train/dev/test splits per language. The dataset is cleaned up by page filtering to remove disambiguation pages, redirect pages, deleted pages, and non-entity pages. Each example contains the wikidata id of the entity, and the full Wikipedia article after page processing that removes non-content sections and structured objects. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### en - **Size of downloaded dataset files:** 0.00 MB - **Size of the generated dataset:** 10.47 GB - **Total amount of disk used:** 10.47 GB An example of 'train' looks as follows. ``` ``` ### Data Fields The data fields are the same among all splits. #### en - `wikidata_id`: a `string` feature. - `text`: a `string` feature. - `version_id`: a `string` feature. ### Data Splits |name| train |validation| test | |----|------:|---------:|-----:| |en |2926536| 163597|162274| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` ``` ### Contributions Thanks to [@jplu](https://github.com/jplu), [@patrickvonplaten](https://github.com/patrickvonplaten), [@thomwolf](https://github.com/thomwolf), [@albertvillanova](https://github.com/albertvillanova), [@lhoestq](https://github.com/lhoestq) for adding this dataset.
wiki40b
[ "language:en", "region:us" ]
2022-03-02T23:29:22+00:00
{"language": ["en"], "paperswithcode_id": "wiki-40b", "pretty_name": "Wiki-40B", "dataset_info": {"features": [{"name": "wikidata_id", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "version_id", "dtype": "string"}], "config_name": "en", "splits": [{"name": "train", "num_bytes": 9423623904, "num_examples": 2926536}, {"name": "validation", "num_bytes": 527383016, "num_examples": 163597}, {"name": "test", "num_bytes": 522219464, "num_examples": 162274}], "download_size": 0, "dataset_size": 10473226384}}
2024-01-18T11:17:58+00:00
[]
[ "en" ]
TAGS #language-English #region-us
Dataset Card for "wiki40b" ========================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: * Point of Contact: * Size of downloaded dataset files: 0.00 MB * Size of the generated dataset: 10.47 GB * Total amount of disk used: 10.47 GB ### Dataset Summary Clean-up text for 40+ Wikipedia languages editions of pages correspond to entities. The datasets have train/dev/test splits per language. The dataset is cleaned up by page filtering to remove disambiguation pages, redirect pages, deleted pages, and non-entity pages. Each example contains the wikidata id of the entity, and the full Wikipedia article after page processing that removes non-content sections and structured objects. ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- ### Data Instances #### en * Size of downloaded dataset files: 0.00 MB * Size of the generated dataset: 10.47 GB * Total amount of disk used: 10.47 GB An example of 'train' looks as follows. ### Data Fields The data fields are the same among all splits. #### en * 'wikidata\_id': a 'string' feature. * 'text': a 'string' feature. * 'version\_id': a 'string' feature. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @jplu, @patrickvonplaten, @thomwolf, @albertvillanova, @lhoestq for adding this dataset.
[ "### Dataset Summary\n\n\nClean-up text for 40+ Wikipedia languages editions of pages\ncorrespond to entities. The datasets have train/dev/test splits per language.\nThe dataset is cleaned up by page filtering to remove disambiguation pages,\nredirect pages, deleted pages, and non-entity pages. Each example contains the\nwikidata id of the entity, and the full Wikipedia article after page processing\nthat removes non-content sections and structured objects.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### en\n\n\n* Size of downloaded dataset files: 0.00 MB\n* Size of the generated dataset: 10.47 GB\n* Total amount of disk used: 10.47 GB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### en\n\n\n* 'wikidata\\_id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'version\\_id': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @jplu, @patrickvonplaten, @thomwolf, @albertvillanova, @lhoestq for adding this dataset." ]
[ "TAGS\n#language-English #region-us \n", "### Dataset Summary\n\n\nClean-up text for 40+ Wikipedia languages editions of pages\ncorrespond to entities. The datasets have train/dev/test splits per language.\nThe dataset is cleaned up by page filtering to remove disambiguation pages,\nredirect pages, deleted pages, and non-entity pages. Each example contains the\nwikidata id of the entity, and the full Wikipedia article after page processing\nthat removes non-content sections and structured objects.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### en\n\n\n* Size of downloaded dataset files: 0.00 MB\n* Size of the generated dataset: 10.47 GB\n* Total amount of disk used: 10.47 GB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### en\n\n\n* 'wikidata\\_id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'version\\_id': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @jplu, @patrickvonplaten, @thomwolf, @albertvillanova, @lhoestq for adding this dataset." ]
[ 10, 110, 10, 11, 6, 49, 17, 43, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 6, 39 ]
[ "passage: TAGS\n#language-English #region-us \n### Dataset Summary\n\n\nClean-up text for 40+ Wikipedia languages editions of pages\ncorrespond to entities. The datasets have train/dev/test splits per language.\nThe dataset is cleaned up by page filtering to remove disambiguation pages,\nredirect pages, deleted pages, and non-entity pages. Each example contains the\nwikidata id of the entity, and the full Wikipedia article after page processing\nthat removes non-content sections and structured objects.### Supported Tasks and Leaderboards### Languages\n\n\nDataset Structure\n-----------------### Data Instances#### en\n\n\n* Size of downloaded dataset files: 0.00 MB\n* Size of the generated dataset: 10.47 GB\n* Total amount of disk used: 10.47 GB\n\n\nAn example of 'train' looks as follows.### Data Fields\n\n\nThe data fields are the same among all splits.#### en\n\n\n* 'wikidata\\_id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'version\\_id': a 'string' feature.### Data Splits\n\n\n\nDataset Creation\n----------------### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------### Social Impact of Dataset### Discussion of Biases### Other Known Limitations\n\n\nAdditional Information\n----------------------### Dataset Curators### Licensing Information### Contributions\n\n\nThanks to @jplu, @patrickvonplaten, @thomwolf, @albertvillanova, @lhoestq for adding this dataset." ]
07b2eb09ff169e12181d6f25e883c03518212a3c
# Dataset Card for WikiAsp ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Wiki Asp](https://github.com/neulab/wikiasp) - **Repository:** [GitHub](https://github.com/neulab/wikiasp) - **Paper:** [WikiAsp: A Dataset for Multi-domain Aspect-based Summarization](https://arxiv.org/abs/2011.07832) ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances An example from the "plant" configuration: ``` { 'exid': 'train-78-8', 'inputs': ['< EOT > calcareous rocks and barrens , wooded cliff edges .', 'plant an erect short - lived perennial ( or biennial ) herb whose slender leafy stems radiate from the base , and are 3 - 5 dm tall , giving it a bushy appearance .', 'leaves densely hairy , grayish - green , simple and alternate on the stem .', 'flowers are bright yellow to yellow - orange , cross - shaped , each having 4 spatula - shaped petals about 5 mm long .', 'fruit is a nearly globe - shaped capsule , about 3 mm in diameter , with 1 or 2 seeds in each cell .', 'flowering period : early april to late may .', 'even though there are many members of the mustard family in the range of this species , no other plant shares this combination of characters : bright yellow flowers , grayish - green stems and foliage , globe - shaped fruits with a long style , perennial habit , and the habitat of limestone rocky cliffs .', 'timber removal may be beneficial and even needed to maintain the open character of the habitat for this species .', 'hand removal of trees in the vicinity of the population is necessary to avoid impacts from timber operations .', 'southwest indiana , north central kentucky , and north central tennessee .', 'email : naturepreserves @ ky . gov feedback naturepreserves @ ky . gov | about the agency | about this site copyright © 2003 - 2013 commonwealth of kentucky .', 'all rights reserved .', '<EOS>' ], 'targets': [ ['description', 'physaria globosa is a small plant covered with dense hairs giving it a grayish appearance . it produces yellow flowers in the spring , and its fruit is globe - shaped . its preferred habitat is dry limestone cliffs , barrens , cedar glades , steep wooded slopes , and talus areas . some have also been found in areas of deeper soil and roadsides .' ], ['conservation', 'the population fluctuates year to year , but on average there are about 2000 living plants at any one time , divided among 33 known locations . threats include forms of habitat degradation and destruction , including road construction and grading , mowing , dumping , herbicides , alteration of waterways , livestock damage , and invasive species of plants such as japanese honeysuckle , garlic mustard , alsike clover , sweet clover , meadow fescue , and multiflora rose . all populations are considered vulnerable to extirpation .' ] ] } ``` ### Data Fields - `exid`: a unique identifier - `input`: the cited references and consists of tokenized sentences (with NLTK) - `targets`: a list of aspect-based summaries, where each element is a pair of a) the target aspect and b) the aspect-based summary ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@katnoria](https://github.com/katnoria) for adding this dataset.
wiki_asp
[ "task_categories:summarization", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:cc-by-sa-4.0", "aspect-based-summarization", "arxiv:2011.07832", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["cc-by-sa-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["summarization"], "task_ids": [], "paperswithcode_id": "wikiasp", "pretty_name": "WikiAsp", "tags": ["aspect-based-summarization"], "dataset_info": [{"config_name": "album", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 1907323642, "num_examples": 24434}, {"name": "test", "num_bytes": 232999001, "num_examples": 3038}, {"name": "validation", "num_bytes": 234990092, "num_examples": 3104}], "download_size": 644173065, "dataset_size": 2375312735}, {"config_name": "animal", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 497474133, "num_examples": 16540}, {"name": "test", "num_bytes": 61315970, "num_examples": 2007}, {"name": "validation", "num_bytes": 57943532, "num_examples": 2005}], "download_size": 150974930, "dataset_size": 616733635}, {"config_name": "artist", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 1876134255, "num_examples": 26754}, {"name": "test", "num_bytes": 237751553, "num_examples": 3329}, {"name": "validation", "num_bytes": 223240910, "num_examples": 3194}], "download_size": 626686303, "dataset_size": 2337126718}, {"config_name": "building", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 1100057273, "num_examples": 20449}, {"name": "test", "num_bytes": 134357678, "num_examples": 2482}, {"name": "validation", "num_bytes": 139387376, "num_examples": 2607}], "download_size": 346224042, "dataset_size": 1373802327}, {"config_name": "company", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 1606057076, "num_examples": 24353}, {"name": "test", "num_bytes": 199282041, "num_examples": 3029}, {"name": "validation", "num_bytes": 200498778, "num_examples": 2946}], "download_size": 504194353, "dataset_size": 2005837895}, {"config_name": "educational_institution", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 1623000534, "num_examples": 17634}, {"name": "test", "num_bytes": 200476681, "num_examples": 2267}, {"name": "validation", "num_bytes": 203262430, "num_examples": 2141}], "download_size": 471033992, "dataset_size": 2026739645}, {"config_name": "event", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 748201660, "num_examples": 6475}, {"name": "test", "num_bytes": 96212295, "num_examples": 828}, {"name": "validation", "num_bytes": 97431395, "num_examples": 807}], "download_size": 240072903, "dataset_size": 941845350}, {"config_name": "film", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 2370068027, "num_examples": 32129}, {"name": "test", "num_bytes": 294918370, "num_examples": 3981}, {"name": "validation", "num_bytes": 290240851, "num_examples": 4014}], "download_size": 808231638, "dataset_size": 2955227248}, {"config_name": "group", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 1025166800, "num_examples": 11966}, {"name": "test", "num_bytes": 114239405, "num_examples": 1444}, {"name": "validation", "num_bytes": 120863870, "num_examples": 1462}], "download_size": 344498865, "dataset_size": 1260270075}, {"config_name": "historic_place", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 256158020, "num_examples": 4919}, {"name": "test", "num_bytes": 31201154, "num_examples": 600}, {"name": "validation", "num_bytes": 29058067, "num_examples": 601}], "download_size": 77289509, "dataset_size": 316417241}, {"config_name": "infrastructure", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 1124486451, "num_examples": 17226}, {"name": "test", "num_bytes": 134820330, "num_examples": 2091}, {"name": "validation", "num_bytes": 125193140, "num_examples": 1984}], "download_size": 328804337, "dataset_size": 1384499921}, {"config_name": "mean_of_transportation", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 650424738, "num_examples": 9277}, {"name": "test", "num_bytes": 89759392, "num_examples": 1170}, {"name": "validation", "num_bytes": 88440901, "num_examples": 1215}], "download_size": 210234418, "dataset_size": 828625031}, {"config_name": "office_holder", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 1643899203, "num_examples": 18177}, {"name": "test", "num_bytes": 207433317, "num_examples": 2333}, {"name": "validation", "num_bytes": 202624275, "num_examples": 2218}], "download_size": 524721727, "dataset_size": 2053956795}, {"config_name": "plant", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 239150885, "num_examples": 6107}, {"name": "test", "num_bytes": 31340125, "num_examples": 774}, {"name": "validation", "num_bytes": 28752150, "num_examples": 786}], "download_size": 77890632, "dataset_size": 299243160}, {"config_name": "single", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 1277277277, "num_examples": 14217}, {"name": "test", "num_bytes": 152328537, "num_examples": 1712}, {"name": "validation", "num_bytes": 160312594, "num_examples": 1734}], "download_size": 429214401, "dataset_size": 1589918408}, {"config_name": "soccer_player", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 604502541, "num_examples": 17599}, {"name": "test", "num_bytes": 72820378, "num_examples": 2280}, {"name": "validation", "num_bytes": 76705685, "num_examples": 2150}], "download_size": 193347234, "dataset_size": 754028604}, {"config_name": "software", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 1122906186, "num_examples": 13516}, {"name": "test", "num_bytes": 133717992, "num_examples": 1638}, {"name": "validation", "num_bytes": 134578157, "num_examples": 1637}], "download_size": 356764908, "dataset_size": 1391202335}, {"config_name": "television_show", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 893325347, "num_examples": 8717}, {"name": "test", "num_bytes": 115155155, "num_examples": 1072}, {"name": "validation", "num_bytes": 119461892, "num_examples": 1128}], "download_size": 302093407, "dataset_size": 1127942394}, {"config_name": "town", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 772504751, "num_examples": 14818}, {"name": "test", "num_bytes": 100975827, "num_examples": 1831}, {"name": "validation", "num_bytes": 101522638, "num_examples": 1911}], "download_size": 243261734, "dataset_size": 975003216}, {"config_name": "written_work", "features": [{"name": "exid", "dtype": "string"}, {"name": "inputs", "sequence": "string"}, {"name": "targets", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 1491395960, "num_examples": 15065}, {"name": "test", "num_bytes": 189537205, "num_examples": 1931}, {"name": "validation", "num_bytes": 185707567, "num_examples": 1843}], "download_size": 498307235, "dataset_size": 1866640732}]}
2024-01-18T11:17:59+00:00
[ "2011.07832" ]
[ "en" ]
TAGS #task_categories-summarization #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-sa-4.0 #aspect-based-summarization #arxiv-2011.07832 #region-us
# Dataset Card for WikiAsp ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: Wiki Asp - Repository: GitHub - Paper: WikiAsp: A Dataset for Multi-domain Aspect-based Summarization ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances An example from the "plant" configuration: ### Data Fields - 'exid': a unique identifier - 'input': the cited references and consists of tokenized sentences (with NLTK) - 'targets': a list of aspect-based summaries, where each element is a pair of a) the target aspect and b) the aspect-based summary ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @katnoria for adding this dataset.
[ "# Dataset Card for WikiAsp", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Wiki Asp\n- Repository: GitHub\n- Paper: WikiAsp: A Dataset for Multi-domain Aspect-based Summarization", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances\n\nAn example from the \"plant\" configuration:", "### Data Fields\n\n- 'exid': a unique identifier\n- 'input': the cited references and consists of tokenized sentences (with NLTK) \n- 'targets': a list of aspect-based summaries, where each element is a pair of a) the target aspect and b) the aspect-based summary", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @katnoria for adding this dataset." ]
[ "TAGS\n#task_categories-summarization #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-sa-4.0 #aspect-based-summarization #arxiv-2011.07832 #region-us \n", "# Dataset Card for WikiAsp", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: Wiki Asp\n- Repository: GitHub\n- Paper: WikiAsp: A Dataset for Multi-domain Aspect-based Summarization", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances\n\nAn example from the \"plant\" configuration:", "### Data Fields\n\n- 'exid': a unique identifier\n- 'input': the cited references and consists of tokenized sentences (with NLTK) \n- 'targets': a list of aspect-based summaries, where each element is a pair of a) the target aspect and b) the aspect-based summary", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @katnoria for adding this dataset." ]
[ 100, 8, 120, 39, 6, 10, 4, 6, 15, 78, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 17 ]
[ "passage: TAGS\n#task_categories-summarization #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-sa-4.0 #aspect-based-summarization #arxiv-2011.07832 #region-us \n# Dataset Card for WikiAsp## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: Wiki Asp\n- Repository: GitHub\n- Paper: WikiAsp: A Dataset for Multi-domain Aspect-based Summarization### Dataset Summary### Supported Tasks and Leaderboards### Languages## Dataset Structure### Data Instances\n\nAn example from the \"plant\" configuration:### Data Fields\n\n- 'exid': a unique identifier\n- 'input': the cited references and consists of tokenized sentences (with NLTK) \n- 'targets': a list of aspect-based summaries, where each element is a pair of a) the target aspect and b) the aspect-based summary### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators### Licensing Information" ]
b38e6e03abd0dc7d42008eb14eac6a91ccb8877d
# Dataset Card for WikiAtomicEdits ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** None - **Repository:** https://github.com/google-research-datasets/wiki-atomic-edits - **Paper:** https://www.aclweb.org/anthology/D18-1028/ - **Leaderboard:** [More Information Needed] - **Point of Contact:** [More Information Needed] ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The languages in the dataset are: - de - en - es - fr - it - jp: Japanese (`ja`) - ru - zh ## Dataset Structure ### Data Instances Here are some examples of questions and facts: ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@abhishekkrthakur](https://github.com/abhishekkrthakur) for adding this dataset.
wiki_atomic_edits
[ "task_categories:summarization", "annotations_creators:found", "language_creators:found", "multilinguality:multilingual", "size_categories:100K<n<1M", "size_categories:10M<n<100M", "size_categories:1M<n<10M", "source_datasets:original", "language:de", "language:en", "language:es", "language:fr", "language:it", "language:ja", "language:ru", "language:zh", "license:cc-by-sa-4.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["de", "en", "es", "fr", "it", "ja", "ru", "zh"], "license": ["cc-by-sa-4.0"], "multilinguality": ["multilingual"], "size_categories": ["100K<n<1M", "10M<n<100M", "1M<n<10M"], "source_datasets": ["original"], "task_categories": ["summarization"], "task_ids": [], "paperswithcode_id": "wikiatomicedits", "pretty_name": "WikiAtomicEdits", "config_names": ["chinese_deletions", "chinese_insertions", "english_deletions", "english_insertions", "french_deletions", "french_insertions", "german_deletions", "german_insertions", "italian_deletions", "italian_insertions", "japanese_deletions", "japanese_insertions", "russian_deletions", "russian_insertions", "spanish_deletions", "spanish_insertions"], "dataset_info": [{"config_name": "german_insertions", "features": [{"name": "id", "dtype": "int32"}, {"name": "base_sentence", "dtype": "string"}, {"name": "phrase", "dtype": "string"}, {"name": "edited_sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1072443082, "num_examples": 3343403}], "download_size": 274280387, "dataset_size": 1072443082}, {"config_name": "german_deletions", "features": [{"name": "id", "dtype": "int32"}, {"name": "base_sentence", "dtype": "string"}, {"name": "phrase", "dtype": "string"}, {"name": "edited_sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 624070402, "num_examples": 1994329}], "download_size": 160133549, "dataset_size": 624070402}, {"config_name": "english_insertions", "features": [{"name": "id", "dtype": "int32"}, {"name": "base_sentence", "dtype": "string"}, {"name": "phrase", "dtype": "string"}, {"name": "edited_sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 4258411914, "num_examples": 13737796}], "download_size": 1090652177, "dataset_size": 4258411914}, {"config_name": "english_deletions", "features": [{"name": "id", "dtype": "int32"}, {"name": "base_sentence", "dtype": "string"}, {"name": "phrase", "dtype": "string"}, {"name": "edited_sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 2865754626, "num_examples": 9352389}], "download_size": 736560902, "dataset_size": 2865754626}, {"config_name": "spanish_insertions", "features": [{"name": "id", "dtype": "int32"}, {"name": "base_sentence", "dtype": "string"}, {"name": "phrase", "dtype": "string"}, {"name": "edited_sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 481145004, "num_examples": 1380934}], "download_size": 118837934, "dataset_size": 481145004}, {"config_name": "spanish_deletions", "features": [{"name": "id", "dtype": "int32"}, {"name": "base_sentence", "dtype": "string"}, {"name": "phrase", "dtype": "string"}, {"name": "edited_sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 317253196, "num_examples": 908276}], "download_size": 78485695, "dataset_size": 317253196}, {"config_name": "french_insertions", "features": [{"name": "id", "dtype": "int32"}, {"name": "base_sentence", "dtype": "string"}, {"name": "phrase", "dtype": "string"}, {"name": "edited_sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 651525210, "num_examples": 2038305}], "download_size": 160442894, "dataset_size": 651525210}, {"config_name": "french_deletions", "features": [{"name": "id", "dtype": "int32"}, {"name": "base_sentence", "dtype": "string"}, {"name": "phrase", "dtype": "string"}, {"name": "edited_sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 626323354, "num_examples": 2060242}], "download_size": 155263358, "dataset_size": 626323354}, {"config_name": "italian_insertions", "features": [{"name": "id", "dtype": "int32"}, {"name": "base_sentence", "dtype": "string"}, {"name": "phrase", "dtype": "string"}, {"name": "edited_sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 372950256, "num_examples": 1078814}], "download_size": 92302006, "dataset_size": 372950256}, {"config_name": "italian_deletions", "features": [{"name": "id", "dtype": "int32"}, {"name": "base_sentence", "dtype": "string"}, {"name": "phrase", "dtype": "string"}, {"name": "edited_sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 198598618, "num_examples": 583316}], "download_size": 49048596, "dataset_size": 198598618}, {"config_name": "japanese_insertions", "features": [{"name": "id", "dtype": "int32"}, {"name": "base_sentence", "dtype": "string"}, {"name": "phrase", "dtype": "string"}, {"name": "edited_sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 765754162, "num_examples": 2249527}], "download_size": 185766012, "dataset_size": 765754162}, {"config_name": "japanese_deletions", "features": [{"name": "id", "dtype": "int32"}, {"name": "base_sentence", "dtype": "string"}, {"name": "phrase", "dtype": "string"}, {"name": "edited_sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 459683880, "num_examples": 1352162}], "download_size": 110513593, "dataset_size": 459683880}, {"config_name": "russian_insertions", "features": [{"name": "id", "dtype": "int32"}, {"name": "base_sentence", "dtype": "string"}, {"name": "phrase", "dtype": "string"}, {"name": "edited_sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 790822192, "num_examples": 1471638}], "download_size": 152985812, "dataset_size": 790822192}, {"config_name": "russian_deletions", "features": [{"name": "id", "dtype": "int32"}, {"name": "base_sentence", "dtype": "string"}, {"name": "phrase", "dtype": "string"}, {"name": "edited_sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 514750186, "num_examples": 960976}], "download_size": 100033230, "dataset_size": 514750186}, {"config_name": "chinese_insertions", "features": [{"name": "id", "dtype": "int32"}, {"name": "base_sentence", "dtype": "string"}, {"name": "phrase", "dtype": "string"}, {"name": "edited_sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 233367646, "num_examples": 746509}], "download_size": 66124094, "dataset_size": 233367646}, {"config_name": "chinese_deletions", "features": [{"name": "id", "dtype": "int32"}, {"name": "base_sentence", "dtype": "string"}, {"name": "phrase", "dtype": "string"}, {"name": "edited_sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 144269112, "num_examples": 467271}], "download_size": 40898651, "dataset_size": 144269112}]}
2024-01-18T11:18:00+00:00
[]
[ "de", "en", "es", "fr", "it", "ja", "ru", "zh" ]
TAGS #task_categories-summarization #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-100K<n<1M #size_categories-10M<n<100M #size_categories-1M<n<10M #source_datasets-original #language-German #language-English #language-Spanish #language-French #language-Italian #language-Japanese #language-Russian #language-Chinese #license-cc-by-sa-4.0 #region-us
# Dataset Card for WikiAtomicEdits ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: None - Repository: URL - Paper: URL - Leaderboard: - Point of Contact: ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages The languages in the dataset are: - de - en - es - fr - it - jp: Japanese ('ja') - ru - zh ## Dataset Structure ### Data Instances Here are some examples of questions and facts: ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @abhishekkrthakur for adding this dataset.
[ "# Dataset Card for WikiAtomicEdits", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: None\n- Repository: URL\n- Paper: URL\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages\n\nThe languages in the dataset are:\n- de\n- en\n- es\n- fr\n- it\n- jp: Japanese ('ja')\n- ru\n- zh", "## Dataset Structure", "### Data Instances\n\nHere are some examples of questions and facts:", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ "TAGS\n#task_categories-summarization #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-100K<n<1M #size_categories-10M<n<100M #size_categories-1M<n<10M #source_datasets-original #language-German #language-English #language-Spanish #language-French #language-Italian #language-Japanese #language-Russian #language-Chinese #license-cc-by-sa-4.0 #region-us \n", "# Dataset Card for WikiAtomicEdits", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: None\n- Repository: URL\n- Paper: URL\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages\n\nThe languages in the dataset are:\n- de\n- en\n- es\n- fr\n- it\n- jp: Japanese ('ja')\n- ru\n- zh", "## Dataset Structure", "### Data Instances\n\nHere are some examples of questions and facts:", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ 137, 10, 120, 28, 6, 10, 38, 6, 17, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 20 ]
[ "passage: TAGS\n#task_categories-summarization #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-100K<n<1M #size_categories-10M<n<100M #size_categories-1M<n<10M #source_datasets-original #language-German #language-English #language-Spanish #language-French #language-Italian #language-Japanese #language-Russian #language-Chinese #license-cc-by-sa-4.0 #region-us \n# Dataset Card for WikiAtomicEdits## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: None\n- Repository: URL\n- Paper: URL\n- Leaderboard: \n- Point of Contact:### Dataset Summary### Supported Tasks and Leaderboards### Languages\n\nThe languages in the dataset are:\n- de\n- en\n- es\n- fr\n- it\n- jp: Japanese ('ja')\n- ru\n- zh## Dataset Structure### Data Instances\n\nHere are some examples of questions and facts:### Data Fields### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators### Licensing Information" ]
1f3a7762819d9afdba523fa7399605a5dd591efa
# Dataset Card for WikiAuto ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Repository:** [WikiAuto github repository](https://github.com/chaojiang06/wiki-auto) - **Paper:** [Neural CRF Model for Sentence Alignment in Text Simplification](https://arxiv.org/abs/2005.02324) - **Point of Contact:** [Chao Jiang](jiang.1530@osu.edu) ### Dataset Summary WikiAuto provides a set of aligned sentences from English Wikipedia and Simple English Wikipedia as a resource to train sentence simplification systems. The authors first crowd-sourced a set of manual alignments between sentences in a subset of the Simple English Wikipedia and their corresponding versions in English Wikipedia (this corresponds to the `manual` config in this version of dataset), then trained a neural CRF system to predict these alignments. The trained alignment prediction model was then applied to the other articles in Simple English Wikipedia with an English counterpart to create a larger corpus of aligned sentences (corresponding to the `auto`, `auto_acl`, `auto_full_no_split`, and `auto_full_with_split` configs here). ### Supported Tasks and Leaderboards The dataset was created to support a `text-simplification` task. Success in these tasks is typically measured using the [SARI](https://huggingface.co/metrics/sari) and [FKBLEU](https://huggingface.co/metrics/fkbleu) metrics described in the paper [Optimizing Statistical Machine Translation for Text Simplification](https://www.aclweb.org/anthology/Q16-1029.pdf). ### Languages While both the input and output of the proposed task are in English (`en`), it should be noted that it is presented as a translation task where Wikipedia Simple English is treated as its own idiom. For a statement of what is intended (but not always observed) to constitute Simple English on this platform, see [Simple English in Wikipedia](https://simple.wikipedia.org/wiki/Wikipedia:About#Simple_English). ## Dataset Structure ### Data Instances The data in all of the configurations looks a little different. A `manual` config instance consists of a sentence from the Simple English Wikipedia article, one from the linked English Wikipedia article, IDs for each of them, and a label indicating whether they are aligned. Sentences on either side can be repeated so that the aligned sentences are in the same instances. For example: ``` {'alignment_label': 1, 'normal_sentence_id': '0_66252-1-0-0', 'simple_sentence_id': '0_66252-0-0-0', 'normal_sentence': 'The Local Government Act 1985 is an Act of Parliament in the United Kingdom.', 'simple_sentence': 'The Local Government Act 1985 was an Act of Parliament in the United Kingdom', 'gleu_score': 0.800000011920929} ``` Is followed by ``` {'alignment_label': 0, 'normal_sentence_id': '0_66252-1-0-1', 'simple_sentence_id': '0_66252-0-0-0', 'normal_sentence': 'Its main effect was to abolish the six county councils of the metropolitan counties that had been set up in 1974, 11 years earlier, by the Local Government Act 1972, along with the Greater London Council that had been established in 1965.', 'simple_sentence': 'The Local Government Act 1985 was an Act of Parliament in the United Kingdom', 'gleu_score': 0.08641975373029709} ``` The `auto` config shows a pair of an English and corresponding Simple English Wikipedia as an instance, with an alignment at the paragraph and sentence level: ``` {'example_id': '0', 'normal': {'normal_article_content': {'normal_sentence': ["Lata Mondal ( ; born: 16 January 1993, Dhaka) is a Bangladeshi cricketer who plays for the Bangladesh national women's cricket team.", 'She is a right handed batter.', 'Mondal was born on January 16, 1993 in Dhaka, Bangladesh.', "Mondal made her ODI career against the Ireland women's cricket team on November 26, 2011.", "Mondal made her T20I career against the Ireland women's cricket team on August 28, 2012.", "In October 2018, she was named in Bangladesh's squad for the 2018 ICC Women's World Twenty20 tournament in the West Indies.", "Mondal was a member of the team that won a silver medal in cricket against the China national women's cricket team at the 2010 Asian Games in Guangzhou, China."], 'normal_sentence_id': ['normal-41918715-0-0', 'normal-41918715-0-1', 'normal-41918715-1-0', 'normal-41918715-2-0', 'normal-41918715-3-0', 'normal-41918715-3-1', 'normal-41918715-4-0']}, 'normal_article_id': 41918715, 'normal_article_title': 'Lata Mondal', 'normal_article_url': 'https://en.wikipedia.org/wiki?curid=41918715'}, 'paragraph_alignment': {'normal_paragraph_id': ['normal-41918715-0'], 'simple_paragraph_id': ['simple-702227-0']}, 'sentence_alignment': {'normal_sentence_id': ['normal-41918715-0-0', 'normal-41918715-0-1'], 'simple_sentence_id': ['simple-702227-0-0', 'simple-702227-0-1']}, 'simple': {'simple_article_content': {'simple_sentence': ["Lata Mondal (born: 16 January 1993) is a Bangladeshi cricketer who plays for the Bangladesh national women's cricket team.", 'She is a right handed bat.'], 'simple_sentence_id': ['simple-702227-0-0', 'simple-702227-0-1']}, 'simple_article_id': 702227, 'simple_article_title': 'Lata Mondal', 'simple_article_url': 'https://simple.wikipedia.org/wiki?curid=702227'}} ``` Finally, the `auto_acl`, the `auto_full_no_split`, and the `auto_full_with_split` configs were obtained by selecting the aligned pairs of sentences from `auto` to provide a ready-to-go aligned dataset to train a sequence-to-sequence system. While `auto_acl` corresponds to the filtered version of the data used to train the systems in the paper, `auto_full_no_split` and `auto_full_with_split` correspond to the unfiltered versions with and without sentence splits respectively. In the `auto_full_with_split` config, we join the sentences in the simple article mapped to the same sentence in the complex article to capture sentence splitting. Split sentences are separated by a `<SEP>` token. In the `auto_full_no_split` config, we do not join the splits and treat them as separate pairs. An instance is a single pair of sentences: ``` {'normal_sentence': 'In early work , Rutherford discovered the concept of radioactive half-life , the radioactive element radon , and differentiated and named alpha and beta radiation .\n', 'simple_sentence': 'Rutherford discovered the radioactive half-life , and the three parts of radiation which he named Alpha , Beta , and Gamma .\n'} ``` ### Data Fields The data has the following field: - `normal_sentence`: a sentence from English Wikipedia. - `normal_sentence_id`: a unique ID for each English Wikipedia sentence. The last two dash-separated numbers correspond to the paragraph number in the article and the sentence number in the paragraph. - `simple_sentence`: a sentence from Simple English Wikipedia. - `simple_sentence_id`: a unique ID for each Simple English Wikipedia sentence. The last two dash-separated numbers correspond to the paragraph number in the article and the sentence number in the paragraph. - `alignment_label`: signifies whether a pair of sentences is aligned: labels are `2:partialAligned`, `1:aligned` and `0:notAligned` - `paragraph_alignment`: a first step of alignment mapping English and Simple English paragraphs from linked articles - `sentence_alignment`: the full alignment mapping English and Simple English sentences from linked articles - `gleu_score`: the sentence level GLEU (Google-BLEU) score for each pair. ### Data Splits In `auto`, the `part_2` split corresponds to the articles used in `manual`, and `part_1` has the rest of Wikipedia. The `manual` config is provided with a `train`/`dev`/`test` split with the following amounts of data: | | train | validation | test | |------------------------|--------:|-----------:|--------:| | Total sentence pairs | 373801 | 73249 | 118074 | | Aligned sentence pairs | 1889 | 346 | 677 | ## Dataset Creation ### Curation Rationale Simple English Wikipedia provides a ready source of training data for text simplification systems, as 1. articles in different languages are linked, making it easier to find parallel data and 2. the Simple English data is written by users for users rather than by professional translators. However, even though articles are aligned, finding a good sentence-level alignment can remain challenging. This work aims to provide a solution for this problem. By manually annotating a sub-set of the articles, they manage to achieve an F1 score of over 88% on predicting alignment, which allows to create a good quality sentence level aligned corpus using all of Simple English Wikipedia. ### Source Data #### Initial Data Collection and Normalization The authors mention that they "extracted 138,095 article pairs from the 2019/09 Wikipedia dump [...] using an improved version of the [WikiExtractor](https://github.com/attardi/wikiextractor) library". The [SpaCy](https://spacy.io/) library is used for sentence splitting. #### Who are the source language producers? The dataset uses langauge from Wikipedia: some demographic information is provided [here](https://en.wikipedia.org/wiki/Wikipedia:Who_writes_Wikipedia%3F). ### Annotations #### Annotation process Sentence alignment labels were obtained for 500 randomly sampled document pairs (10,123 sentence pairs total). The authors pre-selected several alignment candidates from English Wikipedia for each Simple Wikipedia sentence based on various similarity metrics, then asked the crowd-workers to annotate these pairs. #### Who are the annotators? No demographic annotation is provided for the crowd workers. [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators The dataset was created by Chao Jiang, Mounica Maddela, Wuwei Lan, Yang Zhong, and Wei Xu working at Ohio State University. ### Licensing Information The dataset is not licensed by itself, but the source Wikipedia data is under a `cc-by-sa-3.0` license. ### Citation Information You can cite the paper presenting the dataset as: ``` @inproceedings{acl/JiangMLZX20, author = {Chao Jiang and Mounica Maddela and Wuwei Lan and Yang Zhong and Wei Xu}, editor = {Dan Jurafsky and Joyce Chai and Natalie Schluter and Joel R. Tetreault}, title = {Neural {CRF} Model for Sentence Alignment in Text Simplification}, booktitle = {Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, {ACL} 2020, Online, July 5-10, 2020}, pages = {7943--7960}, publisher = {Association for Computational Linguistics}, year = {2020}, url = {https://www.aclweb.org/anthology/2020.acl-main.709/} } ``` ### Contributions Thanks to [@yjernite](https://github.com/yjernite), [@mounicam](https://github.com/mounicam) for adding this dataset.
wiki_auto
[ "task_categories:text2text-generation", "task_ids:text-simplification", "annotations_creators:crowdsourced", "annotations_creators:machine-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:extended|other-wikipedia", "language:en", "license:cc-by-sa-3.0", "arxiv:2005.02324", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced", "machine-generated"], "language_creators": ["found"], "language": ["en"], "license": ["cc-by-sa-3.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["extended|other-wikipedia"], "task_categories": ["text2text-generation"], "task_ids": ["text-simplification"], "pretty_name": "WikiAuto", "config_names": ["auto", "auto_acl", "auto_full_no_split", "auto_full_with_split", "manual"], "dataset_info": [{"config_name": "manual", "features": [{"name": "alignment_label", "dtype": {"class_label": {"names": {"0": "notAligned", "1": "aligned", "2": "partialAligned"}}}}, {"name": "normal_sentence_id", "dtype": "string"}, {"name": "simple_sentence_id", "dtype": "string"}, {"name": "normal_sentence", "dtype": "string"}, {"name": "simple_sentence", "dtype": "string"}, {"name": "gleu_score", "dtype": "float32"}], "splits": [{"name": "train", "num_bytes": 110838475, "num_examples": 373801}, {"name": "dev", "num_bytes": 21112775, "num_examples": 73249}, {"name": "test", "num_bytes": 33851634, "num_examples": 118074}], "download_size": 168957430, "dataset_size": 165802884}, {"config_name": "auto_acl", "features": [{"name": "normal_sentence", "dtype": "string"}, {"name": "simple_sentence", "dtype": "string"}], "splits": [{"name": "full", "num_bytes": 121975414, "num_examples": 488332}], "download_size": 118068366, "dataset_size": 121975414}, {"config_name": "auto", "features": [{"name": "example_id", "dtype": "string"}, {"name": "normal", "struct": [{"name": "normal_article_id", "dtype": "int32"}, {"name": "normal_article_title", "dtype": "string"}, {"name": "normal_article_url", "dtype": "string"}, {"name": "normal_article_content", "sequence": [{"name": "normal_sentence_id", "dtype": "string"}, {"name": "normal_sentence", "dtype": "string"}]}]}, {"name": "simple", "struct": [{"name": "simple_article_id", "dtype": "int32"}, {"name": "simple_article_title", "dtype": "string"}, {"name": "simple_article_url", "dtype": "string"}, {"name": "simple_article_content", "sequence": [{"name": "simple_sentence_id", "dtype": "string"}, {"name": "simple_sentence", "dtype": "string"}]}]}, {"name": "paragraph_alignment", "sequence": [{"name": "normal_paragraph_id", "dtype": "string"}, {"name": "simple_paragraph_id", "dtype": "string"}]}, {"name": "sentence_alignment", "sequence": [{"name": "normal_sentence_id", "dtype": "string"}, {"name": "simple_sentence_id", "dtype": "string"}]}], "splits": [{"name": "part_1", "num_bytes": 1773240295, "num_examples": 125059}, {"name": "part_2", "num_bytes": 80417651, "num_examples": 13036}], "download_size": 2160638921, "dataset_size": 1853657946}, {"config_name": "auto_full_no_split", "features": [{"name": "normal_sentence", "dtype": "string"}, {"name": "simple_sentence", "dtype": "string"}], "splits": [{"name": "full", "num_bytes": 146310611, "num_examples": 591994}], "download_size": 141574179, "dataset_size": 146310611}, {"config_name": "auto_full_with_split", "features": [{"name": "normal_sentence", "dtype": "string"}, {"name": "simple_sentence", "dtype": "string"}], "splits": [{"name": "full", "num_bytes": 124549115, "num_examples": 483801}], "download_size": 120678315, "dataset_size": 124549115}]}
2024-01-18T11:18:01+00:00
[ "2005.02324" ]
[ "en" ]
TAGS #task_categories-text2text-generation #task_ids-text-simplification #annotations_creators-crowdsourced #annotations_creators-machine-generated #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-extended|other-wikipedia #language-English #license-cc-by-sa-3.0 #arxiv-2005.02324 #region-us
Dataset Card for WikiAuto ========================= Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Repository: WikiAuto github repository * Paper: Neural CRF Model for Sentence Alignment in Text Simplification * Point of Contact: Chao Jiang ### Dataset Summary WikiAuto provides a set of aligned sentences from English Wikipedia and Simple English Wikipedia as a resource to train sentence simplification systems. The authors first crowd-sourced a set of manual alignments between sentences in a subset of the Simple English Wikipedia and their corresponding versions in English Wikipedia (this corresponds to the 'manual' config in this version of dataset), then trained a neural CRF system to predict these alignments. The trained alignment prediction model was then applied to the other articles in Simple English Wikipedia with an English counterpart to create a larger corpus of aligned sentences (corresponding to the 'auto', 'auto\_acl', 'auto\_full\_no\_split', and 'auto\_full\_with\_split' configs here). ### Supported Tasks and Leaderboards The dataset was created to support a 'text-simplification' task. Success in these tasks is typically measured using the SARI and FKBLEU metrics described in the paper Optimizing Statistical Machine Translation for Text Simplification. ### Languages While both the input and output of the proposed task are in English ('en'), it should be noted that it is presented as a translation task where Wikipedia Simple English is treated as its own idiom. For a statement of what is intended (but not always observed) to constitute Simple English on this platform, see Simple English in Wikipedia. Dataset Structure ----------------- ### Data Instances The data in all of the configurations looks a little different. A 'manual' config instance consists of a sentence from the Simple English Wikipedia article, one from the linked English Wikipedia article, IDs for each of them, and a label indicating whether they are aligned. Sentences on either side can be repeated so that the aligned sentences are in the same instances. For example: Is followed by The 'auto' config shows a pair of an English and corresponding Simple English Wikipedia as an instance, with an alignment at the paragraph and sentence level: Finally, the 'auto\_acl', the 'auto\_full\_no\_split', and the 'auto\_full\_with\_split' configs were obtained by selecting the aligned pairs of sentences from 'auto' to provide a ready-to-go aligned dataset to train a sequence-to-sequence system. While 'auto\_acl' corresponds to the filtered version of the data used to train the systems in the paper, 'auto\_full\_no\_split' and 'auto\_full\_with\_split' correspond to the unfiltered versions with and without sentence splits respectively. In the 'auto\_full\_with\_split' config, we join the sentences in the simple article mapped to the same sentence in the complex article to capture sentence splitting. Split sentences are separated by a '' token. In the 'auto\_full\_no\_split' config, we do not join the splits and treat them as separate pairs. An instance is a single pair of sentences: ### Data Fields The data has the following field: * 'normal\_sentence': a sentence from English Wikipedia. * 'normal\_sentence\_id': a unique ID for each English Wikipedia sentence. The last two dash-separated numbers correspond to the paragraph number in the article and the sentence number in the paragraph. * 'simple\_sentence': a sentence from Simple English Wikipedia. * 'simple\_sentence\_id': a unique ID for each Simple English Wikipedia sentence. The last two dash-separated numbers correspond to the paragraph number in the article and the sentence number in the paragraph. * 'alignment\_label': signifies whether a pair of sentences is aligned: labels are '2:partialAligned', '1:aligned' and '0:notAligned' * 'paragraph\_alignment': a first step of alignment mapping English and Simple English paragraphs from linked articles * 'sentence\_alignment': the full alignment mapping English and Simple English sentences from linked articles * 'gleu\_score': the sentence level GLEU (Google-BLEU) score for each pair. ### Data Splits In 'auto', the 'part\_2' split corresponds to the articles used in 'manual', and 'part\_1' has the rest of Wikipedia. The 'manual' config is provided with a 'train'/'dev'/'test' split with the following amounts of data: Dataset Creation ---------------- ### Curation Rationale Simple English Wikipedia provides a ready source of training data for text simplification systems, as 1. articles in different languages are linked, making it easier to find parallel data and 2. the Simple English data is written by users for users rather than by professional translators. However, even though articles are aligned, finding a good sentence-level alignment can remain challenging. This work aims to provide a solution for this problem. By manually annotating a sub-set of the articles, they manage to achieve an F1 score of over 88% on predicting alignment, which allows to create a good quality sentence level aligned corpus using all of Simple English Wikipedia. ### Source Data #### Initial Data Collection and Normalization The authors mention that they "extracted 138,095 article pairs from the 2019/09 Wikipedia dump [...] using an improved version of the WikiExtractor library". The SpaCy library is used for sentence splitting. #### Who are the source language producers? The dataset uses langauge from Wikipedia: some demographic information is provided here. ### Annotations #### Annotation process Sentence alignment labels were obtained for 500 randomly sampled document pairs (10,123 sentence pairs total). The authors pre-selected several alignment candidates from English Wikipedia for each Simple Wikipedia sentence based on various similarity metrics, then asked the crowd-workers to annotate these pairs. #### Who are the annotators? No demographic annotation is provided for the crowd workers. ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators The dataset was created by Chao Jiang, Mounica Maddela, Wuwei Lan, Yang Zhong, and Wei Xu working at Ohio State University. ### Licensing Information The dataset is not licensed by itself, but the source Wikipedia data is under a 'cc-by-sa-3.0' license. You can cite the paper presenting the dataset as: ### Contributions Thanks to @yjernite, @mounicam for adding this dataset.
[ "### Dataset Summary\n\n\nWikiAuto provides a set of aligned sentences from English Wikipedia and Simple English Wikipedia as a resource to train sentence simplification systems.\n\n\nThe authors first crowd-sourced a set of manual alignments between sentences in a subset of the Simple English Wikipedia and their corresponding versions in English Wikipedia (this corresponds to the 'manual' config in this version of dataset), then trained a neural CRF system to predict these alignments.\n\n\nThe trained alignment prediction model was then applied to the other articles in Simple English Wikipedia with an English counterpart to create a larger corpus of aligned sentences (corresponding to the 'auto', 'auto\\_acl', 'auto\\_full\\_no\\_split', and 'auto\\_full\\_with\\_split' configs here).", "### Supported Tasks and Leaderboards\n\n\nThe dataset was created to support a 'text-simplification' task. Success in these tasks is typically measured using the SARI and FKBLEU metrics described in the paper Optimizing Statistical Machine Translation for Text Simplification.", "### Languages\n\n\nWhile both the input and output of the proposed task are in English ('en'), it should be noted that it is presented as a translation task where Wikipedia Simple English is treated as its own idiom. For a statement of what is intended (but not always observed) to constitute Simple English on this platform, see Simple English in Wikipedia.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nThe data in all of the configurations looks a little different.\n\n\nA 'manual' config instance consists of a sentence from the Simple English Wikipedia article, one from the linked English Wikipedia article, IDs for each of them, and a label indicating whether they are aligned. Sentences on either side can be repeated so that the aligned sentences are in the same instances. For example:\n\n\nIs followed by\n\n\nThe 'auto' config shows a pair of an English and corresponding Simple English Wikipedia as an instance, with an alignment at the paragraph and sentence level:\n\n\nFinally, the 'auto\\_acl', the 'auto\\_full\\_no\\_split', and the 'auto\\_full\\_with\\_split' configs were obtained by selecting the aligned pairs of sentences from 'auto' to provide a ready-to-go aligned dataset to train a sequence-to-sequence system. While 'auto\\_acl' corresponds to the filtered version of the data used to train the systems in the paper, 'auto\\_full\\_no\\_split' and 'auto\\_full\\_with\\_split' correspond to the unfiltered versions with and without sentence splits respectively. In the 'auto\\_full\\_with\\_split' config, we join the sentences in the simple article mapped to the same sentence in the complex article to capture sentence splitting. Split sentences are separated by a '' token. In the 'auto\\_full\\_no\\_split' config, we do not join the splits and treat them as separate pairs. An instance is a single pair of sentences:", "### Data Fields\n\n\nThe data has the following field:\n\n\n* 'normal\\_sentence': a sentence from English Wikipedia.\n* 'normal\\_sentence\\_id': a unique ID for each English Wikipedia sentence. The last two dash-separated numbers correspond to the paragraph number in the article and the sentence number in the paragraph.\n* 'simple\\_sentence': a sentence from Simple English Wikipedia.\n* 'simple\\_sentence\\_id': a unique ID for each Simple English Wikipedia sentence. The last two dash-separated numbers correspond to the paragraph number in the article and the sentence number in the paragraph.\n* 'alignment\\_label': signifies whether a pair of sentences is aligned: labels are '2:partialAligned', '1:aligned' and '0:notAligned'\n* 'paragraph\\_alignment': a first step of alignment mapping English and Simple English paragraphs from linked articles\n* 'sentence\\_alignment': the full alignment mapping English and Simple English sentences from linked articles\n* 'gleu\\_score': the sentence level GLEU (Google-BLEU) score for each pair.", "### Data Splits\n\n\nIn 'auto', the 'part\\_2' split corresponds to the articles used in 'manual', and 'part\\_1' has the rest of Wikipedia.\n\n\nThe 'manual' config is provided with a 'train'/'dev'/'test' split with the following amounts of data:\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nSimple English Wikipedia provides a ready source of training data for text simplification systems, as 1. articles in different languages are linked, making it easier to find parallel data and 2. the Simple English data is written by users for users rather than by professional translators. However, even though articles are aligned, finding a good sentence-level alignment can remain challenging. This work aims to provide a solution for this problem. By manually annotating a sub-set of the articles, they manage to achieve an F1 score of over 88% on predicting alignment, which allows to create a good quality sentence level aligned corpus using all of Simple English Wikipedia.", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nThe authors mention that they \"extracted 138,095 article pairs from the 2019/09 Wikipedia dump [...] using an improved version of the WikiExtractor library\". The SpaCy library is used for sentence splitting.", "#### Who are the source language producers?\n\n\nThe dataset uses langauge from Wikipedia: some demographic information is provided here.", "### Annotations", "#### Annotation process\n\n\nSentence alignment labels were obtained for 500 randomly sampled document pairs (10,123 sentence pairs total). The authors pre-selected several alignment candidates from English Wikipedia for each Simple Wikipedia sentence based on various similarity metrics, then asked the crowd-workers to annotate these pairs.", "#### Who are the annotators?\n\n\nNo demographic annotation is provided for the crowd workers.", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThe dataset was created by Chao Jiang, Mounica Maddela, Wuwei Lan, Yang Zhong, and Wei Xu working at Ohio State University.", "### Licensing Information\n\n\nThe dataset is not licensed by itself, but the source Wikipedia data is under a 'cc-by-sa-3.0' license.\n\n\nYou can cite the paper presenting the dataset as:", "### Contributions\n\n\nThanks to @yjernite, @mounicam for adding this dataset." ]
[ "TAGS\n#task_categories-text2text-generation #task_ids-text-simplification #annotations_creators-crowdsourced #annotations_creators-machine-generated #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-extended|other-wikipedia #language-English #license-cc-by-sa-3.0 #arxiv-2005.02324 #region-us \n", "### Dataset Summary\n\n\nWikiAuto provides a set of aligned sentences from English Wikipedia and Simple English Wikipedia as a resource to train sentence simplification systems.\n\n\nThe authors first crowd-sourced a set of manual alignments between sentences in a subset of the Simple English Wikipedia and their corresponding versions in English Wikipedia (this corresponds to the 'manual' config in this version of dataset), then trained a neural CRF system to predict these alignments.\n\n\nThe trained alignment prediction model was then applied to the other articles in Simple English Wikipedia with an English counterpart to create a larger corpus of aligned sentences (corresponding to the 'auto', 'auto\\_acl', 'auto\\_full\\_no\\_split', and 'auto\\_full\\_with\\_split' configs here).", "### Supported Tasks and Leaderboards\n\n\nThe dataset was created to support a 'text-simplification' task. Success in these tasks is typically measured using the SARI and FKBLEU metrics described in the paper Optimizing Statistical Machine Translation for Text Simplification.", "### Languages\n\n\nWhile both the input and output of the proposed task are in English ('en'), it should be noted that it is presented as a translation task where Wikipedia Simple English is treated as its own idiom. For a statement of what is intended (but not always observed) to constitute Simple English on this platform, see Simple English in Wikipedia.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nThe data in all of the configurations looks a little different.\n\n\nA 'manual' config instance consists of a sentence from the Simple English Wikipedia article, one from the linked English Wikipedia article, IDs for each of them, and a label indicating whether they are aligned. Sentences on either side can be repeated so that the aligned sentences are in the same instances. For example:\n\n\nIs followed by\n\n\nThe 'auto' config shows a pair of an English and corresponding Simple English Wikipedia as an instance, with an alignment at the paragraph and sentence level:\n\n\nFinally, the 'auto\\_acl', the 'auto\\_full\\_no\\_split', and the 'auto\\_full\\_with\\_split' configs were obtained by selecting the aligned pairs of sentences from 'auto' to provide a ready-to-go aligned dataset to train a sequence-to-sequence system. While 'auto\\_acl' corresponds to the filtered version of the data used to train the systems in the paper, 'auto\\_full\\_no\\_split' and 'auto\\_full\\_with\\_split' correspond to the unfiltered versions with and without sentence splits respectively. In the 'auto\\_full\\_with\\_split' config, we join the sentences in the simple article mapped to the same sentence in the complex article to capture sentence splitting. Split sentences are separated by a '' token. In the 'auto\\_full\\_no\\_split' config, we do not join the splits and treat them as separate pairs. An instance is a single pair of sentences:", "### Data Fields\n\n\nThe data has the following field:\n\n\n* 'normal\\_sentence': a sentence from English Wikipedia.\n* 'normal\\_sentence\\_id': a unique ID for each English Wikipedia sentence. The last two dash-separated numbers correspond to the paragraph number in the article and the sentence number in the paragraph.\n* 'simple\\_sentence': a sentence from Simple English Wikipedia.\n* 'simple\\_sentence\\_id': a unique ID for each Simple English Wikipedia sentence. The last two dash-separated numbers correspond to the paragraph number in the article and the sentence number in the paragraph.\n* 'alignment\\_label': signifies whether a pair of sentences is aligned: labels are '2:partialAligned', '1:aligned' and '0:notAligned'\n* 'paragraph\\_alignment': a first step of alignment mapping English and Simple English paragraphs from linked articles\n* 'sentence\\_alignment': the full alignment mapping English and Simple English sentences from linked articles\n* 'gleu\\_score': the sentence level GLEU (Google-BLEU) score for each pair.", "### Data Splits\n\n\nIn 'auto', the 'part\\_2' split corresponds to the articles used in 'manual', and 'part\\_1' has the rest of Wikipedia.\n\n\nThe 'manual' config is provided with a 'train'/'dev'/'test' split with the following amounts of data:\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nSimple English Wikipedia provides a ready source of training data for text simplification systems, as 1. articles in different languages are linked, making it easier to find parallel data and 2. the Simple English data is written by users for users rather than by professional translators. However, even though articles are aligned, finding a good sentence-level alignment can remain challenging. This work aims to provide a solution for this problem. By manually annotating a sub-set of the articles, they manage to achieve an F1 score of over 88% on predicting alignment, which allows to create a good quality sentence level aligned corpus using all of Simple English Wikipedia.", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nThe authors mention that they \"extracted 138,095 article pairs from the 2019/09 Wikipedia dump [...] using an improved version of the WikiExtractor library\". The SpaCy library is used for sentence splitting.", "#### Who are the source language producers?\n\n\nThe dataset uses langauge from Wikipedia: some demographic information is provided here.", "### Annotations", "#### Annotation process\n\n\nSentence alignment labels were obtained for 500 randomly sampled document pairs (10,123 sentence pairs total). The authors pre-selected several alignment candidates from English Wikipedia for each Simple Wikipedia sentence based on various similarity metrics, then asked the crowd-workers to annotate these pairs.", "#### Who are the annotators?\n\n\nNo demographic annotation is provided for the crowd workers.", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThe dataset was created by Chao Jiang, Mounica Maddela, Wuwei Lan, Yang Zhong, and Wei Xu working at Ohio State University.", "### Licensing Information\n\n\nThe dataset is not licensed by itself, but the source Wikipedia data is under a 'cc-by-sa-3.0' license.\n\n\nYou can cite the paper presenting the dataset as:", "### Contributions\n\n\nThanks to @yjernite, @mounicam for adding this dataset." ]
[ 122, 190, 62, 85, 386, 268, 80, 146, 4, 60, 28, 5, 75, 21, 18, 7, 8, 14, 38, 47, 22 ]
[ "passage: TAGS\n#task_categories-text2text-generation #task_ids-text-simplification #annotations_creators-crowdsourced #annotations_creators-machine-generated #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-extended|other-wikipedia #language-English #license-cc-by-sa-3.0 #arxiv-2005.02324 #region-us \n### Dataset Summary\n\n\nWikiAuto provides a set of aligned sentences from English Wikipedia and Simple English Wikipedia as a resource to train sentence simplification systems.\n\n\nThe authors first crowd-sourced a set of manual alignments between sentences in a subset of the Simple English Wikipedia and their corresponding versions in English Wikipedia (this corresponds to the 'manual' config in this version of dataset), then trained a neural CRF system to predict these alignments.\n\n\nThe trained alignment prediction model was then applied to the other articles in Simple English Wikipedia with an English counterpart to create a larger corpus of aligned sentences (corresponding to the 'auto', 'auto\\_acl', 'auto\\_full\\_no\\_split', and 'auto\\_full\\_with\\_split' configs here).### Supported Tasks and Leaderboards\n\n\nThe dataset was created to support a 'text-simplification' task. Success in these tasks is typically measured using the SARI and FKBLEU metrics described in the paper Optimizing Statistical Machine Translation for Text Simplification.### Languages\n\n\nWhile both the input and output of the proposed task are in English ('en'), it should be noted that it is presented as a translation task where Wikipedia Simple English is treated as its own idiom. For a statement of what is intended (but not always observed) to constitute Simple English on this platform, see Simple English in Wikipedia.\n\n\nDataset Structure\n-----------------", "passage: ### Data Instances\n\n\nThe data in all of the configurations looks a little different.\n\n\nA 'manual' config instance consists of a sentence from the Simple English Wikipedia article, one from the linked English Wikipedia article, IDs for each of them, and a label indicating whether they are aligned. Sentences on either side can be repeated so that the aligned sentences are in the same instances. For example:\n\n\nIs followed by\n\n\nThe 'auto' config shows a pair of an English and corresponding Simple English Wikipedia as an instance, with an alignment at the paragraph and sentence level:\n\n\nFinally, the 'auto\\_acl', the 'auto\\_full\\_no\\_split', and the 'auto\\_full\\_with\\_split' configs were obtained by selecting the aligned pairs of sentences from 'auto' to provide a ready-to-go aligned dataset to train a sequence-to-sequence system. While 'auto\\_acl' corresponds to the filtered version of the data used to train the systems in the paper, 'auto\\_full\\_no\\_split' and 'auto\\_full\\_with\\_split' correspond to the unfiltered versions with and without sentence splits respectively. In the 'auto\\_full\\_with\\_split' config, we join the sentences in the simple article mapped to the same sentence in the complex article to capture sentence splitting. Split sentences are separated by a '' token. In the 'auto\\_full\\_no\\_split' config, we do not join the splits and treat them as separate pairs. An instance is a single pair of sentences:### Data Fields\n\n\nThe data has the following field:\n\n\n* 'normal\\_sentence': a sentence from English Wikipedia.\n* 'normal\\_sentence\\_id': a unique ID for each English Wikipedia sentence. The last two dash-separated numbers correspond to the paragraph number in the article and the sentence number in the paragraph.\n* 'simple\\_sentence': a sentence from Simple English Wikipedia.\n* 'simple\\_sentence\\_id': a unique ID for each Simple English Wikipedia sentence. The last two dash-separated numbers correspond to the paragraph number in the article and the sentence number in the paragraph.\n* 'alignment\\_label': signifies whether a pair of sentences is aligned: labels are '2:partialAligned', '1:aligned' and '0:notAligned'\n* 'paragraph\\_alignment': a first step of alignment mapping English and Simple English paragraphs from linked articles\n* 'sentence\\_alignment': the full alignment mapping English and Simple English sentences from linked articles\n* 'gleu\\_score': the sentence level GLEU (Google-BLEU) score for each pair.### Data Splits\n\n\nIn 'auto', the 'part\\_2' split corresponds to the articles used in 'manual', and 'part\\_1' has the rest of Wikipedia.\n\n\nThe 'manual' config is provided with a 'train'/'dev'/'test' split with the following amounts of data:\n\n\n\nDataset Creation\n----------------### Curation Rationale\n\n\nSimple English Wikipedia provides a ready source of training data for text simplification systems, as 1. articles in different languages are linked, making it easier to find parallel data and 2. the Simple English data is written by users for users rather than by professional translators. However, even though articles are aligned, finding a good sentence-level alignment can remain challenging. This work aims to provide a solution for this problem. By manually annotating a sub-set of the articles, they manage to achieve an F1 score of over 88% on predicting alignment, which allows to create a good quality sentence level aligned corpus using all of Simple English Wikipedia.### Source Data" ]
3a00c01f03c79c4799a4b884bd9508a1eea47b6a
# Dataset Card for [Dataset Name] ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Repository:** https://github.com/DavidGrangier/wikipedia-biography-dataset - **Paper:** https://arxiv.org/pdf/1603.07771.pdf - **GitHub:** https://github.com/DavidGrangier/wikipedia-biography-dataset ### Dataset Summary This Dataset contains 728321 biographies extracted from Wikipedia containing the first paragraph of the biography and the tabular infobox. ### Supported Tasks and Leaderboards The main purpose of this dataset is developing text generation models. ### Languages English. ## Dataset Structure ### Data Instances More Information Needed ### Data Fields The structure of a single sample is the following: ```json { "input_text":{ "context":"pope michael iii of alexandria\n", "table":{ "column_header":[ "type", "ended", "death_date", "title", "enthroned", "name", "buried", "religion", "predecessor", "nationality", "article_title", "feast_day", "birth_place", "residence", "successor" ], "content":[ "pope", "16 march 907", "16 march 907", "56th of st. mark pope of alexandria & patriarch of the see", "25 april 880", "michael iii of alexandria", "monastery of saint macarius the great", "coptic orthodox christian", "shenouda i", "egyptian", "pope michael iii of alexandria\n", "16 -rrb- march -lrb- 20 baramhat in the coptic calendar", "egypt", "saint mark 's church", "gabriel i" ], "row_number":[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1] } }, "target_text":"pope michael iii of alexandria -lrb- also known as khail iii -rrb- was the coptic pope of alexandria and patriarch of the see of st. mark -lrb- 880 -- 907 -rrb- .\nin 882 , the governor of egypt , ahmad ibn tulun , forced khail to pay heavy contributions , forcing him to sell a church and some attached properties to the local jewish community .\nthis building was at one time believed to have later become the site of the cairo geniza .\n" } ``` where, in the `"table"` field, all the information of the Wikpedia infobox is stored (the header of the infobox is stored in `"column_header"` and the information in the `"content"` field). ### Data Splits - Train: 582659 samples. - Test: 72831 samples. - Validation: 72831 samples. ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data This dataset was announced in the paper <em>Neural Text Generation from Structured Data with Application to the Biography Domain</em> [(arxiv link)](https://arxiv.org/pdf/1603.07771.pdf) and is stored in [this](https://github.com/DavidGrangier/wikipedia-biography-dataset) repo (owned by DavidGrangier). #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information This dataset is ditributed under Creative Comons CC BY-SA 3.0 License. ### Citation Information For refering the original paper in BibTex format: ``` @article{DBLP:journals/corr/LebretGA16, author = {R{\'{e}}mi Lebret and David Grangier and Michael Auli}, title = {Generating Text from Structured Data with Application to the Biography Domain}, journal = {CoRR}, volume = {abs/1603.07771}, year = {2016}, url = {http://arxiv.org/abs/1603.07771}, archivePrefix = {arXiv}, eprint = {1603.07771}, timestamp = {Mon, 13 Aug 2018 16:48:30 +0200}, biburl = {https://dblp.org/rec/journals/corr/LebretGA16.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ``` ### Contributions Thanks to [@alejandrocros](https://github.com/alejandrocros) for adding this dataset.
wiki_bio
[ "task_categories:table-to-text", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:en", "license:cc-by-sa-3.0", "arxiv:1603.07771", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["en"], "license": ["cc-by-sa-3.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["table-to-text"], "task_ids": [], "paperswithcode_id": "wikibio", "pretty_name": "WikiBio", "dataset_info": {"features": [{"name": "input_text", "struct": [{"name": "table", "sequence": [{"name": "column_header", "dtype": "string"}, {"name": "row_number", "dtype": "int16"}, {"name": "content", "dtype": "string"}]}, {"name": "context", "dtype": "string"}]}, {"name": "target_text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 619269257, "num_examples": 582659}, {"name": "test", "num_bytes": 77264695, "num_examples": 72831}, {"name": "val", "num_bytes": 77335069, "num_examples": 72831}], "download_size": 333998704, "dataset_size": 773869021}}
2024-01-18T11:18:02+00:00
[ "1603.07771" ]
[ "en" ]
TAGS #task_categories-table-to-text #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-sa-3.0 #arxiv-1603.07771 #region-us
# Dataset Card for [Dataset Name] ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Repository: URL - Paper: URL - GitHub: URL ### Dataset Summary This Dataset contains 728321 biographies extracted from Wikipedia containing the first paragraph of the biography and the tabular infobox. ### Supported Tasks and Leaderboards The main purpose of this dataset is developing text generation models. ### Languages English. ## Dataset Structure ### Data Instances ### Data Fields The structure of a single sample is the following: where, in the '"table"' field, all the information of the Wikpedia infobox is stored (the header of the infobox is stored in '"column_header"' and the information in the '"content"' field). ### Data Splits - Train: 582659 samples. - Test: 72831 samples. - Validation: 72831 samples. ## Dataset Creation ### Curation Rationale ### Source Data This dataset was announced in the paper <em>Neural Text Generation from Structured Data with Application to the Biography Domain</em> (arxiv link) and is stored in this repo (owned by DavidGrangier). #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information This dataset is ditributed under Creative Comons CC BY-SA 3.0 License. For refering the original paper in BibTex format: ### Contributions Thanks to @alejandrocros for adding this dataset.
[ "# Dataset Card for [Dataset Name]", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Repository: URL\n- Paper: URL\n- GitHub: URL", "### Dataset Summary\n\nThis Dataset contains 728321 biographies extracted from Wikipedia containing the first paragraph of the biography and the tabular infobox.", "### Supported Tasks and Leaderboards\n\nThe main purpose of this dataset is developing text generation models.", "### Languages\n\nEnglish.", "## Dataset Structure", "### Data Instances", "### Data Fields\n\nThe structure of a single sample is the following:\n\nwhere, in the '\"table\"' field, all the information of the Wikpedia infobox is stored (the header of the infobox is stored in '\"column_header\"' and the information in the '\"content\"' field).", "### Data Splits\n\n- Train: 582659 samples.\n- Test: 72831 samples.\n- Validation: 72831 samples.", "## Dataset Creation", "### Curation Rationale", "### Source Data\nThis dataset was announced in the paper <em>Neural Text Generation from Structured Data with Application to the Biography Domain</em> (arxiv link) and is stored in this repo (owned by DavidGrangier).", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nThis dataset is ditributed under Creative Comons CC BY-SA 3.0 License.\n\n\nFor refering the original paper in BibTex format:", "### Contributions\n\nThanks to @alejandrocros for adding this dataset." ]
[ "TAGS\n#task_categories-table-to-text #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-sa-3.0 #arxiv-1603.07771 #region-us \n", "# Dataset Card for [Dataset Name]", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Repository: URL\n- Paper: URL\n- GitHub: URL", "### Dataset Summary\n\nThis Dataset contains 728321 biographies extracted from Wikipedia containing the first paragraph of the biography and the tabular infobox.", "### Supported Tasks and Leaderboards\n\nThe main purpose of this dataset is developing text generation models.", "### Languages\n\nEnglish.", "## Dataset Structure", "### Data Instances", "### Data Fields\n\nThe structure of a single sample is the following:\n\nwhere, in the '\"table\"' field, all the information of the Wikpedia infobox is stored (the header of the infobox is stored in '\"column_header\"' and the information in the '\"content\"' field).", "### Data Splits\n\n- Train: 582659 samples.\n- Test: 72831 samples.\n- Validation: 72831 samples.", "## Dataset Creation", "### Curation Rationale", "### Source Data\nThis dataset was announced in the paper <em>Neural Text Generation from Structured Data with Application to the Biography Domain</em> (arxiv link) and is stored in this repo (owned by DavidGrangier).", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nThis dataset is ditributed under Creative Comons CC BY-SA 3.0 License.\n\n\nFor refering the original paper in BibTex format:", "### Contributions\n\nThanks to @alejandrocros for adding this dataset." ]
[ 87, 10, 120, 20, 37, 23, 6, 6, 6, 73, 34, 5, 7, 55, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 37, 19 ]
[ "passage: TAGS\n#task_categories-table-to-text #annotations_creators-found #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-sa-3.0 #arxiv-1603.07771 #region-us \n# Dataset Card for [Dataset Name]## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Repository: URL\n- Paper: URL\n- GitHub: URL### Dataset Summary\n\nThis Dataset contains 728321 biographies extracted from Wikipedia containing the first paragraph of the biography and the tabular infobox.### Supported Tasks and Leaderboards\n\nThe main purpose of this dataset is developing text generation models.### Languages\n\nEnglish.## Dataset Structure### Data Instances### Data Fields\n\nThe structure of a single sample is the following:\n\nwhere, in the '\"table\"' field, all the information of the Wikpedia infobox is stored (the header of the infobox is stored in '\"column_header\"' and the information in the '\"content\"' field).### Data Splits\n\n- Train: 582659 samples.\n- Test: 72831 samples.\n- Validation: 72831 samples.## Dataset Creation### Curation Rationale### Source Data\nThis dataset was announced in the paper <em>Neural Text Generation from Structured Data with Application to the Biography Domain</em> (arxiv link) and is stored in this repo (owned by DavidGrangier).#### Initial Data Collection and Normalization" ]
b24a417d802a583f8922946c1c75210290e93108
# Dataset Card for "wiki_dpr" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Repository:** https://github.com/facebookresearch/DPR - **Paper:** https://arxiv.org/abs/2004.04906 - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 425.79 GB - **Size of the generated dataset:** 470.52 GB - **Total amount of disk used:** 978.05 GB ### Dataset Summary This is the wikipedia split used to evaluate the Dense Passage Retrieval (DPR) model. It contains 21M passages from wikipedia along with their DPR embeddings. The wikipedia articles were split into multiple, disjoint text blocks of 100 words as passages. The wikipedia dump is the one from Dec. 20, 2018. There are two types of DPR embeddings based on two different models: - `nq`: the model is trained on the Natural Questions dataset - `multiset`: the model is trained on multiple datasets Additionally, a FAISS index can be created from the embeddings: - `exact`: with an exact FAISS index (high RAM usage) - `compressed`: with a compressed FAISS index (approximate, but lower RAM usage) - `no_index`: without FAISS index Finally, there is the possibility of generating the dataset without the embeddings: - `no_embeddings` ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances Each instance contains a paragraph of at most 100 words, as well as the title of the wikipedia page it comes from, and the DPR embedding (a 768-d vector). #### psgs_w100.multiset.compressed - **Size of downloaded dataset files:** 70.97 GB - **Size of the generated dataset:** 78.42 GB - **Total amount of disk used:** 152.26 GB An example of 'train' looks as follows. ``` This example was too long and was cropped: {'id': '1', 'text': 'Aaron Aaron ( or ; "Ahärôn") is a prophet, high priest, and the brother of Moses in the Abrahamic religions. Knowledge of Aaron, along with his brother Moses, comes exclusively from religious texts, such as the Bible and Quran. The Hebrew Bible relates that, unlike Moses, who grew up in the Egyptian royal court, Aaron and his elder sister Miriam remained with their kinsmen in the eastern border-land of Egypt (Goshen). When Moses first confronted the Egyptian king about the Israelites, Aaron served as his brother\'s spokesman ("prophet") to the Pharaoh. Part of the Law (Torah) that Moses received from'], 'title': 'Aaron', 'embeddings': [-0.07233893871307373, 0.48035329580307007, 0.18650995194911957, -0.5287084579467773, -0.37329429388046265, 0.37622880935668945, 0.25524479150772095, ... -0.336689829826355, 0.6313082575798035, -0.7025573253631592]} ``` #### psgs_w100.multiset.exact - **Size of downloaded dataset files:** 70.97 GB - **Size of the generated dataset:** 78.42 GB - **Total amount of disk used:** 187.38 GB An example of 'train' looks as follows. ``` This example was too long and was cropped: {'id': '1', 'text': 'Aaron Aaron ( or ; "Ahärôn") is a prophet, high priest, and the brother of Moses in the Abrahamic religions. Knowledge of Aaron, along with his brother Moses, comes exclusively from religious texts, such as the Bible and Quran. The Hebrew Bible relates that, unlike Moses, who grew up in the Egyptian royal court, Aaron and his elder sister Miriam remained with their kinsmen in the eastern border-land of Egypt (Goshen). When Moses first confronted the Egyptian king about the Israelites, Aaron served as his brother\'s spokesman ("prophet") to the Pharaoh. Part of the Law (Torah) that Moses received from'], 'title': 'Aaron', 'embeddings': [-0.07233893871307373, 0.48035329580307007, 0.18650995194911957, -0.5287084579467773, -0.37329429388046265, 0.37622880935668945, 0.25524479150772095, ... -0.336689829826355, 0.6313082575798035, -0.7025573253631592]} ``` #### psgs_w100.multiset.no_index - **Size of downloaded dataset files:** 70.97 GB - **Size of the generated dataset:** 78.42 GB - **Total amount of disk used:** 149.38 GB An example of 'train' looks as follows. ``` This example was too long and was cropped: {'id': '1', 'text': 'Aaron Aaron ( or ; "Ahärôn") is a prophet, high priest, and the brother of Moses in the Abrahamic religions. Knowledge of Aaron, along with his brother Moses, comes exclusively from religious texts, such as the Bible and Quran. The Hebrew Bible relates that, unlike Moses, who grew up in the Egyptian royal court, Aaron and his elder sister Miriam remained with their kinsmen in the eastern border-land of Egypt (Goshen). When Moses first confronted the Egyptian king about the Israelites, Aaron served as his brother\'s spokesman ("prophet") to the Pharaoh. Part of the Law (Torah) that Moses received from'], 'title': 'Aaron', 'embeddings': [-0.07233893871307373, 0.48035329580307007, 0.18650995194911957, -0.5287084579467773, -0.37329429388046265, 0.37622880935668945, 0.25524479150772095, ... -0.336689829826355, 0.6313082575798035, -0.7025573253631592]} ``` #### psgs_w100.nq.compressed - **Size of downloaded dataset files:** 70.97 GB - **Size of the generated dataset:** 78.42 GB - **Total amount of disk used:** 152.26 GB An example of 'train' looks as follows. ``` This example was too long and was cropped: {'id': '1', 'text': 'Aaron Aaron ( or ; "Ahärôn") is a prophet, high priest, and the brother of Moses in the Abrahamic religions. Knowledge of Aaron, along with his brother Moses, comes exclusively from religious texts, such as the Bible and Quran. The Hebrew Bible relates that, unlike Moses, who grew up in the Egyptian royal court, Aaron and his elder sister Miriam remained with their kinsmen in the eastern border-land of Egypt (Goshen). When Moses first confronted the Egyptian king about the Israelites, Aaron served as his brother\'s spokesman ("prophet") to the Pharaoh. Part of the Law (Torah) that Moses received from'], 'title': 'Aaron', 'embeddings': [0.013342111371457577, 0.582173764705658, -0.31309744715690613, -0.6991612911224365, -0.5583199858665466, 0.5187504887580872, 0.7152731418609619, ... -0.5385938286781311, 0.8093984127044678, -0.4741983711719513]} ``` #### psgs_w100.nq.exact - **Size of downloaded dataset files:** 70.97 GB - **Size of the generated dataset:** 78.42 GB - **Total amount of disk used:** 187.38 GB An example of 'train' looks as follows. ``` This example was too long and was cropped: {'id': '1', 'text': 'Aaron Aaron ( or ; "Ahärôn") is a prophet, high priest, and the brother of Moses in the Abrahamic religions. Knowledge of Aaron, along with his brother Moses, comes exclusively from religious texts, such as the Bible and Quran. The Hebrew Bible relates that, unlike Moses, who grew up in the Egyptian royal court, Aaron and his elder sister Miriam remained with their kinsmen in the eastern border-land of Egypt (Goshen). When Moses first confronted the Egyptian king about the Israelites, Aaron served as his brother\'s spokesman ("prophet") to the Pharaoh. Part of the Law (Torah) that Moses received from'], 'title': 'Aaron', 'embeddings': [0.013342111371457577, 0.582173764705658, -0.31309744715690613, -0.6991612911224365, -0.5583199858665466, 0.5187504887580872, 0.7152731418609619, ... -0.5385938286781311, 0.8093984127044678, -0.4741983711719513]} ``` ### Data Fields The data fields are the same among all splits. #### psgs_w100.multiset.compressed - `id`: a `string` feature. - `text`: a `string` feature. - `title`: a `string` feature. - `embeddings`: a `list` of `float32` features. #### psgs_w100.multiset.exact - `id`: a `string` feature. - `text`: a `string` feature. - `title`: a `string` feature. - `embeddings`: a `list` of `float32` features. #### psgs_w100.multiset.no_index - `id`: a `string` feature. - `text`: a `string` feature. - `title`: a `string` feature. - `embeddings`: a `list` of `float32` features. #### psgs_w100.nq.compressed - `id`: a `string` feature. - `text`: a `string` feature. - `title`: a `string` feature. - `embeddings`: a `list` of `float32` features. #### psgs_w100.nq.exact - `id`: a `string` feature. - `text`: a `string` feature. - `title`: a `string` feature. - `embeddings`: a `list` of `float32` features. ### Data Splits | name | train | |-----------------------------|-------:| |psgs_w100.multiset.compressed|21015300| |psgs_w100.multiset.exact |21015300| |psgs_w100.multiset.no_index |21015300| |psgs_w100.nq.compressed |21015300| |psgs_w100.nq.exact |21015300| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information DPR is CC-BY-NC 4.0 licensed: https://github.com/facebookresearch/DPR/blob/main/LICENSE ### Citation Information ``` @inproceedings{karpukhin-etal-2020-dense, title = "Dense Passage Retrieval for Open-Domain Question Answering", author = "Karpukhin, Vladimir and Oguz, Barlas and Min, Sewon and Lewis, Patrick and Wu, Ledell and Edunov, Sergey and Chen, Danqi and Yih, Wen-tau", booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)", month = nov, year = "2020", address = "Online", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/2020.emnlp-main.550", doi = "10.18653/v1/2020.emnlp-main.550", pages = "6769--6781", } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@lewtun](https://github.com/lewtun), [@lhoestq](https://github.com/lhoestq) for adding this dataset.
wiki_dpr
[ "task_categories:fill-mask", "task_categories:text-generation", "task_ids:language-modeling", "task_ids:masked-language-modeling", "annotations_creators:no-annotation", "language_creators:crowdsourced", "multilinguality:multilingual", "size_categories:10M<n<100M", "source_datasets:original", "language:en", "license:cc-by-nc-4.0", "text-search", "arxiv:2004.04906", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["crowdsourced"], "language": ["en"], "license": "cc-by-nc-4.0", "multilinguality": ["multilingual"], "size_categories": ["10M<n<100M"], "source_datasets": ["original"], "task_categories": ["fill-mask", "text-generation"], "task_ids": ["language-modeling", "masked-language-modeling"], "pretty_name": "Wiki-DPR", "tags": ["text-search"], "dataset_info": [{"config_name": "psgs_w100.nq.exact", "features": [{"name": "id", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "embeddings", "sequence": "float32"}], "splits": [{"name": "train", "num_bytes": 78419281788, "num_examples": 21015300}], "download_size": 70965697456, "dataset_size": 78419281788}, {"config_name": "psgs_w100.nq.compressed", "features": [{"name": "id", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "embeddings", "sequence": "float32"}], "splits": [{"name": "train", "num_bytes": 78419281788, "num_examples": 21015300}], "download_size": 70965697456, "dataset_size": 78419281788}, {"config_name": "psgs_w100.nq.no_index", "features": [{"name": "id", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "embeddings", "sequence": "float32"}], "splits": [{"name": "train", "num_bytes": 78419281788, "num_examples": 21015300}], "download_size": 70965697456, "dataset_size": 78419281788}, {"config_name": "psgs_w100.multiset.exact", "features": [{"name": "id", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "embeddings", "sequence": "float32"}], "splits": [{"name": "train", "num_bytes": 78419281788, "num_examples": 21015300}], "download_size": 70965697456, "dataset_size": 78419281788}, {"config_name": "psgs_w100.multiset.compressed", "features": [{"name": "id", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "embeddings", "sequence": "float32"}], "splits": [{"name": "train", "num_bytes": 78419281788, "num_examples": 21015300}], "download_size": 70965697456, "dataset_size": 78419281788}, {"config_name": "psgs_w100.multiset.no_index", "features": [{"name": "id", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "embeddings", "sequence": "float32"}], "splits": [{"name": "train", "num_bytes": 78419281788, "num_examples": 21015300}], "download_size": 70965697456, "dataset_size": 78419281788}]}
2024-01-30T15:16:56+00:00
[ "2004.04906" ]
[ "en" ]
TAGS #task_categories-fill-mask #task_categories-text-generation #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-crowdsourced #multilinguality-multilingual #size_categories-10M<n<100M #source_datasets-original #language-English #license-cc-by-nc-4.0 #text-search #arxiv-2004.04906 #region-us
Dataset Card for "wiki\_dpr" ============================ Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Repository: URL * Paper: URL * Point of Contact: * Size of downloaded dataset files: 425.79 GB * Size of the generated dataset: 470.52 GB * Total amount of disk used: 978.05 GB ### Dataset Summary This is the wikipedia split used to evaluate the Dense Passage Retrieval (DPR) model. It contains 21M passages from wikipedia along with their DPR embeddings. The wikipedia articles were split into multiple, disjoint text blocks of 100 words as passages. The wikipedia dump is the one from Dec. 20, 2018. There are two types of DPR embeddings based on two different models: * 'nq': the model is trained on the Natural Questions dataset * 'multiset': the model is trained on multiple datasets Additionally, a FAISS index can be created from the embeddings: * 'exact': with an exact FAISS index (high RAM usage) * 'compressed': with a compressed FAISS index (approximate, but lower RAM usage) * 'no\_index': without FAISS index Finally, there is the possibility of generating the dataset without the embeddings: * 'no\_embeddings' ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- ### Data Instances Each instance contains a paragraph of at most 100 words, as well as the title of the wikipedia page it comes from, and the DPR embedding (a 768-d vector). #### psgs\_w100.multiset.compressed * Size of downloaded dataset files: 70.97 GB * Size of the generated dataset: 78.42 GB * Total amount of disk used: 152.26 GB An example of 'train' looks as follows. #### psgs\_w100.URL * Size of downloaded dataset files: 70.97 GB * Size of the generated dataset: 78.42 GB * Total amount of disk used: 187.38 GB An example of 'train' looks as follows. #### psgs\_w100.multiset.no\_index * Size of downloaded dataset files: 70.97 GB * Size of the generated dataset: 78.42 GB * Total amount of disk used: 149.38 GB An example of 'train' looks as follows. #### psgs\_w100.nq.compressed * Size of downloaded dataset files: 70.97 GB * Size of the generated dataset: 78.42 GB * Total amount of disk used: 152.26 GB An example of 'train' looks as follows. #### psgs\_w100.URL * Size of downloaded dataset files: 70.97 GB * Size of the generated dataset: 78.42 GB * Total amount of disk used: 187.38 GB An example of 'train' looks as follows. ### Data Fields The data fields are the same among all splits. #### psgs\_w100.multiset.compressed * 'id': a 'string' feature. * 'text': a 'string' feature. * 'title': a 'string' feature. * 'embeddings': a 'list' of 'float32' features. #### psgs\_w100.URL * 'id': a 'string' feature. * 'text': a 'string' feature. * 'title': a 'string' feature. * 'embeddings': a 'list' of 'float32' features. #### psgs\_w100.multiset.no\_index * 'id': a 'string' feature. * 'text': a 'string' feature. * 'title': a 'string' feature. * 'embeddings': a 'list' of 'float32' features. #### psgs\_w100.nq.compressed * 'id': a 'string' feature. * 'text': a 'string' feature. * 'title': a 'string' feature. * 'embeddings': a 'list' of 'float32' features. #### psgs\_w100.URL * 'id': a 'string' feature. * 'text': a 'string' feature. * 'title': a 'string' feature. * 'embeddings': a 'list' of 'float32' features. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information DPR is CC-BY-NC 4.0 licensed: URL ### Contributions Thanks to @thomwolf, @lewtun, @lhoestq for adding this dataset.
[ "### Dataset Summary\n\n\nThis is the wikipedia split used to evaluate the Dense Passage Retrieval (DPR) model.\nIt contains 21M passages from wikipedia along with their DPR embeddings.\nThe wikipedia articles were split into multiple, disjoint text blocks of 100 words as passages.\n\n\nThe wikipedia dump is the one from Dec. 20, 2018.\n\n\nThere are two types of DPR embeddings based on two different models:\n\n\n* 'nq': the model is trained on the Natural Questions dataset\n* 'multiset': the model is trained on multiple datasets\n\n\nAdditionally, a FAISS index can be created from the embeddings:\n\n\n* 'exact': with an exact FAISS index (high RAM usage)\n* 'compressed': with a compressed FAISS index (approximate, but lower RAM usage)\n* 'no\\_index': without FAISS index\n\n\nFinally, there is the possibility of generating the dataset without the embeddings:\n\n\n* 'no\\_embeddings'", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nEach instance contains a paragraph of at most 100 words, as well as the title of the wikipedia page it comes from, and the DPR embedding (a 768-d vector).", "#### psgs\\_w100.multiset.compressed\n\n\n* Size of downloaded dataset files: 70.97 GB\n* Size of the generated dataset: 78.42 GB\n* Total amount of disk used: 152.26 GB\n\n\nAn example of 'train' looks as follows.", "#### psgs\\_w100.URL\n\n\n* Size of downloaded dataset files: 70.97 GB\n* Size of the generated dataset: 78.42 GB\n* Total amount of disk used: 187.38 GB\n\n\nAn example of 'train' looks as follows.", "#### psgs\\_w100.multiset.no\\_index\n\n\n* Size of downloaded dataset files: 70.97 GB\n* Size of the generated dataset: 78.42 GB\n* Total amount of disk used: 149.38 GB\n\n\nAn example of 'train' looks as follows.", "#### psgs\\_w100.nq.compressed\n\n\n* Size of downloaded dataset files: 70.97 GB\n* Size of the generated dataset: 78.42 GB\n* Total amount of disk used: 152.26 GB\n\n\nAn example of 'train' looks as follows.", "#### psgs\\_w100.URL\n\n\n* Size of downloaded dataset files: 70.97 GB\n* Size of the generated dataset: 78.42 GB\n* Total amount of disk used: 187.38 GB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### psgs\\_w100.multiset.compressed\n\n\n* 'id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'title': a 'string' feature.\n* 'embeddings': a 'list' of 'float32' features.", "#### psgs\\_w100.URL\n\n\n* 'id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'title': a 'string' feature.\n* 'embeddings': a 'list' of 'float32' features.", "#### psgs\\_w100.multiset.no\\_index\n\n\n* 'id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'title': a 'string' feature.\n* 'embeddings': a 'list' of 'float32' features.", "#### psgs\\_w100.nq.compressed\n\n\n* 'id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'title': a 'string' feature.\n* 'embeddings': a 'list' of 'float32' features.", "#### psgs\\_w100.URL\n\n\n* 'id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'title': a 'string' feature.\n* 'embeddings': a 'list' of 'float32' features.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nDPR is CC-BY-NC 4.0 licensed: URL", "### Contributions\n\n\nThanks to @thomwolf, @lewtun, @lhoestq for adding this dataset." ]
[ "TAGS\n#task_categories-fill-mask #task_categories-text-generation #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-crowdsourced #multilinguality-multilingual #size_categories-10M<n<100M #source_datasets-original #language-English #license-cc-by-nc-4.0 #text-search #arxiv-2004.04906 #region-us \n", "### Dataset Summary\n\n\nThis is the wikipedia split used to evaluate the Dense Passage Retrieval (DPR) model.\nIt contains 21M passages from wikipedia along with their DPR embeddings.\nThe wikipedia articles were split into multiple, disjoint text blocks of 100 words as passages.\n\n\nThe wikipedia dump is the one from Dec. 20, 2018.\n\n\nThere are two types of DPR embeddings based on two different models:\n\n\n* 'nq': the model is trained on the Natural Questions dataset\n* 'multiset': the model is trained on multiple datasets\n\n\nAdditionally, a FAISS index can be created from the embeddings:\n\n\n* 'exact': with an exact FAISS index (high RAM usage)\n* 'compressed': with a compressed FAISS index (approximate, but lower RAM usage)\n* 'no\\_index': without FAISS index\n\n\nFinally, there is the possibility of generating the dataset without the embeddings:\n\n\n* 'no\\_embeddings'", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nEach instance contains a paragraph of at most 100 words, as well as the title of the wikipedia page it comes from, and the DPR embedding (a 768-d vector).", "#### psgs\\_w100.multiset.compressed\n\n\n* Size of downloaded dataset files: 70.97 GB\n* Size of the generated dataset: 78.42 GB\n* Total amount of disk used: 152.26 GB\n\n\nAn example of 'train' looks as follows.", "#### psgs\\_w100.URL\n\n\n* Size of downloaded dataset files: 70.97 GB\n* Size of the generated dataset: 78.42 GB\n* Total amount of disk used: 187.38 GB\n\n\nAn example of 'train' looks as follows.", "#### psgs\\_w100.multiset.no\\_index\n\n\n* Size of downloaded dataset files: 70.97 GB\n* Size of the generated dataset: 78.42 GB\n* Total amount of disk used: 149.38 GB\n\n\nAn example of 'train' looks as follows.", "#### psgs\\_w100.nq.compressed\n\n\n* Size of downloaded dataset files: 70.97 GB\n* Size of the generated dataset: 78.42 GB\n* Total amount of disk used: 152.26 GB\n\n\nAn example of 'train' looks as follows.", "#### psgs\\_w100.URL\n\n\n* Size of downloaded dataset files: 70.97 GB\n* Size of the generated dataset: 78.42 GB\n* Total amount of disk used: 187.38 GB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### psgs\\_w100.multiset.compressed\n\n\n* 'id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'title': a 'string' feature.\n* 'embeddings': a 'list' of 'float32' features.", "#### psgs\\_w100.URL\n\n\n* 'id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'title': a 'string' feature.\n* 'embeddings': a 'list' of 'float32' features.", "#### psgs\\_w100.multiset.no\\_index\n\n\n* 'id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'title': a 'string' feature.\n* 'embeddings': a 'list' of 'float32' features.", "#### psgs\\_w100.nq.compressed\n\n\n* 'id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'title': a 'string' feature.\n* 'embeddings': a 'list' of 'float32' features.", "#### psgs\\_w100.URL\n\n\n* 'id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'title': a 'string' feature.\n* 'embeddings': a 'list' of 'float32' features.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nDPR is CC-BY-NC 4.0 licensed: URL", "### Contributions\n\n\nThanks to @thomwolf, @lewtun, @lhoestq for adding this dataset." ]
[ 130, 237, 10, 11, 47, 64, 59, 65, 64, 59, 17, 67, 62, 68, 67, 62, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 18, 26 ]
[ "passage: TAGS\n#task_categories-fill-mask #task_categories-text-generation #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-crowdsourced #multilinguality-multilingual #size_categories-10M<n<100M #source_datasets-original #language-English #license-cc-by-nc-4.0 #text-search #arxiv-2004.04906 #region-us \n### Dataset Summary\n\n\nThis is the wikipedia split used to evaluate the Dense Passage Retrieval (DPR) model.\nIt contains 21M passages from wikipedia along with their DPR embeddings.\nThe wikipedia articles were split into multiple, disjoint text blocks of 100 words as passages.\n\n\nThe wikipedia dump is the one from Dec. 20, 2018.\n\n\nThere are two types of DPR embeddings based on two different models:\n\n\n* 'nq': the model is trained on the Natural Questions dataset\n* 'multiset': the model is trained on multiple datasets\n\n\nAdditionally, a FAISS index can be created from the embeddings:\n\n\n* 'exact': with an exact FAISS index (high RAM usage)\n* 'compressed': with a compressed FAISS index (approximate, but lower RAM usage)\n* 'no\\_index': without FAISS index\n\n\nFinally, there is the possibility of generating the dataset without the embeddings:\n\n\n* 'no\\_embeddings'### Supported Tasks and Leaderboards### Languages\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nEach instance contains a paragraph of at most 100 words, as well as the title of the wikipedia page it comes from, and the DPR embedding (a 768-d vector).#### psgs\\_w100.multiset.compressed\n\n\n* Size of downloaded dataset files: 70.97 GB\n* Size of the generated dataset: 78.42 GB\n* Total amount of disk used: 152.26 GB\n\n\nAn example of 'train' looks as follows.", "passage: #### psgs\\_w100.URL\n\n\n* Size of downloaded dataset files: 70.97 GB\n* Size of the generated dataset: 78.42 GB\n* Total amount of disk used: 187.38 GB\n\n\nAn example of 'train' looks as follows.#### psgs\\_w100.multiset.no\\_index\n\n\n* Size of downloaded dataset files: 70.97 GB\n* Size of the generated dataset: 78.42 GB\n* Total amount of disk used: 149.38 GB\n\n\nAn example of 'train' looks as follows.#### psgs\\_w100.nq.compressed\n\n\n* Size of downloaded dataset files: 70.97 GB\n* Size of the generated dataset: 78.42 GB\n* Total amount of disk used: 152.26 GB\n\n\nAn example of 'train' looks as follows.#### psgs\\_w100.URL\n\n\n* Size of downloaded dataset files: 70.97 GB\n* Size of the generated dataset: 78.42 GB\n* Total amount of disk used: 187.38 GB\n\n\nAn example of 'train' looks as follows.### Data Fields\n\n\nThe data fields are the same among all splits.#### psgs\\_w100.multiset.compressed\n\n\n* 'id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'title': a 'string' feature.\n* 'embeddings': a 'list' of 'float32' features.#### psgs\\_w100.URL\n\n\n* 'id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'title': a 'string' feature.\n* 'embeddings': a 'list' of 'float32' features.#### psgs\\_w100.multiset.no\\_index\n\n\n* 'id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'title': a 'string' feature.\n* 'embeddings': a 'list' of 'float32' features.#### psgs\\_w100.nq.compressed\n\n\n* 'id': a 'string' feature.\n* 'text': a 'string' feature.\n* 'title': a 'string' feature.\n* 'embeddings': a 'list' of 'float32' features." ]
15b7dfb153aeff20243913f4f3632548e2deabfa
# Dataset Card for WikiHop ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [QAngaroo](http://qangaroo.cs.ucl.ac.uk/) - **Repository:** [If the dataset is hosted on github or has a github homepage, add URL here]() - **Paper:** [Constructing Datasets for Multi-hop Reading Comprehension Across Documents](https://arxiv.org/abs/1710.06481) - **Leaderboard:** [leaderboard](http://qangaroo.cs.ucl.ac.uk/leaderboard.html) - **Point of Contact:** [Johannes Welbl](j.welbl@cs.ucl.ac.uk) ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@patil-suraj](https://github.com/patil-suraj) for adding this dataset.
wiki_hop
[ "task_categories:question-answering", "task_ids:extractive-qa", "annotations_creators:crowdsourced", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:cc-by-sa-3.0", "multi-hop", "arxiv:1710.06481", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["expert-generated"], "language": ["en"], "license": ["cc-by-sa-3.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["question-answering"], "task_ids": ["extractive-qa"], "paperswithcode_id": "wikihop", "pretty_name": "WikiHop", "tags": ["multi-hop"], "dataset_info": [{"config_name": "original", "features": [{"name": "id", "dtype": "string"}, {"name": "query", "dtype": "string"}, {"name": "answer", "dtype": "string"}, {"name": "candidates", "sequence": "string"}, {"name": "supports", "sequence": "string"}, {"name": "annotations", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 325952974, "num_examples": 43738}, {"name": "validation", "num_bytes": 41246536, "num_examples": 5129}], "download_size": 339843061, "dataset_size": 367199510}, {"config_name": "masked", "features": [{"name": "id", "dtype": "string"}, {"name": "question", "dtype": "string"}, {"name": "answer", "dtype": "string"}, {"name": "candidates", "sequence": "string"}, {"name": "supports", "sequence": "string"}, {"name": "annotations", "sequence": {"sequence": "string"}}], "splits": [{"name": "train", "num_bytes": 348249138, "num_examples": 43738}, {"name": "validation", "num_bytes": 44066862, "num_examples": 5129}], "download_size": 339843061, "dataset_size": 392316000}]}
2024-01-18T11:18:04+00:00
[ "1710.06481" ]
[ "en" ]
TAGS #task_categories-question-answering #task_ids-extractive-qa #annotations_creators-crowdsourced #language_creators-expert-generated #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-sa-3.0 #multi-hop #arxiv-1710.06481 #region-us
# Dataset Card for WikiHop ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: QAngaroo - Repository: [If the dataset is hosted on github or has a github homepage, add URL here]() - Paper: Constructing Datasets for Multi-hop Reading Comprehension Across Documents - Leaderboard: leaderboard - Point of Contact: Johannes Welbl ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @patil-suraj for adding this dataset.
[ "# Dataset Card for WikiHop", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: QAngaroo\n- Repository: [If the dataset is hosted on github or has a github homepage, add URL here]()\n- Paper: Constructing Datasets for Multi-hop Reading Comprehension Across Documents\n- Leaderboard: leaderboard\n- Point of Contact: Johannes Welbl", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @patil-suraj for adding this dataset." ]
[ "TAGS\n#task_categories-question-answering #task_ids-extractive-qa #annotations_creators-crowdsourced #language_creators-expert-generated #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-sa-3.0 #multi-hop #arxiv-1710.06481 #region-us \n", "# Dataset Card for WikiHop", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: QAngaroo\n- Repository: [If the dataset is hosted on github or has a github homepage, add URL here]()\n- Paper: Constructing Datasets for Multi-hop Reading Comprehension Across Documents\n- Leaderboard: leaderboard\n- Point of Contact: Johannes Welbl", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @patil-suraj for adding this dataset." ]
[ 109, 8, 120, 76, 6, 10, 4, 6, 6, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 19 ]
[ "passage: TAGS\n#task_categories-question-answering #task_ids-extractive-qa #annotations_creators-crowdsourced #language_creators-expert-generated #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-cc-by-sa-3.0 #multi-hop #arxiv-1710.06481 #region-us \n# Dataset Card for WikiHop## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: QAngaroo\n- Repository: [If the dataset is hosted on github or has a github homepage, add URL here]()\n- Paper: Constructing Datasets for Multi-hop Reading Comprehension Across Documents\n- Leaderboard: leaderboard\n- Point of Contact: Johannes Welbl### Dataset Summary### Supported Tasks and Leaderboards### Languages## Dataset Structure### Data Instances### Data Fields### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators### Licensing Information### Contributions\n\nThanks to @patil-suraj for adding this dataset." ]
ea3db3510cbd34d0f8dc612419ae40e4732f3b40
# Dataset Card for "wiki_lingua" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Repository:** [URL](https://github.com/esdurmus/Wikilingua) - **Paper:** [WikiLingua: A Multilingual Abstractive Summarization Dataset](https://arxiv.org/abs/2010.03093) ### Dataset Summary We introduce WikiLingua, a large-scale, multilingual dataset for the evaluation of cross-lingual abstractive summarization systems. We extract article and summary pairs in 18 languages from WikiHow, a high quality, collaborative resource of how-to guides on a diverse set of topics written by human authors. We create gold-standard article-summary alignments across languages by aligning the images that are used to describe each how-to step in an article. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The table below shows number of article-summary pairs with a parallel article-summary pair in English. ______________________________ | Language | Num. parallel | | ----------- | --------------| | English | 141,457 | | Spanish | 113,215 | | Portuguese | 81,695 | | French | 63,692 | | German | 58,375 | | Russian | 52,928 | | Italian | 50,968 | | Indonesian | 47,511 | | Dutch | 31,270 | | Arabic | 29,229 | | Vietnamese | 19,600 | | Chinese | 18,887 | | Thai | 14,770 | | Japanese | 12,669 | | Korean | 12,189 | | Hindi | 9,929 | | Czech | 7,200 | | Turkish | 4,503 | ## Dataset Structure ### Data Instances ``` { 'article': { 'document': ['make sure that the area is a safe place, especially if you plan on walking home at night. It’s always a good idea to practice the buddy system. Have a friend meet up and walk with you. Research the bus, train, or streetcar routes available in your area to find safe and affordable travel to your destination. Make sure you check the schedule for your outgoing and return travel. Some public transportation will cease to run late at night. Be sure if you take public transportation to the venue that you will also be able to get home late at night. Check the routes. Even if some public transit is still running late at night, the routing may change. Some may run express past many of the stops, or not travel all the way to the ends. Be sure that your stop will still be available when you need it for your return trip. If you are taking public transit in a vulnerable state after drinking, it is always a good idea to travel in groups. Having friends available is a good way to stay safe and make sure that you reach your destination. This is more expensive option than a taxi or ride share service, but could be a fun and fancy way to stay safe and ensure that you will have a ride home. Plan this service in advance with a scheduled time to pick you up from your home and the venue. You want to be sure that the service will still be available when you need to get home. This may be easy in a large city, but taxis may be less frequent in smaller towns. This is especially true late at night, so this is a less reliable option than scheduling a ride in advance. Have a friend accompany you and help you flag a cab to make sure you are able to get one. Set up a plan to call a friend when you get home to make sure that you made it safely to your destination. If there are no taxis readily available call a local service to send a car to pick you up. You can share a ride with your friends, or other people using the app at the same moment. If you are in a vulnerable state it is best to share the ride with your friends to make sure you get home safe. You can request the car to yourself rather than sharing rides with strangers. If you travel home on your own or are the last of your group to be dropped off, make plans to call a friend when you get home so they know you made it safely to your destination. There may be a designated driver service in your area which can chauffeur your group. Make reservations with them in advance and keep their contact information handy while you are drinking.', "Designating a driver is a very popular tactic to avoid drinking and driving. It is important to plan in advance, because your brain function will slow down and your decision making skills will be impaired once you start drinking. Decide before you begin drinking that you will not drive. Figure out who will be getting you home before you leave. Make sure this person is responsible and keep them in your sight while you are drinking. Have their contact information handy in case you can’t find them when you are ready to leave. Choose a friend who doesn’t drink alcohol. You likely have someone in your friend group who doesn’t drink. This person is the most likely to remain sober. Decide on one person who will remain sober. You can take turns within your friend group, alternating who will be the designated driver on each occasion. Be sure that the designated driver actually remains sober. The person who has drank the least is still not sober. If you don’t have your car with you, you can guarantee that you won’t make the choice to drive it home. If you are drinking at your home. Give your keys to a responsible friend to ensure that you don't choose to drive somewhere after you have been drinking. It may be tempting to stay longer or leave with someone else. Stick to the plan you made in advance and only leave with your sober, designated driver. Keep the phone number of your driver handy in case you can't find them when you are ready to leave. If your designated driver drinks alcohol, find alternate transportation to get home.", 'If you have been drinking at all you are at least on the spectrum of drunkenness. You could be showing signs of impairment and slower brain function including lack of motor skills and slower reaction time, leading to the inability to operate a motor vehicle. Some of these signs could be: Poor balance or stumbling. Difficulty speaking clearly and slurred words. Abnormal behavior leading to you doing things you wouldn’t normally do if you were sober. As soon as you notice that you are showing signs of impairment, give your keys to a friend, the host or the bartender to ensure that you won’t drive until you are sober. Make sure to only give them your car key. Hold onto your house keys. If your friend, the host or the bartender are advising you not to drive, you are likely too drunk. Listen to their advice and acknowledge that they are trying to help you. Bystander intervention is common when it comes to drinking and driving. Many people will be willing to step in, take your keys and help you get home safely. If no one if offering to help, you may need to ask. Take a ride from a sober friend. It is best to get in a car with someone you trust when you are in this vulnerable state. Allow the host or bartender to call a cab or car service to take you home. If you are having a difficult time finding a safe way to get home, find a place to stay which does not involve you driving. Ask the host of the party if there is a place you can sleep. Give them your keys and ask that they keep them in a safe place until the morning. Stay with a friend if they live nearby and are on their way home. Find a hotel within walking distance. Call them to book a room, or have a friend help you secure one. Ask the friend if they will walk you to the hotel and make sure you get checked in safely. There are people in your life who care about you and want to be sure that you are safe. It may seem scary or embarrassing to call your parents or your siblings if you are too drunk to drive, but they will be glad you did. Your safety is the most important. You may need your phone to call someone for a ride or get help from a friend. Be sure to charge your phone before you leave the house. It is also a good idea to bring a charger with you in case your battery dies before the end of the night or you end up staying where you are and need to get home the next morning. You may also want to invest in a portable battery charger for your phone should there not be a power outlet available. Make sure it is fully charged before you leave your house. Keep it handy in your pocket or your bag throughout the night.' ], 'section_name': ['Finding Other Transportation', 'Designating a Driver', 'Staying Safe' ], 'summary': ['Walk to the venue where you will be drinking if it is close enough. Take public transit. Show up in style by hiring a limo or black car service. Flag a taxi cab for a convenient option to get where you’re going. Request a rideshare service like Uber or Lyft using an app on your phone. Reserve a designated driver service.', 'Plan in advance. Assign a designated driver. Leave your car at home. Leave the venue with your designated driver.', 'Pay attention to your body. Give up your keys. Listen to other people. Accept help. Stay where you are. Have an emergency back-up plan. Make sure that your phone is charged.' ] }, 'url': 'https://www.wikihow.com/Avoid-Drinking-and-Driving' } ``` ### Data Fields - `url`: WikiHow URL of the article - `article`: A dictionary containing `section_name`, `document` and `summary` - `section_name`: List of section headings in an article - `document`: List of documents, one for each section in the `section_name` list - `summary`: List of summarized document ### Data Splits | | train | |:-----------|--------:| | arabic | 9995 | | chinese | 6541 | | czech | 2520 | | dutch | 10862 | | english | 57945 | | french | 21690 | | german | 20103 | | hindi | 3402 | | indonesian | 16308 | | italian | 17673 | | japanese | 4372 | | korean | 4111 | | portuguese | 28143 | | russian | 18143 | | spanish | 6616 | | thai | 5093 | | turkish | 1512 | | vietnamese | 6616 | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information - Article provided by wikiHow https://www.wikihow.com/Main-Page, a wiki building the world's largest, highest quality how-to manual. Please edit this article and find author credits at wikiHow.com. Content on wikiHow can be shared under a [Creative Commons license](http://creativecommons.org/licenses/by-nc-sa/3.0/). - Refer to [this webpage](https://www.wikihow.com/wikiHow:Attribution) for the specific attribution guidelines. - also see https://gem-benchmark.com/data_cards/WikiLingua ### Citation Information ```bibtex @inproceedings{ladhak-etal-2020-wikilingua, title = "{W}iki{L}ingua: A New Benchmark Dataset for Cross-Lingual Abstractive Summarization", author = "Ladhak, Faisal and Durmus, Esin and Cardie, Claire and McKeown, Kathleen", booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2020", month = nov, year = "2020", address = "Online", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2020.findings-emnlp.360", doi = "10.18653/v1/2020.findings-emnlp.360", pages = "4034--4048", } ``` ### Contributions Thanks to [@katnoria](https://github.com/katnoria) for adding this dataset.
wiki_lingua
[ "task_categories:summarization", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:multilingual", "size_categories:10K<n<100K", "size_categories:1K<n<10K", "source_datasets:original", "language:ar", "language:cs", "language:de", "language:en", "language:es", "language:fr", "language:hi", "language:id", "language:it", "language:ja", "language:ko", "language:nl", "language:pt", "language:ru", "language:th", "language:tr", "language:vi", "language:zh", "license:cc-by-3.0", "arxiv:2010.03093", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["ar", "cs", "de", "en", "es", "fr", "hi", "id", "it", "ja", "ko", "nl", "pt", "ru", "th", "tr", "vi", "zh"], "license": ["cc-by-3.0"], "multilinguality": ["multilingual"], "size_categories": ["10K<n<100K", "1K<n<10K"], "source_datasets": ["original"], "task_categories": ["summarization"], "task_ids": [], "paperswithcode_id": "wikilingua", "pretty_name": "WikiLingua", "config_names": ["arabic", "chinese", "czech", "dutch", "english", "french", "german", "hindi", "indonesian", "italian", "japanese", "korean", "portuguese", "russian", "spanish", "thai", "turkish", "vietnamese"], "dataset_info": [{"config_name": "arabic", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 119116075, "num_examples": 9995}], "download_size": 55808460, "dataset_size": 119116075}, {"config_name": "chinese", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 41170645, "num_examples": 6541}], "download_size": 25187026, "dataset_size": 41170645}, {"config_name": "czech", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 20816346, "num_examples": 2520}], "download_size": 12480761, "dataset_size": 20816346}, {"config_name": "dutch", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 87257952, "num_examples": 10862}], "download_size": 47651076, "dataset_size": 87257952}, {"config_name": "english", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 333699946, "num_examples": 57945}], "download_size": 187189233, "dataset_size": 333699946}, {"config_name": "french", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 197550244, "num_examples": 21690}], "download_size": 105158840, "dataset_size": 197550244}, {"config_name": "german", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 168674208, "num_examples": 20103}], "download_size": 93078076, "dataset_size": 168674208}, {"config_name": "hindi", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 63785007, "num_examples": 3402}], "download_size": 22774620, "dataset_size": 63785007}, {"config_name": "indonesian", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 136408773, "num_examples": 16308}], "download_size": 67658970, "dataset_size": 136408773}, {"config_name": "italian", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 138119439, "num_examples": 17673}], "download_size": 78108134, "dataset_size": 138119439}, {"config_name": "japanese", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 40144987, "num_examples": 4372}], "download_size": 19794488, "dataset_size": 40144987}, {"config_name": "korean", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 38647570, "num_examples": 4111}], "download_size": 20029486, "dataset_size": 38647570}, {"config_name": "portuguese", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 204270713, "num_examples": 28143}], "download_size": 114735912, "dataset_size": 204270713}, {"config_name": "russian", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 241923944, "num_examples": 18143}], "download_size": 111025228, "dataset_size": 241923944}, {"config_name": "spanish", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 314618442, "num_examples": 38795}], "download_size": 170995186, "dataset_size": 314618442}, {"config_name": "thai", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 86982807, "num_examples": 5093}], "download_size": 31944979, "dataset_size": 86982807}, {"config_name": "turkish", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 11371777, "num_examples": 1512}], "download_size": 5964904, "dataset_size": 11371777}, {"config_name": "vietnamese", "features": [{"name": "url", "dtype": "string"}, {"name": "article", "sequence": [{"name": "section_name", "dtype": "string"}, {"name": "document", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "english_url", "dtype": "string"}, {"name": "english_section_name", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 69868744, "num_examples": 6616}], "download_size": 33194150, "dataset_size": 69868744}], "configs": [{"config_name": "arabic", "data_files": [{"split": "train", "path": "arabic/train-*"}]}, {"config_name": "chinese", "data_files": [{"split": "train", "path": "chinese/train-*"}]}, {"config_name": "czech", "data_files": [{"split": "train", "path": "czech/train-*"}]}, {"config_name": "dutch", "data_files": [{"split": "train", "path": "dutch/train-*"}]}, {"config_name": "english", "data_files": [{"split": "train", "path": "english/train-*"}], "default": true}, {"config_name": "french", "data_files": [{"split": "train", "path": "french/train-*"}]}, {"config_name": "german", "data_files": [{"split": "train", "path": "german/train-*"}]}, {"config_name": "hindi", "data_files": [{"split": "train", "path": "hindi/train-*"}]}, {"config_name": "indonesian", "data_files": [{"split": "train", "path": "indonesian/train-*"}]}, {"config_name": "italian", "data_files": [{"split": "train", "path": "italian/train-*"}]}, {"config_name": "japanese", "data_files": [{"split": "train", "path": "japanese/train-*"}]}, {"config_name": "korean", "data_files": [{"split": "train", "path": "korean/train-*"}]}, {"config_name": "portuguese", "data_files": [{"split": "train", "path": "portuguese/train-*"}]}, {"config_name": "russian", "data_files": [{"split": "train", "path": "russian/train-*"}]}, {"config_name": "spanish", "data_files": [{"split": "train", "path": "spanish/train-*"}]}, {"config_name": "thai", "data_files": [{"split": "train", "path": "thai/train-*"}]}, {"config_name": "turkish", "data_files": [{"split": "train", "path": "turkish/train-*"}]}, {"config_name": "vietnamese", "data_files": [{"split": "train", "path": "vietnamese/train-*"}]}]}
2024-01-05T08:06:54+00:00
[ "2010.03093" ]
[ "ar", "cs", "de", "en", "es", "fr", "hi", "id", "it", "ja", "ko", "nl", "pt", "ru", "th", "tr", "vi", "zh" ]
TAGS #task_categories-summarization #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-multilingual #size_categories-10K<n<100K #size_categories-1K<n<10K #source_datasets-original #language-Arabic #language-Czech #language-German #language-English #language-Spanish #language-French #language-Hindi #language-Indonesian #language-Italian #language-Japanese #language-Korean #language-Dutch #language-Portuguese #language-Russian #language-Thai #language-Turkish #language-Vietnamese #language-Chinese #license-cc-by-3.0 #arxiv-2010.03093 #region-us
Dataset Card for "wiki\_lingua" =============================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Repository: URL * Paper: WikiLingua: A Multilingual Abstractive Summarization Dataset ### Dataset Summary We introduce WikiLingua, a large-scale, multilingual dataset for the evaluation of cross-lingual abstractive summarization systems. We extract article and summary pairs in 18 languages from WikiHow, a high quality, collaborative resource of how-to guides on a diverse set of topics written by human authors. We create gold-standard article-summary alignments across languages by aligning the images that are used to describe each how-to step in an article. ### Supported Tasks and Leaderboards ### Languages The table below shows number of article-summary pairs with a parallel article-summary pair in English. --- Dataset Structure ----------------- ### Data Instances ### Data Fields * 'url': WikiHow URL of the article * 'article': A dictionary containing 'section\_name', 'document' and 'summary' + 'section\_name': List of section headings in an article + 'document': List of documents, one for each section in the 'section\_name' list + 'summary': List of summarized document ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information * Article provided by wikiHow URL a wiki building the world's largest, highest quality how-to manual. Please edit this article and find author credits at URL. Content on wikiHow can be shared under a Creative Commons license. * Refer to this webpage for the specific attribution guidelines. * also see URL ### Contributions Thanks to @katnoria for adding this dataset.
[ "### Dataset Summary\n\n\nWe introduce WikiLingua, a large-scale, multilingual dataset for the evaluation of cross-lingual abstractive summarization systems. We extract article and summary pairs in 18 languages from WikiHow, a high quality, collaborative resource of how-to guides on a diverse set of topics written by human authors. We create gold-standard article-summary alignments across languages by aligning the images that are used to describe each how-to step in an article.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe table below shows number of article-summary pairs with a parallel article-summary pair in English.\n\n\n\n\n---\n\n\n\nDataset Structure\n-----------------", "### Data Instances", "### Data Fields\n\n\n* 'url': WikiHow URL of the article\n* 'article': A dictionary containing 'section\\_name', 'document' and 'summary'\n\t+ 'section\\_name': List of section headings in an article\n\t+ 'document': List of documents, one for each section in the 'section\\_name' list\n\t+ 'summary': List of summarized document", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\n* Article provided by wikiHow URL a wiki building the world's largest, highest quality how-to manual. Please edit this article and find author credits at URL. Content on wikiHow can be shared under a Creative Commons license.\n* Refer to this webpage for the specific attribution guidelines.\n* also see URL", "### Contributions\n\n\nThanks to @katnoria for adding this dataset." ]
[ "TAGS\n#task_categories-summarization #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-multilingual #size_categories-10K<n<100K #size_categories-1K<n<10K #source_datasets-original #language-Arabic #language-Czech #language-German #language-English #language-Spanish #language-French #language-Hindi #language-Indonesian #language-Italian #language-Japanese #language-Korean #language-Dutch #language-Portuguese #language-Russian #language-Thai #language-Turkish #language-Vietnamese #language-Chinese #license-cc-by-3.0 #arxiv-2010.03093 #region-us \n", "### Dataset Summary\n\n\nWe introduce WikiLingua, a large-scale, multilingual dataset for the evaluation of cross-lingual abstractive summarization systems. We extract article and summary pairs in 18 languages from WikiHow, a high quality, collaborative resource of how-to guides on a diverse set of topics written by human authors. We create gold-standard article-summary alignments across languages by aligning the images that are used to describe each how-to step in an article.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe table below shows number of article-summary pairs with a parallel article-summary pair in English.\n\n\n\n\n---\n\n\n\nDataset Structure\n-----------------", "### Data Instances", "### Data Fields\n\n\n* 'url': WikiHow URL of the article\n* 'article': A dictionary containing 'section\\_name', 'document' and 'summary'\n\t+ 'section\\_name': List of section headings in an article\n\t+ 'document': List of documents, one for each section in the 'section\\_name' list\n\t+ 'summary': List of summarized document", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\n* Article provided by wikiHow URL a wiki building the world's largest, highest quality how-to manual. Please edit this article and find author credits at URL. Content on wikiHow can be shared under a Creative Commons license.\n* Refer to this webpage for the specific attribution guidelines.\n* also see URL", "### Contributions\n\n\nThanks to @katnoria for adding this dataset." ]
[ 192, 114, 10, 35, 6, 95, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 71, 17 ]
[ "passage: TAGS\n#task_categories-summarization #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-multilingual #size_categories-10K<n<100K #size_categories-1K<n<10K #source_datasets-original #language-Arabic #language-Czech #language-German #language-English #language-Spanish #language-French #language-Hindi #language-Indonesian #language-Italian #language-Japanese #language-Korean #language-Dutch #language-Portuguese #language-Russian #language-Thai #language-Turkish #language-Vietnamese #language-Chinese #license-cc-by-3.0 #arxiv-2010.03093 #region-us \n### Dataset Summary\n\n\nWe introduce WikiLingua, a large-scale, multilingual dataset for the evaluation of cross-lingual abstractive summarization systems. We extract article and summary pairs in 18 languages from WikiHow, a high quality, collaborative resource of how-to guides on a diverse set of topics written by human authors. We create gold-standard article-summary alignments across languages by aligning the images that are used to describe each how-to step in an article.### Supported Tasks and Leaderboards### Languages\n\n\nThe table below shows number of article-summary pairs with a parallel article-summary pair in English.\n\n\n\n\n---\n\n\n\nDataset Structure\n-----------------### Data Instances### Data Fields\n\n\n* 'url': WikiHow URL of the article\n* 'article': A dictionary containing 'section\\_name', 'document' and 'summary'\n\t+ 'section\\_name': List of section headings in an article\n\t+ 'document': List of documents, one for each section in the 'section\\_name' list\n\t+ 'summary': List of summarized document### Data Splits\n\n\n\nDataset Creation\n----------------### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process" ]
6faeebe97e325caefbb0aeaf6f3a5b4d749bf5af
# Dataset Card for WikiMovies ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [WikiMovies Homepage](https://research.fb.com/downloads/babi/) - **Repository:** - **Paper:** [Key-Value Memory Networks for Directly Reading Documents](https://arxiv.org/pdf/1606.03126.pdf) - **Leaderboard:** - **Point of Contact:** ### Dataset Summary The WikiMovies dataset consists of roughly 100k (templated) questions over 75k entitiesbased on questions with answers in the open movie database (OMDb). It is the QA part of the Movie Dialog dataset. ### Supported Tasks and Leaderboards - Question Answering ### Languages The text in the dataset is written in English. ## Dataset Structure ### Data Instances The raw data consists of question answer pairs separated by a tab. Here are 3 examples: ```buildoutcfg 1 what does Grégoire Colin appear in? Before the Rain 1 Joe Thomas appears in which movies? The Inbetweeners Movie, The Inbetweeners 2 1 what films did Michelle Trachtenberg star in? Inspector Gadget, Black Christmas, Ice Princess, Harriet the Spy, The Scribbler ``` It is unclear what the `1` is for at the beginning of each line, but it has been removed in the `Dataset` object. ### Data Fields Here is an example of the raw data ingested by `Datasets`: ```buildoutcfg { 'answer': 'Before the Rain', 'question': 'what does Grégoire Colin appear in?' } ``` `answer`: a string containing the answer to a corresponding question. `question`: a string containing the relevant question. ### Data Splits The data is split into train, test, and dev sets. The split sizes are as follows: | wiki-entities_qa_* | n examples| | ----- | ---- | | train.txt | 96185 | | dev.txt | 10000 | | test.txt | 9952 | ## Dataset Creation ### Curation Rationale WikiMovies was built with the following goals in mind: (i) machine learning techniques should have ample training examples for learning; and (ii) one can analyze easily the performance of different representations of knowledge and break down the results by question type. The datasetcan be downloaded fromhttp://fb.ai/babi ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ``` @misc{miller2016keyvalue, title={Key-Value Memory Networks for Directly Reading Documents}, author={Alexander Miller and Adam Fisch and Jesse Dodge and Amir-Hossein Karimi and Antoine Bordes and Jason Weston}, year={2016}, eprint={1606.03126}, archivePrefix={arXiv}, primaryClass={cs.CL} ``` ### Contributions Thanks to [@aclifton314](https://github.com/aclifton314) for adding this dataset.
wiki_movies
[ "task_categories:question-answering", "task_ids:closed-domain-qa", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:en", "license:cc-by-3.0", "arxiv:1606.03126", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["cc-by-3.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["question-answering"], "task_ids": ["closed-domain-qa"], "paperswithcode_id": "wikimovies", "pretty_name": "WikiMovies", "dataset_info": {"features": [{"name": "question", "dtype": "string"}, {"name": "answer", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 7274490, "num_examples": 96185}, {"name": "test", "num_bytes": 755258, "num_examples": 9952}, {"name": "validation", "num_bytes": 754755, "num_examples": 10000}], "download_size": 57070041, "dataset_size": 8784503}}
2024-01-18T11:18:06+00:00
[ "1606.03126" ]
[ "en" ]
TAGS #task_categories-question-answering #task_ids-closed-domain-qa #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-3.0 #arxiv-1606.03126 #region-us
Dataset Card for WikiMovies =========================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: WikiMovies Homepage * Repository: * Paper: Key-Value Memory Networks for Directly Reading Documents * Leaderboard: * Point of Contact: ### Dataset Summary The WikiMovies dataset consists of roughly 100k (templated) questions over 75k entitiesbased on questions with answers in the open movie database (OMDb). It is the QA part of the Movie Dialog dataset. ### Supported Tasks and Leaderboards * Question Answering ### Languages The text in the dataset is written in English. Dataset Structure ----------------- ### Data Instances The raw data consists of question answer pairs separated by a tab. Here are 3 examples: It is unclear what the '1' is for at the beginning of each line, but it has been removed in the 'Dataset' object. ### Data Fields Here is an example of the raw data ingested by 'Datasets': 'answer': a string containing the answer to a corresponding question. 'question': a string containing the relevant question. ### Data Splits The data is split into train, test, and dev sets. The split sizes are as follows: Dataset Creation ---------------- ### Curation Rationale WikiMovies was built with the following goals in mind: (i) machine learning techniques should have ample training examples for learning; and (ii) one can analyze easily the performance of different representations of knowledge and break down the results by question type. The datasetcan be downloaded fromhttp://URL ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @aclifton314 for adding this dataset.
[ "### Dataset Summary\n\n\nThe WikiMovies dataset consists of roughly 100k (templated) questions over 75k entitiesbased on questions with answers in the open movie database (OMDb). It is the QA part of the Movie Dialog dataset.", "### Supported Tasks and Leaderboards\n\n\n* Question Answering", "### Languages\n\n\nThe text in the dataset is written in English.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nThe raw data consists of question answer pairs separated by a tab. Here are 3 examples:\n\n\nIt is unclear what the '1' is for at the beginning of each line, but it has been removed in the 'Dataset' object.", "### Data Fields\n\n\nHere is an example of the raw data ingested by 'Datasets':\n\n\n'answer': a string containing the answer to a corresponding question.\n'question': a string containing the relevant question.", "### Data Splits\n\n\nThe data is split into train, test, and dev sets. The split sizes are as follows:\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nWikiMovies was built with the following goals in mind: (i) machine learning techniques should have ample training examples for learning; and (ii) one can analyze easily the performance of different representations of knowledge and break down the results by question type. The datasetcan be downloaded fromhttp://URL", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @aclifton314 for adding this dataset." ]
[ "TAGS\n#task_categories-question-answering #task_ids-closed-domain-qa #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-3.0 #arxiv-1606.03126 #region-us \n", "### Dataset Summary\n\n\nThe WikiMovies dataset consists of roughly 100k (templated) questions over 75k entitiesbased on questions with answers in the open movie database (OMDb). It is the QA part of the Movie Dialog dataset.", "### Supported Tasks and Leaderboards\n\n\n* Question Answering", "### Languages\n\n\nThe text in the dataset is written in English.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nThe raw data consists of question answer pairs separated by a tab. Here are 3 examples:\n\n\nIt is unclear what the '1' is for at the beginning of each line, but it has been removed in the 'Dataset' object.", "### Data Fields\n\n\nHere is an example of the raw data ingested by 'Datasets':\n\n\n'answer': a string containing the answer to a corresponding question.\n'question': a string containing the relevant question.", "### Data Splits\n\n\nThe data is split into train, test, and dev sets. The split sizes are as follows:\n\n\n\nDataset Creation\n----------------", "### Curation Rationale\n\n\nWikiMovies was built with the following goals in mind: (i) machine learning techniques should have ample training examples for learning; and (ii) one can analyze easily the performance of different representations of knowledge and break down the results by question type. The datasetcan be downloaded fromhttp://URL", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @aclifton314 for adding this dataset." ]
[ 104, 59, 14, 22, 59, 52, 34, 72, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 6, 19 ]
[ "passage: TAGS\n#task_categories-question-answering #task_ids-closed-domain-qa #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-3.0 #arxiv-1606.03126 #region-us \n### Dataset Summary\n\n\nThe WikiMovies dataset consists of roughly 100k (templated) questions over 75k entitiesbased on questions with answers in the open movie database (OMDb). It is the QA part of the Movie Dialog dataset.### Supported Tasks and Leaderboards\n\n\n* Question Answering### Languages\n\n\nThe text in the dataset is written in English.\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nThe raw data consists of question answer pairs separated by a tab. Here are 3 examples:\n\n\nIt is unclear what the '1' is for at the beginning of each line, but it has been removed in the 'Dataset' object.### Data Fields\n\n\nHere is an example of the raw data ingested by 'Datasets':\n\n\n'answer': a string containing the answer to a corresponding question.\n'question': a string containing the relevant question.### Data Splits\n\n\nThe data is split into train, test, and dev sets. The split sizes are as follows:\n\n\n\nDataset Creation\n----------------### Curation Rationale\n\n\nWikiMovies was built with the following goals in mind: (i) machine learning techniques should have ample training examples for learning; and (ii) one can analyze easily the performance of different representations of knowledge and break down the results by question type. The datasetcan be downloaded fromhttp://URL### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------### Social Impact of Dataset### Discussion of Biases### Other Known Limitations\n\n\nAdditional Information\n----------------------" ]
3f104672b5de699878fe7907afc486f0de325eb5
# Dataset Card for "wiki_qa" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://www.microsoft.com/en-us/download/details.aspx?id=52419](https://www.microsoft.com/en-us/download/details.aspx?id=52419) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [WikiQA: A Challenge Dataset for Open-Domain Question Answering](https://aclanthology.org/D15-1237/) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 7.10 MB - **Size of the generated dataset:** 6.40 MB - **Total amount of disk used:** 13.50 MB ### Dataset Summary Wiki Question Answering corpus from Microsoft. The WikiQA corpus is a publicly available set of question and sentence pairs, collected and annotated for research on open-domain question answering. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### default - **Size of downloaded dataset files:** 7.10 MB - **Size of the generated dataset:** 6.40 MB - **Total amount of disk used:** 13.50 MB An example of 'train' looks as follows. ``` { "answer": "Glacier caves are often called ice caves , but this term is properly used to describe bedrock caves that contain year-round ice.", "document_title": "Glacier cave", "label": 0, "question": "how are glacier caves formed?", "question_id": "Q1" } ``` ### Data Fields The data fields are the same among all splits. #### default - `question_id`: a `string` feature. - `question`: a `string` feature. - `document_title`: a `string` feature. - `answer`: a `string` feature. - `label`: a classification label, with possible values including `0` (0), `1` (1). ### Data Splits | name |train|validation|test| |-------|----:|---------:|---:| |default|20360| 2733|6165| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information MICROSOFT RESEARCH DATA LICENSE AGREEMENT FOR MICROSOFT RESEARCH WIKIQA CORPUS These license terms are an agreement between Microsoft Corporation (or based on where you live, one of its affiliates) and you. Please read them. They apply to the data associated with this license above, which includes the media on which you received it, if any. The terms also apply to any Microsoft: - updates, - supplements, - Internet-based services, and - support services for this data, unless other terms accompany those items. If so, those terms apply. BY USING THE DATA, YOU ACCEPT THESE TERMS. IF YOU DO NOT ACCEPT THEM, DO NOT USE THE DATA. If you comply with these license terms, you have the rights below. 1. SCOPE OF LICENSE. a. You may use, copy, modify, create derivative works, and distribute the Dataset: i. for research and technology development purposes only. Examples of research and technology development uses are teaching, academic research, public demonstrations and experimentation ; and ii. to publish (or present papers or articles) on your results from using such Dataset. b. The data is licensed, not sold. This agreement only gives you some rights to use the data. Microsoft reserves all other rights. Unless applicable law gives you more rights despite this limitation, you may use the data only as expressly permitted in this agreement. In doing so, you must comply with any technical limitations in the data that only allow you to use it in certain ways. You may not - work around any technical limitations in the data; - reverse engineer, decompile or disassemble the data, except and only to the extent that applicable law expressly permits, despite this limitation; - rent, lease or lend the data; - transfer the data or this agreement to any third party; or - use the data directly in a commercial product without Microsoft’s permission. 2. DISTRIBUTION REQUIREMENTS: a. If you distribute the Dataset or any derivative works of the Dataset, you will distribute them under the same terms and conditions as in this Agreement, and you will not grant other rights to the Dataset or derivative works that are different from those provided by this Agreement. b. If you have created derivative works of the Dataset, and distribute such derivative works, you will cause the modified files to carry prominent notices so that recipients know that they are not receiving Page 1 of 3the original Dataset. Such notices must state: (i) that you have changed the Dataset; and (ii) the date of any changes. 3. DISTRIBUTION RESTRICTIONS. You may not: (a) alter any copyright, trademark or patent notice in the Dataset; (b) use Microsoft’s trademarks in a way that suggests your derivative works or modifications come from or are endorsed by Microsoft; (c) include the Dataset in malicious, deceptive or unlawful programs. 4. OWNERSHIP. Microsoft retains all right, title, and interest in and to any Dataset provided to you under this Agreement. You acquire no interest in the Dataset you may receive under the terms of this Agreement. 5. LICENSE TO MICROSOFT. Microsoft is granted back, without any restrictions or limitations, a non-exclusive, perpetual, irrevocable, royalty-free, assignable and sub-licensable license, to reproduce, publicly perform or display, use, modify, post, distribute, make and have made, sell and transfer your modifications to and/or derivative works of the Dataset, for any purpose. 6. FEEDBACK. If you give feedback about the Dataset to Microsoft, you give to Microsoft, without charge, the right to use, share and commercialize your feedback in any way and for any purpose. You also give to third parties, without charge, any patent rights needed for their products, technologies and services to use or interface with any specific parts of a Microsoft dataset or service that includes the feedback. You will not give feedback that is subject to a license that requires Microsoft to license its Dataset or documentation to third parties because we include your feedback in them. These rights survive this Agreement. 7. EXPORT RESTRICTIONS. The Dataset is subject to United States export laws and regulations. You must comply with all domestic and international export laws and regulations that apply to the Dataset. These laws include restrictions on destinations, end users and end use. For additional information, see www.microsoft.com/exporting. 8. ENTIRE AGREEMENT. This Agreement, and the terms for supplements, updates, Internet-based services and support services that you use, are the entire agreement for the Dataset. 9. SUPPORT SERVICES. Because this data is “as is,” we may not provide support services for it. 10. APPLICABLE LAW. a. United States. If you acquired the software in the United States, Washington state law governs the interpretation of this agreement and applies to claims for breach of it, regardless of conflict of laws principles. The laws of the state where you live govern all other claims, including claims under state consumer protection laws, unfair competition laws, and in tort. b. Outside the United States. If you acquired the software in any other country, the laws of that country apply. 11. LEGAL EFFECT. This Agreement describes certain legal rights. You may have other rights under the laws of your country. You may also have rights with respect to the party from whom you acquired the Dataset. This Agreement does not change your rights under the laws of your country if the laws of your country do not permit it to do so. 12. DISCLAIMER OF WARRANTY. The Dataset is licensed “as-is.” You bear the risk of using it. Microsoft gives no express warranties, guarantees or conditions. You may have additional consumer rights or statutory guarantees under your local laws which this agreement cannot change. To the extent permitted under your local laws, Microsoft excludes the implied warranties of merchantability, fitness for a particular purpose and non- infringement. 13. LIMITATION ON AND EXCLUSION OF REMEDIES AND DAMAGES. YOU CAN RECOVER FROM MICROSOFT AND ITS SUPPLIERS ONLY DIRECT DAMAGES UP TO U.S. $5.00. YOU CANNOT RECOVER ANY OTHER DAMAGES, INCLUDING CONSEQUENTIAL, LOST PROFITS, SPECIAL, INDIRECT OR INCIDENTAL DAMAGES. This limitation applies to - anything related to the software, services, content (including code) on third party Internet sites, or third party programs; and Page 2 of 3 - claims for breach of contract, breach of warranty, guarantee or condition, strict liability, negligence, or other tort to the extent permitted by applicable law. It also applies even if Microsoft knew or should have known about the possibility of the damages. The above limitation or exclusion may not apply to you because your country may not allow the exclusion or limitation of incidental, consequential or other damages. ### Citation Information ``` @inproceedings{yang-etal-2015-wikiqa, title = "{W}iki{QA}: A Challenge Dataset for Open-Domain Question Answering", author = "Yang, Yi and Yih, Wen-tau and Meek, Christopher", booktitle = "Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing", month = sep, year = "2015", address = "Lisbon, Portugal", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/D15-1237", doi = "10.18653/v1/D15-1237", pages = "2013--2018", } ``` ### Contributions Thanks to [@patrickvonplaten](https://github.com/patrickvonplaten), [@mariamabarham](https://github.com/mariamabarham), [@lewtun](https://github.com/lewtun), [@thomwolf](https://github.com/thomwolf) for adding this dataset.
wiki_qa
[ "task_categories:question-answering", "task_ids:open-domain-qa", "annotations_creators:crowdsourced", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:other", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["found"], "language": ["en"], "license": ["other"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["question-answering"], "task_ids": ["open-domain-qa"], "paperswithcode_id": "wikiqa", "pretty_name": "WikiQA", "dataset_info": {"features": [{"name": "question_id", "dtype": "string"}, {"name": "question", "dtype": "string"}, {"name": "document_title", "dtype": "string"}, {"name": "answer", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "0", "1": "1"}}}}], "splits": [{"name": "test", "num_bytes": 1333261, "num_examples": 6165}, {"name": "validation", "num_bytes": 589765, "num_examples": 2733}, {"name": "train", "num_bytes": 4453862, "num_examples": 20360}], "download_size": 2861208, "dataset_size": 6376888}, "configs": [{"config_name": "default", "data_files": [{"split": "test", "path": "data/test-*"}, {"split": "validation", "path": "data/validation-*"}, {"split": "train", "path": "data/train-*"}]}]}
2024-01-04T16:41:46+00:00
[]
[ "en" ]
TAGS #task_categories-question-answering #task_ids-open-domain-qa #annotations_creators-crowdsourced #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-other #region-us
Dataset Card for "wiki\_qa" =========================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: WikiQA: A Challenge Dataset for Open-Domain Question Answering * Point of Contact: * Size of downloaded dataset files: 7.10 MB * Size of the generated dataset: 6.40 MB * Total amount of disk used: 13.50 MB ### Dataset Summary Wiki Question Answering corpus from Microsoft. The WikiQA corpus is a publicly available set of question and sentence pairs, collected and annotated for research on open-domain question answering. ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- ### Data Instances #### default * Size of downloaded dataset files: 7.10 MB * Size of the generated dataset: 6.40 MB * Total amount of disk used: 13.50 MB An example of 'train' looks as follows. ### Data Fields The data fields are the same among all splits. #### default * 'question\_id': a 'string' feature. * 'question': a 'string' feature. * 'document\_title': a 'string' feature. * 'answer': a 'string' feature. * 'label': a classification label, with possible values including '0' (0), '1' (1). ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information MICROSOFT RESEARCH DATA LICENSE AGREEMENT FOR MICROSOFT RESEARCH WIKIQA CORPUS These license terms are an agreement between Microsoft Corporation (or based on where you live, one of its affiliates) and you. Please read them. They apply to the data associated with this license above, which includes the media on which you received it, if any. The terms also apply to any Microsoft: * updates, * supplements, * Internet-based services, and * support services for this data, unless other terms accompany those items. If so, those terms apply. BY USING THE DATA, YOU ACCEPT THESE TERMS. IF YOU DO NOT ACCEPT THEM, DO NOT USE THE DATA. If you comply with these license terms, you have the rights below. 1. SCOPE OF LICENSE. a. You may use, copy, modify, create derivative works, and distribute the Dataset: i. for research and technology development purposes only. Examples of research and technology development uses are teaching, academic research, public demonstrations and experimentation ; and ii. to publish (or present papers or articles) on your results from using such Dataset. b. The data is licensed, not sold. This agreement only gives you some rights to use the data. Microsoft reserves all other rights. Unless applicable law gives you more rights despite this limitation, you may use the data only as expressly permitted in this agreement. In doing so, you must comply with any technical limitations in the data that only allow you to use it in certain ways. You may not * work around any technical limitations in the data; * reverse engineer, decompile or disassemble the data, except and only to the extent that applicable law expressly permits, despite this limitation; * rent, lease or lend the data; * transfer the data or this agreement to any third party; or * use the data directly in a commercial product without Microsoft’s permission. 2. DISTRIBUTION REQUIREMENTS: a. If you distribute the Dataset or any derivative works of the Dataset, you will distribute them under the same terms and conditions as in this Agreement, and you will not grant other rights to the Dataset or derivative works that are different from those provided by this Agreement. b. If you have created derivative works of the Dataset, and distribute such derivative works, you will cause the modified files to carry prominent notices so that recipients know that they are not receiving Page 1 of 3the original Dataset. Such notices must state: (i) that you have changed the Dataset; and (ii) the date of any changes. 3. DISTRIBUTION RESTRICTIONS. You may not: (a) alter any copyright, trademark or patent notice in the Dataset; (b) use Microsoft’s trademarks in a way that suggests your derivative works or modifications come from or are endorsed by Microsoft; (c) include the Dataset in malicious, deceptive or unlawful programs. 4. OWNERSHIP. Microsoft retains all right, title, and interest in and to any Dataset provided to you under this Agreement. You acquire no interest in the Dataset you may receive under the terms of this Agreement. 5. LICENSE TO MICROSOFT. Microsoft is granted back, without any restrictions or limitations, a non-exclusive, perpetual, irrevocable, royalty-free, assignable and sub-licensable license, to reproduce, publicly perform or display, use, modify, post, distribute, make and have made, sell and transfer your modifications to and/or derivative works of the Dataset, for any purpose. 6. FEEDBACK. If you give feedback about the Dataset to Microsoft, you give to Microsoft, without charge, the right to use, share and commercialize your feedback in any way and for any purpose. You also give to third parties, without charge, any patent rights needed for their products, technologies and services to use or interface with any specific parts of a Microsoft dataset or service that includes the feedback. You will not give feedback that is subject to a license that requires Microsoft to license its Dataset or documentation to third parties because we include your feedback in them. These rights survive this Agreement. 7. EXPORT RESTRICTIONS. The Dataset is subject to United States export laws and regulations. You must comply with all domestic and international export laws and regulations that apply to the Dataset. These laws include restrictions on destinations, end users and end use. For additional information, see URL 8. ENTIRE AGREEMENT. This Agreement, and the terms for supplements, updates, Internet-based services and support services that you use, are the entire agreement for the Dataset. 9. SUPPORT SERVICES. Because this data is “as is,” we may not provide support services for it. 10. APPLICABLE LAW. a. United States. If you acquired the software in the United States, Washington state law governs the interpretation of this agreement and applies to claims for breach of it, regardless of conflict of laws principles. The laws of the state where you live govern all other claims, including claims under state consumer protection laws, unfair competition laws, and in tort. b. Outside the United States. If you acquired the software in any other country, the laws of that country apply. 11. LEGAL EFFECT. This Agreement describes certain legal rights. You may have other rights under the laws of your country. You may also have rights with respect to the party from whom you acquired the Dataset. This Agreement does not change your rights under the laws of your country if the laws of your country do not permit it to do so. 12. DISCLAIMER OF WARRANTY. The Dataset is licensed “as-is.” You bear the risk of using it. Microsoft gives no express warranties, guarantees or conditions. You may have additional consumer rights or statutory guarantees under your local laws which this agreement cannot change. To the extent permitted under your local laws, Microsoft excludes the implied warranties of merchantability, fitness for a particular purpose and non- infringement. 13. LIMITATION ON AND EXCLUSION OF REMEDIES AND DAMAGES. YOU CAN RECOVER FROM MICROSOFT AND ITS SUPPLIERS ONLY DIRECT DAMAGES UP TO U.S. $5.00. YOU CANNOT RECOVER ANY OTHER DAMAGES, INCLUDING CONSEQUENTIAL, LOST PROFITS, SPECIAL, INDIRECT OR INCIDENTAL DAMAGES. This limitation applies to * anything related to the software, services, content (including code) on third party Internet sites, or third party programs; and Page 2 of 3 * claims for breach of contract, breach of warranty, guarantee or condition, strict liability, negligence, or other tort to the extent permitted by applicable law. It also applies even if Microsoft knew or should have known about the possibility of the damages. The above limitation or exclusion may not apply to you because your country may not allow the exclusion or limitation of incidental, consequential or other damages. ### Contributions Thanks to @patrickvonplaten, @mariamabarham, @lewtun, @thomwolf for adding this dataset.
[ "### Dataset Summary\n\n\nWiki Question Answering corpus from Microsoft.\n\n\nThe WikiQA corpus is a publicly available set of question and sentence pairs, collected and annotated for research on open-domain question answering.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### default\n\n\n* Size of downloaded dataset files: 7.10 MB\n* Size of the generated dataset: 6.40 MB\n* Total amount of disk used: 13.50 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### default\n\n\n* 'question\\_id': a 'string' feature.\n* 'question': a 'string' feature.\n* 'document\\_title': a 'string' feature.\n* 'answer': a 'string' feature.\n* 'label': a classification label, with possible values including '0' (0), '1' (1).", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nMICROSOFT RESEARCH DATA LICENSE AGREEMENT\nFOR\nMICROSOFT RESEARCH WIKIQA CORPUS\n\n\nThese license terms are an agreement between Microsoft Corporation (or based on where you live, one of its\naffiliates) and you. Please read them. They apply to the data associated with this license above, which includes\nthe media on which you received it, if any. The terms also apply to any Microsoft:\n\n\n* updates,\n* supplements,\n* Internet-based services, and\n* support services\nfor this data, unless other terms accompany those items. If so, those terms apply.\nBY USING THE DATA, YOU ACCEPT THESE TERMS. IF YOU DO NOT ACCEPT THEM, DO NOT USE THE DATA.\nIf you comply with these license terms, you have the rights below.\n\n\n1. SCOPE OF LICENSE.\na. You may use, copy, modify, create derivative works, and distribute the Dataset:\ni. for research and technology development purposes only. Examples of research and technology\ndevelopment uses are teaching, academic research, public demonstrations and experimentation ;\nand\nii. to publish (or present papers or articles) on your results from using such Dataset.\nb. The data is licensed, not sold. This agreement only gives you some rights to use the data. Microsoft reserves\nall other rights. Unless applicable law gives you more rights despite this limitation, you may use the data only\nas expressly permitted in this agreement. In doing so, you must comply with any technical limitations in the\ndata that only allow you to use it in certain ways.\nYou may not\n\n\n* work around any technical limitations in the data;\n* reverse engineer, decompile or disassemble the data, except and only to the extent that applicable law\nexpressly permits, despite this limitation;\n* rent, lease or lend the data;\n* transfer the data or this agreement to any third party; or\n* use the data directly in a commercial product without Microsoft’s permission.\n\n\n2. DISTRIBUTION REQUIREMENTS:\na. If you distribute the Dataset or any derivative works of the Dataset, you will distribute them under the\nsame terms and conditions as in this Agreement, and you will not grant other rights to the Dataset or\nderivative works that are different from those provided by this Agreement.\nb. If you have created derivative works of the Dataset, and distribute such derivative works, you will\ncause the modified files to carry prominent notices so that recipients know that they are not receiving\nPage 1 of 3the original Dataset. Such notices must state: (i) that you have changed the Dataset; and (ii) the date\nof any changes.\n3. DISTRIBUTION RESTRICTIONS. You may not: (a) alter any copyright, trademark or patent notice in the\nDataset; (b) use Microsoft’s trademarks in a way that suggests your derivative works or modifications come from\nor are endorsed by Microsoft; (c) include the Dataset in malicious, deceptive or unlawful programs.\n4. OWNERSHIP. Microsoft retains all right, title, and interest in and to any Dataset provided to you under this\nAgreement. You acquire no interest in the Dataset you may receive under the terms of this Agreement.\n5. LICENSE TO MICROSOFT. Microsoft is granted back, without any restrictions or limitations, a non-exclusive,\nperpetual, irrevocable, royalty-free, assignable and sub-licensable license, to reproduce, publicly perform or\ndisplay, use, modify, post, distribute, make and have made, sell and transfer your modifications to and/or\nderivative works of the Dataset, for any purpose.\n6. FEEDBACK. If you give feedback about the Dataset to Microsoft, you give to Microsoft, without charge, the right\nto use, share and commercialize your feedback in any way and for any purpose. You also give to third parties,\nwithout charge, any patent rights needed for their products, technologies and services to use or interface with\nany specific parts of a Microsoft dataset or service that includes the feedback. You will not give feedback that is\nsubject to a license that requires Microsoft to license its Dataset or documentation to third parties because we\ninclude your feedback in them. These rights survive this Agreement.\n7. EXPORT RESTRICTIONS. The Dataset is subject to United States export laws and regulations. You must\ncomply with all domestic and international export laws and regulations that apply to the Dataset. These laws\ninclude restrictions on destinations, end users and end use. For additional information, see\nURL\n8. ENTIRE AGREEMENT. This Agreement, and the terms for supplements, updates, Internet-based services and\nsupport services that you use, are the entire agreement for the Dataset.\n9. SUPPORT SERVICES. Because this data is “as is,” we may not provide support services for it.\n10. APPLICABLE LAW.\na. United States. If you acquired the software in the United States, Washington state law governs the\ninterpretation of this agreement and applies to claims for breach of it, regardless of conflict of laws principles.\nThe laws of the state where you live govern all other claims, including claims under state consumer protection\nlaws, unfair competition laws, and in tort.\nb. Outside the United States. If you acquired the software in any other country, the laws of that country\napply.\n11. LEGAL EFFECT. This Agreement describes certain legal rights. You may have other rights under the laws of your\ncountry. You may also have rights with respect to the party from whom you acquired the Dataset. This\nAgreement does not change your rights under the laws of your country if the laws of your country do not permit\nit to do so.\n12. DISCLAIMER OF WARRANTY. The Dataset is licensed “as-is.” You bear the risk of using it. Microsoft gives no\nexpress warranties, guarantees or conditions. You may have additional consumer rights or statutory guarantees\nunder your local laws which this agreement cannot change. To the extent permitted under your local laws,\nMicrosoft excludes the implied warranties of merchantability, fitness for a particular purpose and non-\ninfringement.\n13. LIMITATION ON AND EXCLUSION OF REMEDIES AND DAMAGES. YOU CAN RECOVER FROM\nMICROSOFT AND ITS SUPPLIERS ONLY DIRECT DAMAGES UP TO U.S. $5.00. YOU CANNOT RECOVER ANY\nOTHER DAMAGES, INCLUDING CONSEQUENTIAL, LOST PROFITS, SPECIAL, INDIRECT OR INCIDENTAL\nDAMAGES.\n\n\nThis limitation applies to\n\n\n* anything related to the software, services, content (including code) on third party Internet sites, or third party\nprograms; and Page 2 of 3\n* claims for breach of contract, breach of warranty, guarantee or condition, strict liability, negligence, or other\ntort to the extent permitted by applicable law.\n\n\nIt also applies even if Microsoft knew or should have known about the possibility of the damages. The above\nlimitation or exclusion may not apply to you because your country may not allow the exclusion or limitation of\nincidental, consequential or other damages.", "### Contributions\n\n\nThanks to @patrickvonplaten, @mariamabarham, @lewtun, @thomwolf for adding this dataset." ]
[ "TAGS\n#task_categories-question-answering #task_ids-open-domain-qa #annotations_creators-crowdsourced #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-other #region-us \n", "### Dataset Summary\n\n\nWiki Question Answering corpus from Microsoft.\n\n\nThe WikiQA corpus is a publicly available set of question and sentence pairs, collected and annotated for research on open-domain question answering.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### default\n\n\n* Size of downloaded dataset files: 7.10 MB\n* Size of the generated dataset: 6.40 MB\n* Total amount of disk used: 13.50 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### default\n\n\n* 'question\\_id': a 'string' feature.\n* 'question': a 'string' feature.\n* 'document\\_title': a 'string' feature.\n* 'answer': a 'string' feature.\n* 'label': a classification label, with possible values including '0' (0), '1' (1).", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nMICROSOFT RESEARCH DATA LICENSE AGREEMENT\nFOR\nMICROSOFT RESEARCH WIKIQA CORPUS\n\n\nThese license terms are an agreement between Microsoft Corporation (or based on where you live, one of its\naffiliates) and you. Please read them. They apply to the data associated with this license above, which includes\nthe media on which you received it, if any. The terms also apply to any Microsoft:\n\n\n* updates,\n* supplements,\n* Internet-based services, and\n* support services\nfor this data, unless other terms accompany those items. If so, those terms apply.\nBY USING THE DATA, YOU ACCEPT THESE TERMS. IF YOU DO NOT ACCEPT THEM, DO NOT USE THE DATA.\nIf you comply with these license terms, you have the rights below.\n\n\n1. SCOPE OF LICENSE.\na. You may use, copy, modify, create derivative works, and distribute the Dataset:\ni. for research and technology development purposes only. Examples of research and technology\ndevelopment uses are teaching, academic research, public demonstrations and experimentation ;\nand\nii. to publish (or present papers or articles) on your results from using such Dataset.\nb. The data is licensed, not sold. This agreement only gives you some rights to use the data. Microsoft reserves\nall other rights. Unless applicable law gives you more rights despite this limitation, you may use the data only\nas expressly permitted in this agreement. In doing so, you must comply with any technical limitations in the\ndata that only allow you to use it in certain ways.\nYou may not\n\n\n* work around any technical limitations in the data;\n* reverse engineer, decompile or disassemble the data, except and only to the extent that applicable law\nexpressly permits, despite this limitation;\n* rent, lease or lend the data;\n* transfer the data or this agreement to any third party; or\n* use the data directly in a commercial product without Microsoft’s permission.\n\n\n2. DISTRIBUTION REQUIREMENTS:\na. If you distribute the Dataset or any derivative works of the Dataset, you will distribute them under the\nsame terms and conditions as in this Agreement, and you will not grant other rights to the Dataset or\nderivative works that are different from those provided by this Agreement.\nb. If you have created derivative works of the Dataset, and distribute such derivative works, you will\ncause the modified files to carry prominent notices so that recipients know that they are not receiving\nPage 1 of 3the original Dataset. Such notices must state: (i) that you have changed the Dataset; and (ii) the date\nof any changes.\n3. DISTRIBUTION RESTRICTIONS. You may not: (a) alter any copyright, trademark or patent notice in the\nDataset; (b) use Microsoft’s trademarks in a way that suggests your derivative works or modifications come from\nor are endorsed by Microsoft; (c) include the Dataset in malicious, deceptive or unlawful programs.\n4. OWNERSHIP. Microsoft retains all right, title, and interest in and to any Dataset provided to you under this\nAgreement. You acquire no interest in the Dataset you may receive under the terms of this Agreement.\n5. LICENSE TO MICROSOFT. Microsoft is granted back, without any restrictions or limitations, a non-exclusive,\nperpetual, irrevocable, royalty-free, assignable and sub-licensable license, to reproduce, publicly perform or\ndisplay, use, modify, post, distribute, make and have made, sell and transfer your modifications to and/or\nderivative works of the Dataset, for any purpose.\n6. FEEDBACK. If you give feedback about the Dataset to Microsoft, you give to Microsoft, without charge, the right\nto use, share and commercialize your feedback in any way and for any purpose. You also give to third parties,\nwithout charge, any patent rights needed for their products, technologies and services to use or interface with\nany specific parts of a Microsoft dataset or service that includes the feedback. You will not give feedback that is\nsubject to a license that requires Microsoft to license its Dataset or documentation to third parties because we\ninclude your feedback in them. These rights survive this Agreement.\n7. EXPORT RESTRICTIONS. The Dataset is subject to United States export laws and regulations. You must\ncomply with all domestic and international export laws and regulations that apply to the Dataset. These laws\ninclude restrictions on destinations, end users and end use. For additional information, see\nURL\n8. ENTIRE AGREEMENT. This Agreement, and the terms for supplements, updates, Internet-based services and\nsupport services that you use, are the entire agreement for the Dataset.\n9. SUPPORT SERVICES. Because this data is “as is,” we may not provide support services for it.\n10. APPLICABLE LAW.\na. United States. If you acquired the software in the United States, Washington state law governs the\ninterpretation of this agreement and applies to claims for breach of it, regardless of conflict of laws principles.\nThe laws of the state where you live govern all other claims, including claims under state consumer protection\nlaws, unfair competition laws, and in tort.\nb. Outside the United States. If you acquired the software in any other country, the laws of that country\napply.\n11. LEGAL EFFECT. This Agreement describes certain legal rights. You may have other rights under the laws of your\ncountry. You may also have rights with respect to the party from whom you acquired the Dataset. This\nAgreement does not change your rights under the laws of your country if the laws of your country do not permit\nit to do so.\n12. DISCLAIMER OF WARRANTY. The Dataset is licensed “as-is.” You bear the risk of using it. Microsoft gives no\nexpress warranties, guarantees or conditions. You may have additional consumer rights or statutory guarantees\nunder your local laws which this agreement cannot change. To the extent permitted under your local laws,\nMicrosoft excludes the implied warranties of merchantability, fitness for a particular purpose and non-\ninfringement.\n13. LIMITATION ON AND EXCLUSION OF REMEDIES AND DAMAGES. YOU CAN RECOVER FROM\nMICROSOFT AND ITS SUPPLIERS ONLY DIRECT DAMAGES UP TO U.S. $5.00. YOU CANNOT RECOVER ANY\nOTHER DAMAGES, INCLUDING CONSEQUENTIAL, LOST PROFITS, SPECIAL, INDIRECT OR INCIDENTAL\nDAMAGES.\n\n\nThis limitation applies to\n\n\n* anything related to the software, services, content (including code) on third party Internet sites, or third party\nprograms; and Page 2 of 3\n* claims for breach of contract, breach of warranty, guarantee or condition, strict liability, negligence, or other\ntort to the extent permitted by applicable law.\n\n\nIt also applies even if Microsoft knew or should have known about the possibility of the damages. The above\nlimitation or exclusion may not apply to you because your country may not allow the exclusion or limitation of\nincidental, consequential or other damages.", "### Contributions\n\n\nThanks to @patrickvonplaten, @mariamabarham, @lewtun, @thomwolf for adding this dataset." ]
[ 88, 48, 10, 11, 6, 49, 17, 80, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 1566, 34 ]
[ "passage: TAGS\n#task_categories-question-answering #task_ids-open-domain-qa #annotations_creators-crowdsourced #language_creators-found #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-other #region-us \n### Dataset Summary\n\n\nWiki Question Answering corpus from Microsoft.\n\n\nThe WikiQA corpus is a publicly available set of question and sentence pairs, collected and annotated for research on open-domain question answering.### Supported Tasks and Leaderboards### Languages\n\n\nDataset Structure\n-----------------### Data Instances#### default\n\n\n* Size of downloaded dataset files: 7.10 MB\n* Size of the generated dataset: 6.40 MB\n* Total amount of disk used: 13.50 MB\n\n\nAn example of 'train' looks as follows.### Data Fields\n\n\nThe data fields are the same among all splits.#### default\n\n\n* 'question\\_id': a 'string' feature.\n* 'question': a 'string' feature.\n* 'document\\_title': a 'string' feature.\n* 'answer': a 'string' feature.\n* 'label': a classification label, with possible values including '0' (0), '1' (1).### Data Splits\n\n\n\nDataset Creation\n----------------### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------### Social Impact of Dataset### Discussion of Biases### Other Known Limitations\n\n\nAdditional Information\n----------------------### Dataset Curators" ]
a69a448e485b66d112c31d140dddc8a7a9194cdf
# Dataset Card for WikiQAar ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [WikiQaAr](https://github.com/qcri/WikiQAar) - **Repository:** [WikiQaAr](https://github.com/qcri/WikiQAar) - **Paper:** - **Point of Contact:** [Ines Abbes ](abbes.ines@yahoo.com) ### Dataset Summary Arabic Version of WikiQA by automatic automatic machine translators and crowdsourced the selection of the best one to be incorporated into the corpus ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The dataset is based on Arabic. ## Dataset Structure ### Data Instances Each data point contains the question and whether the answer is a valid or not. ### Data Fields - `question_id`: the question id. - `question`: the question text. - `document_id`: the wikipedia document id. - `answer_id` : the answer id. - `answer` : a candidate answer to the question. - `label` : 1 if the `answer` is correct or 0 otherwise. ### Data Splits The dataset is not split. | | train | validation | test | |------------|-------:|-----------:|-------:| | Data split | 70,264 | 20,632 | 10,387 | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization Translation of WikiQA. #### Who are the source language producers? Translation of WikiQA. ### Annotations The dataset does not contain any additional annotations. #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ``` @InProceedings{YangYihMeek:EMNLP2015:WikiQA, author = {{Yi}, Yang and {Wen-tau}, Yih and {Christopher} Meek}, title = "{WikiQA: A Challenge Dataset for Open-Domain Question Answering}", journal = {Association for Computational Linguistics}, year = 2015, doi = {10.18653/v1/D15-1237}, pages = {2013–2018}, } ``` ### Contributions Thanks to [@zaidalyafeai](https://github.com/zaidalyafeai) for adding this dataset.
wiki_qa_ar
[ "task_categories:question-answering", "task_ids:open-domain-qa", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:ar", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["ar"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["question-answering"], "task_ids": ["open-domain-qa"], "paperswithcode_id": "wikiqaar", "pretty_name": "English-Arabic Wikipedia Question-Answering", "dataset_info": {"features": [{"name": "question_id", "dtype": "string"}, {"name": "question", "dtype": "string"}, {"name": "document_id", "dtype": "string"}, {"name": "answer_id", "dtype": "string"}, {"name": "answer", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "0", "1": "1"}}}}], "config_name": "plain_text", "splits": [{"name": "test", "num_bytes": 7563127, "num_examples": 20632}, {"name": "validation", "num_bytes": 3740721, "num_examples": 10387}, {"name": "train", "num_bytes": 26009979, "num_examples": 70264}], "download_size": 35226436, "dataset_size": 37313827}}
2024-01-18T11:18:07+00:00
[]
[ "ar" ]
TAGS #task_categories-question-answering #task_ids-open-domain-qa #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Arabic #license-unknown #region-us
Dataset Card for WikiQAar ========================= Table of Contents ----------------- * Table of Contents * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: WikiQaAr * Repository: WikiQaAr * Paper: * Point of Contact: Ines Abbes ### Dataset Summary Arabic Version of WikiQA by automatic automatic machine translators and crowdsourced the selection of the best one to be incorporated into the corpus ### Supported Tasks and Leaderboards ### Languages The dataset is based on Arabic. Dataset Structure ----------------- ### Data Instances Each data point contains the question and whether the answer is a valid or not. ### Data Fields * 'question\_id': the question id. * 'question': the question text. * 'document\_id': the wikipedia document id. * 'answer\_id' : the answer id. * 'answer' : a candidate answer to the question. * 'label' : 1 if the 'answer' is correct or 0 otherwise. ### Data Splits The dataset is not split. Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization Translation of WikiQA. #### Who are the source language producers? Translation of WikiQA. ### Annotations The dataset does not contain any additional annotations. #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @zaidalyafeai for adding this dataset.
[ "### Dataset Summary\n\n\nArabic Version of WikiQA by automatic automatic machine translators\nand crowdsourced the selection of the best one to be incorporated into the corpus", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe dataset is based on Arabic.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nEach data point contains the question and whether the answer is a valid or not.", "### Data Fields\n\n\n* 'question\\_id': the question id.\n* 'question': the question text.\n* 'document\\_id': the wikipedia document id.\n* 'answer\\_id' : the answer id.\n* 'answer' : a candidate answer to the question.\n* 'label' : 1 if the 'answer' is correct or 0 otherwise.", "### Data Splits\n\n\nThe dataset is not split.\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nTranslation of WikiQA.", "#### Who are the source language producers?\n\n\nTranslation of WikiQA.", "### Annotations\n\n\nThe dataset does not contain any additional annotations.", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @zaidalyafeai for adding this dataset." ]
[ "TAGS\n#task_categories-question-answering #task_ids-open-domain-qa #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Arabic #license-unknown #region-us \n", "### Dataset Summary\n\n\nArabic Version of WikiQA by automatic automatic machine translators\nand crowdsourced the selection of the best one to be incorporated into the corpus", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe dataset is based on Arabic.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nEach data point contains the question and whether the answer is a valid or not.", "### Data Fields\n\n\n* 'question\\_id': the question id.\n* 'question': the question text.\n* 'document\\_id': the wikipedia document id.\n* 'answer\\_id' : the answer id.\n* 'answer' : a candidate answer to the question.\n* 'label' : 1 if the 'answer' is correct or 0 otherwise.", "### Data Splits\n\n\nThe dataset is not split.\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization\n\n\nTranslation of WikiQA.", "#### Who are the source language producers?\n\n\nTranslation of WikiQA.", "### Annotations\n\n\nThe dataset does not contain any additional annotations.", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @zaidalyafeai for adding this dataset." ]
[ 94, 35, 10, 19, 23, 87, 18, 7, 4, 15, 15, 17, 5, 9, 18, 7, 8, 14, 6, 6, 20 ]
[ "passage: TAGS\n#task_categories-question-answering #task_ids-open-domain-qa #annotations_creators-crowdsourced #language_creators-crowdsourced #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-Arabic #license-unknown #region-us \n### Dataset Summary\n\n\nArabic Version of WikiQA by automatic automatic machine translators\nand crowdsourced the selection of the best one to be incorporated into the corpus### Supported Tasks and Leaderboards### Languages\n\n\nThe dataset is based on Arabic.\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nEach data point contains the question and whether the answer is a valid or not.### Data Fields\n\n\n* 'question\\_id': the question id.\n* 'question': the question text.\n* 'document\\_id': the wikipedia document id.\n* 'answer\\_id' : the answer id.\n* 'answer' : a candidate answer to the question.\n* 'label' : 1 if the 'answer' is correct or 0 otherwise.### Data Splits\n\n\nThe dataset is not split.\n\n\n\nDataset Creation\n----------------### Curation Rationale### Source Data#### Initial Data Collection and Normalization\n\n\nTranslation of WikiQA.#### Who are the source language producers?\n\n\nTranslation of WikiQA.### Annotations\n\n\nThe dataset does not contain any additional annotations.#### Annotation process#### Who are the annotators?### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------### Social Impact of Dataset### Discussion of Biases### Other Known Limitations\n\n\nAdditional Information\n----------------------### Dataset Curators### Licensing Information### Contributions\n\n\nThanks to @zaidalyafeai for adding this dataset." ]
934c6c842af4dcc3a33c77a87d31c65c04080d2f
# Dataset Card for "wiki_snippets" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://dumps.wikimedia.org](https://dumps.wikimedia.org) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Dataset Summary Wikipedia version split into plain text snippets for dense semantic indexing. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure We show detailed information for 2 configurations of the dataset (with 100 snippet passage length and 0 overlap) in English: - wiki40b_en_100_0: Wiki-40B - wikipedia_en_100_0: Wikipedia ### Data Instances #### wiki40b_en_100_0 - **Size of downloaded dataset files:** 0.00 MB - **Size of the generated dataset:** 12.94 GB - **Total amount of disk used:** 12.94 GB An example of 'train' looks as follows: ``` {'_id': '{"datasets_id": 0, "wiki_id": "Q1294448", "sp": 2, "sc": 0, "ep": 6, "ec": 610}', 'datasets_id': 0, 'wiki_id': 'Q1294448', 'start_paragraph': 2, 'start_character': 0, 'end_paragraph': 6, 'end_character': 610, 'article_title': 'Ági Szalóki', 'section_title': 'Life', 'passage_text': "Ági Szalóki Life She started singing as a toddler, considering Márta Sebestyén a role model. Her musical background is traditional folk music; she first won recognition for singing with Ökrös in a traditional folk style, and Besh o droM, a Balkan gypsy brass band. With these ensembles she toured around the world from the Montreal Jazz Festival, through Glastonbury Festival to the Théatre de la Ville in Paris, from New York to Beijing.\nSince 2005, she began to pursue her solo career and explore various genres, such as jazz, thirties ballads, or children's songs.\nUntil now, three of her six released albums"} ``` #### wikipedia_en_100_0 - **Size of downloaded dataset files:** 0.00 MB - **Size of the generated dataset:** 26.41 GB - **Total amount of disk used:** 26.41 GB An example of 'train' looks as follows: ``` {'_id': '{"datasets_id": 0, "wiki_id": "Anarchism", "sp": 0, "sc": 0, "ep": 2, "ec": 129}', 'datasets_id': 0, 'wiki_id': 'Anarchism', 'start_paragraph': 0, 'start_character': 0, 'end_paragraph': 2, 'end_character': 129, 'article_title': 'Anarchism', 'section_title': 'Start', 'passage_text': 'Anarchism is a political philosophy and movement that is sceptical of authority and rejects all involuntary, coercive forms of hierarchy. Anarchism calls for the abolition of the state, which it holds to be unnecessary, undesirable, and harmful. As a historically left-wing movement, placed on the farthest left of the political spectrum, it is usually described alongside communalism and libertarian Marxism as the libertarian wing (libertarian socialism) of the socialist movement, and has a strong historical association with anti-capitalism and socialism. Humans lived in societies without formal hierarchies long before the establishment of formal states, realms, or empires. With the'} ``` ### Data Fields The data fields are the same for all configurations: - `_id`: a `string` feature. - `datasets_id`: a `int32` feature. - `wiki_id`: a `string` feature. - `start_paragraph`: a `int32` feature. - `start_character`: a `int32` feature. - `end_paragraph`: a `int32` feature. - `end_character`: a `int32` feature. - `article_title`: a `string` feature. - `section_title`: a `string` feature. - `passage_text`: a `string` feature. ### Data Splits | name | train | |:-------------------|---------:| | wiki40b_en_100_0 | 17553713 | | wikipedia_en_100_0 | 33849898 | ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information See licensing information of source datasets. ### Citation Information Cite source datasets: - Wiki-40B: ``` @inproceedings{49029, title = {Wiki-40B: Multilingual Language Model Dataset}, author = {Mandy Guo and Zihang Dai and Denny Vrandecic and Rami Al-Rfou}, year = {2020}, booktitle = {LREC 2020} } ``` - Wikipedia: ``` @ONLINE{wikidump, author = "Wikimedia Foundation", title = "Wikimedia Downloads", url = "https://dumps.wikimedia.org" } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@lhoestq](https://github.com/lhoestq), [@mariamabarham](https://github.com/mariamabarham), [@yjernite](https://github.com/yjernite) for adding this dataset.
wiki_snippets
[ "task_categories:text-generation", "task_categories:other", "task_ids:language-modeling", "annotations_creators:no-annotation", "language_creators:crowdsourced", "multilinguality:multilingual", "size_categories:10M<n<100M", "source_datasets:extended|wiki40b", "source_datasets:extended|wikipedia", "language:en", "license:unknown", "text-search", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["unknown"], "multilinguality": ["multilingual"], "size_categories": ["10M<n<100M"], "source_datasets": ["extended|wiki40b", "extended|wikipedia"], "task_categories": ["text-generation", "other"], "task_ids": ["language-modeling"], "pretty_name": "WikiSnippets", "tags": ["text-search"], "dataset_info": [{"config_name": "wiki40b_en_100_0", "features": [{"name": "_id", "dtype": "string"}, {"name": "datasets_id", "dtype": "int32"}, {"name": "wiki_id", "dtype": "string"}, {"name": "start_paragraph", "dtype": "int32"}, {"name": "start_character", "dtype": "int32"}, {"name": "end_paragraph", "dtype": "int32"}, {"name": "end_character", "dtype": "int32"}, {"name": "article_title", "dtype": "string"}, {"name": "section_title", "dtype": "string"}, {"name": "passage_text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 12938641686, "num_examples": 17553713}], "download_size": 0, "dataset_size": 12938641686}, {"config_name": "wikipedia_en_100_0", "features": [{"name": "_id", "dtype": "string"}, {"name": "datasets_id", "dtype": "int32"}, {"name": "wiki_id", "dtype": "string"}, {"name": "start_paragraph", "dtype": "int32"}, {"name": "start_character", "dtype": "int32"}, {"name": "end_paragraph", "dtype": "int32"}, {"name": "end_character", "dtype": "int32"}, {"name": "article_title", "dtype": "string"}, {"name": "section_title", "dtype": "string"}, {"name": "passage_text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 26407884393, "num_examples": 33849898}], "download_size": 0, "dataset_size": 26407884393}]}
2024-01-18T11:18:09+00:00
[]
[ "en" ]
TAGS #task_categories-text-generation #task_categories-other #task_ids-language-modeling #annotations_creators-no-annotation #language_creators-crowdsourced #multilinguality-multilingual #size_categories-10M<n<100M #source_datasets-extended|wiki40b #source_datasets-extended|wikipedia #language-English #license-unknown #text-search #region-us
Dataset Card for "wiki\_snippets" ================================= Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: * Point of Contact: ### Dataset Summary Wikipedia version split into plain text snippets for dense semantic indexing. ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- We show detailed information for 2 configurations of the dataset (with 100 snippet passage length and 0 overlap) in English: * wiki40b\_en\_100\_0: Wiki-40B * wikipedia\_en\_100\_0: Wikipedia ### Data Instances #### wiki40b\_en\_100\_0 * Size of downloaded dataset files: 0.00 MB * Size of the generated dataset: 12.94 GB * Total amount of disk used: 12.94 GB An example of 'train' looks as follows: #### wikipedia\_en\_100\_0 * Size of downloaded dataset files: 0.00 MB * Size of the generated dataset: 26.41 GB * Total amount of disk used: 26.41 GB An example of 'train' looks as follows: ### Data Fields The data fields are the same for all configurations: * '\_id': a 'string' feature. * 'datasets\_id': a 'int32' feature. * 'wiki\_id': a 'string' feature. * 'start\_paragraph': a 'int32' feature. * 'start\_character': a 'int32' feature. * 'end\_paragraph': a 'int32' feature. * 'end\_character': a 'int32' feature. * 'article\_title': a 'string' feature. * 'section\_title': a 'string' feature. * 'passage\_text': a 'string' feature. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information See licensing information of source datasets. Cite source datasets: * Wiki-40B: * Wikipedia: ### Contributions Thanks to @thomwolf, @lhoestq, @mariamabarham, @yjernite for adding this dataset.
[ "### Dataset Summary\n\n\nWikipedia version split into plain text snippets for dense semantic indexing.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------\n\n\nWe show detailed information for 2 configurations of the dataset (with 100 snippet passage length and 0 overlap) in\nEnglish:\n\n\n* wiki40b\\_en\\_100\\_0: Wiki-40B\n* wikipedia\\_en\\_100\\_0: Wikipedia", "### Data Instances", "#### wiki40b\\_en\\_100\\_0\n\n\n* Size of downloaded dataset files: 0.00 MB\n* Size of the generated dataset: 12.94 GB\n* Total amount of disk used: 12.94 GB\n\n\nAn example of 'train' looks as follows:", "#### wikipedia\\_en\\_100\\_0\n\n\n* Size of downloaded dataset files: 0.00 MB\n* Size of the generated dataset: 26.41 GB\n* Total amount of disk used: 26.41 GB\n\n\nAn example of 'train' looks as follows:", "### Data Fields\n\n\nThe data fields are the same for all configurations:\n\n\n* '\\_id': a 'string' feature.\n* 'datasets\\_id': a 'int32' feature.\n* 'wiki\\_id': a 'string' feature.\n* 'start\\_paragraph': a 'int32' feature.\n* 'start\\_character': a 'int32' feature.\n* 'end\\_paragraph': a 'int32' feature.\n* 'end\\_character': a 'int32' feature.\n* 'article\\_title': a 'string' feature.\n* 'section\\_title': a 'string' feature.\n* 'passage\\_text': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nSee licensing information of source datasets.\n\n\nCite source datasets:\n\n\n* Wiki-40B:\n* Wikipedia:", "### Contributions\n\n\nThanks to @thomwolf, @lhoestq, @mariamabarham, @yjernite for adding this dataset." ]
[ "TAGS\n#task_categories-text-generation #task_categories-other #task_ids-language-modeling #annotations_creators-no-annotation #language_creators-crowdsourced #multilinguality-multilingual #size_categories-10M<n<100M #source_datasets-extended|wiki40b #source_datasets-extended|wikipedia #language-English #license-unknown #text-search #region-us \n", "### Dataset Summary\n\n\nWikipedia version split into plain text snippets for dense semantic indexing.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------\n\n\nWe show detailed information for 2 configurations of the dataset (with 100 snippet passage length and 0 overlap) in\nEnglish:\n\n\n* wiki40b\\_en\\_100\\_0: Wiki-40B\n* wikipedia\\_en\\_100\\_0: Wikipedia", "### Data Instances", "#### wiki40b\\_en\\_100\\_0\n\n\n* Size of downloaded dataset files: 0.00 MB\n* Size of the generated dataset: 12.94 GB\n* Total amount of disk used: 12.94 GB\n\n\nAn example of 'train' looks as follows:", "#### wikipedia\\_en\\_100\\_0\n\n\n* Size of downloaded dataset files: 0.00 MB\n* Size of the generated dataset: 26.41 GB\n* Total amount of disk used: 26.41 GB\n\n\nAn example of 'train' looks as follows:", "### Data Fields\n\n\nThe data fields are the same for all configurations:\n\n\n* '\\_id': a 'string' feature.\n* 'datasets\\_id': a 'int32' feature.\n* 'wiki\\_id': a 'string' feature.\n* 'start\\_paragraph': a 'int32' feature.\n* 'start\\_character': a 'int32' feature.\n* 'end\\_paragraph': a 'int32' feature.\n* 'end\\_character': a 'int32' feature.\n* 'article\\_title': a 'string' feature.\n* 'section\\_title': a 'string' feature.\n* 'passage\\_text': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nSee licensing information of source datasets.\n\n\nCite source datasets:\n\n\n* Wiki-40B:\n* Wikipedia:", "### Contributions\n\n\nThanks to @thomwolf, @lhoestq, @mariamabarham, @yjernite for adding this dataset." ]
[ 120, 24, 10, 70, 6, 60, 59, 171, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 32, 33 ]
[ "passage: TAGS\n#task_categories-text-generation #task_categories-other #task_ids-language-modeling #annotations_creators-no-annotation #language_creators-crowdsourced #multilinguality-multilingual #size_categories-10M<n<100M #source_datasets-extended|wiki40b #source_datasets-extended|wikipedia #language-English #license-unknown #text-search #region-us \n### Dataset Summary\n\n\nWikipedia version split into plain text snippets for dense semantic indexing.### Supported Tasks and Leaderboards### Languages\n\n\nDataset Structure\n-----------------\n\n\nWe show detailed information for 2 configurations of the dataset (with 100 snippet passage length and 0 overlap) in\nEnglish:\n\n\n* wiki40b\\_en\\_100\\_0: Wiki-40B\n* wikipedia\\_en\\_100\\_0: Wikipedia### Data Instances#### wiki40b\\_en\\_100\\_0\n\n\n* Size of downloaded dataset files: 0.00 MB\n* Size of the generated dataset: 12.94 GB\n* Total amount of disk used: 12.94 GB\n\n\nAn example of 'train' looks as follows:#### wikipedia\\_en\\_100\\_0\n\n\n* Size of downloaded dataset files: 0.00 MB\n* Size of the generated dataset: 26.41 GB\n* Total amount of disk used: 26.41 GB\n\n\nAn example of 'train' looks as follows:" ]
f98f194f7818e500b2b8438737870488398708fd
# Dataset Card for WikiSource ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** http://opus.nlpl.eu/WikiSource.php - **Repository:** None - **Paper:** http://www.lrec-conf.org/proceedings/lrec2012/pdf/463_Paper.pdf - **Leaderboard:** [More Information Needed] - **Point of Contact:** [More Information Needed] ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data [More Information Needed] #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations [More Information Needed] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@abhishekkrthakur](https://github.com/abhishekkrthakur) for adding this dataset.
wiki_source
[ "task_categories:translation", "annotations_creators:found", "language_creators:found", "multilinguality:multilingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "language:sv", "license:unknown", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["en", "sv"], "license": ["unknown"], "multilinguality": ["multilingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["translation"], "task_ids": [], "pretty_name": "WikiSource", "dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "translation", "dtype": {"translation": {"languages": ["en", "sv"]}}}], "config_name": "en-sv", "splits": [{"name": "train", "num_bytes": 8153542, "num_examples": 33283}], "download_size": 2375052, "dataset_size": 8153542}}
2024-01-18T11:18:10+00:00
[]
[ "en", "sv" ]
TAGS #task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-10K<n<100K #source_datasets-original #language-English #language-Swedish #license-unknown #region-us
# Dataset Card for WikiSource ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: None - Paper: URL - Leaderboard: - Point of Contact: ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information ### Contributions Thanks to @abhishekkrthakur for adding this dataset.
[ "# Dataset Card for WikiSource", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: URL\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ "TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-10K<n<100K #source_datasets-original #language-English #language-Swedish #license-unknown #region-us \n", "# Dataset Card for WikiSource", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: URL\n- Leaderboard: \n- Point of Contact:", "### Dataset Summary", "### Supported Tasks and Leaderboards", "### Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information", "### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
[ 78, 7, 120, 28, 6, 10, 4, 6, 6, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 6, 20 ]
[ "passage: TAGS\n#task_categories-translation #annotations_creators-found #language_creators-found #multilinguality-multilingual #size_categories-10K<n<100K #source_datasets-original #language-English #language-Swedish #license-unknown #region-us \n# Dataset Card for WikiSource## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: None\n- Paper: URL\n- Leaderboard: \n- Point of Contact:### Dataset Summary### Supported Tasks and Leaderboards### Languages## Dataset Structure### Data Instances### Data Fields### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators### Licensing Information### Contributions\n\nThanks to @abhishekkrthakur for adding this dataset." ]
7f022326bb901cc2e32ec659cb9a9022140e0b01
# Dataset Card for "wiki_split" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://dataset-homepage/](https://dataset-homepage/) - **Repository:** https://github.com/google-research-datasets/wiki-split - **Paper:** [Learning To Split and Rephrase From Wikipedia Edit History](https://arxiv.org/abs/1808.09468) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 100.28 MB - **Size of the generated dataset:** 388.40 MB - **Total amount of disk used:** 488.68 MB ### Dataset Summary One million English sentences, each split into two sentences that together preserve the original meaning, extracted from Wikipedia Google's WikiSplit dataset was constructed automatically from the publicly available Wikipedia revision history. Although the dataset contains some inherent noise, it can serve as valuable training data for models that split or merge sentences. ### Supported Tasks and Leaderboards - Split and Rephrase ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### default - **Size of downloaded dataset files:** 100.28 MB - **Size of the generated dataset:** 388.40 MB - **Total amount of disk used:** 488.68 MB An example of 'train' looks as follows. ``` { "complex_sentence": " '' As she translates from one language to another , she tries to find the appropriate wording and context in English that would correspond to the work in Spanish her poems and stories started to have differing meanings in their respective languages .", "simple_sentence_1": "' '' As she translates from one language to another , she tries to find the appropriate wording and context in English that would correspond to the work in Spanish . ", "simple_sentence_2": " Ergo , her poems and stories started to have differing meanings in their respective languages ." } ``` ### Data Fields The data fields are the same among all splits. #### default - `complex_sentence`: a `string` feature. - `simple_sentence_1`: a `string` feature. - `simple_sentence_2`: a `string` feature. ### Data Splits | name |train |validation|test| |-------|-----:|---------:|---:| |default|989944| 5000|5000| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information The WikiSplit dataset is a verbatim copy of certain content from the publicly available Wikipedia revision history. The dataset is therefore licensed under [CC BY-SA 4.0](http://creativecommons.org/licenses/by-sa/4.0/). Any third party content or data is provided "As Is" without any warranty, express or implied. ### Citation Information ``` @inproceedings{botha-etal-2018-learning, title = "Learning To Split and Rephrase From {W}ikipedia Edit History", author = "Botha, Jan A. and Faruqui, Manaal and Alex, John and Baldridge, Jason and Das, Dipanjan", booktitle = "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing", month = oct # "-" # nov, year = "2018", address = "Brussels, Belgium", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/D18-1080", doi = "10.18653/v1/D18-1080", pages = "732--737", } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@patrickvonplaten](https://github.com/patrickvonplaten), [@albertvillanova](https://github.com/albertvillanova), [@lewtun](https://github.com/lewtun) for adding this dataset.
wiki_split
[ "task_categories:text2text-generation", "annotations_creators:machine-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:en", "license:cc-by-4.0", "split-and-rephrase", "arxiv:1808.09468", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["machine-generated"], "language_creators": ["found"], "language": ["en"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["text2text-generation"], "task_ids": [], "paperswithcode_id": "wikisplit", "pretty_name": "WikiSplit", "tags": ["split-and-rephrase"], "dataset_info": {"features": [{"name": "complex_sentence", "dtype": "string"}, {"name": "simple_sentence_1", "dtype": "string"}, {"name": "simple_sentence_2", "dtype": "string"}], "splits": [{"name": "test", "num_bytes": 1949294, "num_examples": 5000}, {"name": "train", "num_bytes": 384513073, "num_examples": 989944}, {"name": "validation", "num_bytes": 1935459, "num_examples": 5000}], "download_size": 100279164, "dataset_size": 388397826}}
2024-01-18T11:18:11+00:00
[ "1808.09468" ]
[ "en" ]
TAGS #task_categories-text2text-generation #annotations_creators-machine-generated #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-4.0 #split-and-rephrase #arxiv-1808.09468 #region-us
Dataset Card for "wiki\_split" ============================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: https://dataset-homepage/ * Repository: URL * Paper: Learning To Split and Rephrase From Wikipedia Edit History * Point of Contact: * Size of downloaded dataset files: 100.28 MB * Size of the generated dataset: 388.40 MB * Total amount of disk used: 488.68 MB ### Dataset Summary One million English sentences, each split into two sentences that together preserve the original meaning, extracted from Wikipedia Google's WikiSplit dataset was constructed automatically from the publicly available Wikipedia revision history. Although the dataset contains some inherent noise, it can serve as valuable training data for models that split or merge sentences. ### Supported Tasks and Leaderboards * Split and Rephrase ### Languages Dataset Structure ----------------- ### Data Instances #### default * Size of downloaded dataset files: 100.28 MB * Size of the generated dataset: 388.40 MB * Total amount of disk used: 488.68 MB An example of 'train' looks as follows. ### Data Fields The data fields are the same among all splits. #### default * 'complex\_sentence': a 'string' feature. * 'simple\_sentence\_1': a 'string' feature. * 'simple\_sentence\_2': a 'string' feature. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information The WikiSplit dataset is a verbatim copy of certain content from the publicly available Wikipedia revision history. The dataset is therefore licensed under CC BY-SA 4.0. Any third party content or data is provided "As Is" without any warranty, express or implied. ### Contributions Thanks to @thomwolf, @patrickvonplaten, @albertvillanova, @lewtun for adding this dataset.
[ "### Dataset Summary\n\n\nOne million English sentences, each split into two sentences that together preserve the original meaning, extracted from Wikipedia\nGoogle's WikiSplit dataset was constructed automatically from the publicly available Wikipedia revision history. Although\nthe dataset contains some inherent noise, it can serve as valuable training data for models that split or merge sentences.", "### Supported Tasks and Leaderboards\n\n\n* Split and Rephrase", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### default\n\n\n* Size of downloaded dataset files: 100.28 MB\n* Size of the generated dataset: 388.40 MB\n* Total amount of disk used: 488.68 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### default\n\n\n* 'complex\\_sentence': a 'string' feature.\n* 'simple\\_sentence\\_1': a 'string' feature.\n* 'simple\\_sentence\\_2': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nThe WikiSplit dataset is a verbatim copy of certain content from the publicly available Wikipedia revision history.\nThe dataset is therefore licensed under CC BY-SA 4.0.\nAny third party content or data is provided \"As Is\" without any warranty, express or implied.", "### Contributions\n\n\nThanks to @thomwolf, @patrickvonplaten, @albertvillanova, @lewtun for adding this dataset." ]
[ "TAGS\n#task_categories-text2text-generation #annotations_creators-machine-generated #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-4.0 #split-and-rephrase #arxiv-1808.09468 #region-us \n", "### Dataset Summary\n\n\nOne million English sentences, each split into two sentences that together preserve the original meaning, extracted from Wikipedia\nGoogle's WikiSplit dataset was constructed automatically from the publicly available Wikipedia revision history. Although\nthe dataset contains some inherent noise, it can serve as valuable training data for models that split or merge sentences.", "### Supported Tasks and Leaderboards\n\n\n* Split and Rephrase", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### default\n\n\n* Size of downloaded dataset files: 100.28 MB\n* Size of the generated dataset: 388.40 MB\n* Total amount of disk used: 488.68 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### default\n\n\n* 'complex\\_sentence': a 'string' feature.\n* 'simple\\_sentence\\_1': a 'string' feature.\n* 'simple\\_sentence\\_2': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nThe WikiSplit dataset is a verbatim copy of certain content from the publicly available Wikipedia revision history.\nThe dataset is therefore licensed under CC BY-SA 4.0.\nAny third party content or data is provided \"As Is\" without any warranty, express or implied.", "### Contributions\n\n\nThanks to @thomwolf, @patrickvonplaten, @albertvillanova, @lewtun for adding this dataset." ]
[ 98, 78, 16, 11, 6, 54, 17, 54, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 65, 34 ]
[ "passage: TAGS\n#task_categories-text2text-generation #annotations_creators-machine-generated #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #source_datasets-original #language-English #license-cc-by-4.0 #split-and-rephrase #arxiv-1808.09468 #region-us \n### Dataset Summary\n\n\nOne million English sentences, each split into two sentences that together preserve the original meaning, extracted from Wikipedia\nGoogle's WikiSplit dataset was constructed automatically from the publicly available Wikipedia revision history. Although\nthe dataset contains some inherent noise, it can serve as valuable training data for models that split or merge sentences.### Supported Tasks and Leaderboards\n\n\n* Split and Rephrase### Languages\n\n\nDataset Structure\n-----------------### Data Instances#### default\n\n\n* Size of downloaded dataset files: 100.28 MB\n* Size of the generated dataset: 388.40 MB\n* Total amount of disk used: 488.68 MB\n\n\nAn example of 'train' looks as follows.### Data Fields\n\n\nThe data fields are the same among all splits.#### default\n\n\n* 'complex\\_sentence': a 'string' feature.\n* 'simple\\_sentence\\_1': a 'string' feature.\n* 'simple\\_sentence\\_2': a 'string' feature.### Data Splits\n\n\n\nDataset Creation\n----------------### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------### Social Impact of Dataset### Discussion of Biases### Other Known Limitations\n\n\nAdditional Information\n----------------------### Dataset Curators" ]
f0322a2ce7ab2fa84e3dd17e43c9863bf0fc9b47
# Dataset Card for [Needs More Information] ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://github.com/m3hrdadfi/wiki-summary - **Repository:** https://github.com/m3hrdadfi/wiki-summary - **Paper:** [More Information Needed] - **Leaderboard:** [More Information Needed] - **Point of Contact:** [Mehrdad Farahani](mailto:m3hrdadphi@gmail.com) ### Dataset Summary The dataset extracted from Persian Wikipedia into the form of articles and highlights and cleaned the dataset into pairs of articles and highlights and reduced the articles' length (only version 1.0.0) and highlights' length to a maximum of 512 and 128, respectively, suitable for parsBERT. This dataset is created to achieve state-of-the-art results on some interesting NLP tasks like Text Summarization. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The text in the dataset is in Percy. ## Dataset Structure ### Data Instances ``` { 'id' :'0598cfd2ac491a928615945054ab7602034a8f4f', 'link': 'https://fa.wikipedia.org/wiki/انقلاب_1917_روسیه', 'title': 'انقلاب 1917 روسیه', 'article': 'نخست انقلاب فوریه ۱۹۱۷ رخ داد . در این انقلاب پس از یک‌سری اعتصابات ، تظاهرات و درگیری‌ها ، نیکولای دوم ، آخرین تزار روسیه از سلطنت خلع شد و یک دولت موقت به قدرت رسید . دولت موقت زیر نظر گئورگی لووف و الکساندر کرنسکی تشکیل شد . اکثر اعضای دولت موقت ، از شاخه منشویک حزب سوسیال دموکرات کارگری روسیه بودند . دومین مرحله ، انقلاب اکتبر ۱۹۱۷ بود . انقلاب اکتبر ، تحت نظارت حزب بلشویک (شاخه رادیکال از حزب سوسیال دموکرات کارگری روسیه) و به رهبری ولادیمیر لنین به پیش رفت و طی یک یورش نظامی همه‌جانبه به کاخ زمستانی سن پترزبورگ و سایر اماکن مهم ، قدرت را از دولت موقت گرفت . در این انقلاب افراد بسیار کمی کشته شدند . از زمان شکست روسیه در جنگ ۱۹۰۵ با ژاپن ، اوضاع بد اقتصادی ، گرسنگی ، عقب‌ماندگی و سرمایه‌داری و نارضایتی‌های گوناگون در بین مردم ، سربازان ، کارگران ، کشاورزان و نخبگان روسیه به‌وجود آمده‌بود . سرکوبهای تزار و ایجاد مجلس دوما نظام مشروطه حاصل آن دوران است . حزب سوسیال دموکرات ، اصلی‌ترین معترض به سیاست‌های نیکلای دوم بود که به‌طور گسترده بین دهقانان کشاورزان و کارگران کارخانجات صنعتی علیه سیاست‌های سیستم تزار فعالیت داشت . در اوت ۱۹۱۴ میلادی ، امپراتوری روسیه به دستور تزار وقت و به منظور حمایت از اسلاوهای صربستان وارد جنگ جهانی اول در برابر امپراتوری آلمان و امپراتوری اتریش-مجارستان شد . نخست فقط بلشویک‌ها ، مخالف ورود روسیه به این جنگ بودند و می‌گفتند که این جنگ ، سبب بدتر شدن اوضاع نابسامان اقتصادی و اجتماعی روسیه خواهد شد . در سال ۱۹۱۴ میلادی ، یعنی در آغاز جنگ جهانی اول ، روسیه بزرگترین ارتش جهان را داشت ، حدود ۱۲ میلیون سرباز و ۶ میلیون سرباز ذخیره ؛ ولی در پایان سال ۱۹۱۶ میلادی ، پنج میلیون نفر از سربازان روسیه کشته ، زخمی یا اسیر شده بودند . حدود دو میلیون سرباز نیز محل خدمت خود را ترک کرده و غالبا با اسلحه به شهر و دیار خود بازگشته بودند . در میان ۱۰ یا ۱۱ میلیون سرباز باقی‌مانده نیز ، اعتبار تزار و سلسله مراتب ارتش و اتوریته افسران بالا دست از بین رفته بود . عوامل نابسامان داخلی اعم از اجتماعی کشاورزی و فرماندهی نظامی در شکستهای روسیه بسیار مؤثر بود . شکست‌های روسیه در جنگ جهانی اول ، حامیان نیکلای دوم در روسیه را به حداقل خود رساند . در اوایل فوریه ۱۹۱۷ میلادی اکثر کارگران صنعتی در پتروگراد و مسکو دست به اعتصاب زدند . سپس شورش به پادگان‌ها و سربازان رسید . اعتراضات دهقانان نیز گسترش یافت . سوسیال دموکرات‌ها هدایت اعتراضات را در دست گرفتند . در ۱۱ مارس ۱۹۱۷ میلادی ، تزار وقت روسیه ، نیکلای دوم ، فرمان انحلال مجلس روسیه را صادر کرد ، اما اکثر نمایندگان مجلس متفرق نشدند و با تصمیمات نیکلای دوم مخالفت کردند . سرانجام در پی تظاهرات گسترده کارگران و سپس نافرمانی سربازان در سرکوب تظاهرکنندگان در پتروگراد ، نیکلای دوم از مقام خود استعفا داد . بدین ترتیب حکم‌رانی دودمان رومانوف‌ها بر روسیه پس از حدود سیصد سال پایان یافت .', 'highlights': 'انقلاب ۱۹۱۷ روسیه ، جنبشی اعتراضی ، ضد امپراتوری روسیه بود که در سال ۱۹۱۷ رخ داد و به سرنگونی حکومت تزارها و برپایی اتحاد جماهیر شوروی انجامید . مبانی انقلاب بر پایه صلح-نان-زمین استوار بود . این انقلاب در دو مرحله صورت گرفت : در طول این انقلاب در شهرهای اصلی روسیه همانند مسکو و سن پترزبورگ رویدادهای تاریخی برجسته‌ای رخ داد . انقلاب در مناطق روستایی و رعیتی نیز پا به پای مناطق شهری در حال پیشروی بود و دهقانان زمین‌ها را تصرف کرده و در حال بازتوزیع آن در میان خود بودند .' } ``` ### Data Fields - `id`: Article id - `link`: Article link - `title`: Title of the article - `article`: Full text content in the article - `highlights`: Summary of the article ### Data Splits | Train | Test | Validation | |-------------|-------------|-------------| | 45,654 | 5,638 | 5,074 | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process No annotations. #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators The dataset was created by Mehrdad Farahani. ### Licensing Information [Apache License 2.0](https://github.com/m3hrdadfi/wiki-summary/blob/master/LICENSE) ### Citation Information ``` @misc{Bert2BertWikiSummaryPersian, author = {Mehrdad Farahani}, title = {Summarization using Bert2Bert model on WikiSummary dataset}, year = {2020}, publisher = {GitHub}, journal = {GitHub repository}, howpublished = {https://github.com/m3hrdadfi/wiki-summary}, } ``` ### Contributions Thanks to [@tanmoyio](https://github.com/tanmoyio) for adding this dataset.
wiki_summary
[ "task_categories:text2text-generation", "task_categories:translation", "task_categories:question-answering", "task_categories:summarization", "task_ids:abstractive-qa", "task_ids:explanation-generation", "task_ids:extractive-qa", "task_ids:open-domain-qa", "task_ids:open-domain-abstractive-qa", "task_ids:text-simplification", "annotations_creators:no-annotation", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:fa", "license:apache-2.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["crowdsourced"], "language": ["fa"], "license": ["apache-2.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text2text-generation", "translation", "question-answering", "summarization"], "task_ids": ["abstractive-qa", "explanation-generation", "extractive-qa", "open-domain-qa", "open-domain-abstractive-qa", "text-simplification"], "pretty_name": "WikiSummary", "dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "link", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "article", "dtype": "string"}, {"name": "highlights", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 207186608, "num_examples": 45654}, {"name": "test", "num_bytes": 25693509, "num_examples": 5638}, {"name": "validation", "num_bytes": 23130954, "num_examples": 5074}], "download_size": 255168504, "dataset_size": 256011071}}
2024-01-18T11:18:12+00:00
[]
[ "fa" ]
TAGS #task_categories-text2text-generation #task_categories-translation #task_categories-question-answering #task_categories-summarization #task_ids-abstractive-qa #task_ids-explanation-generation #task_ids-extractive-qa #task_ids-open-domain-qa #task_ids-open-domain-abstractive-qa #task_ids-text-simplification #annotations_creators-no-annotation #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Persian #license-apache-2.0 #region-us
Dataset Card for ================ Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: URL * Paper: * Leaderboard: * Point of Contact: Mehrdad Farahani ### Dataset Summary The dataset extracted from Persian Wikipedia into the form of articles and highlights and cleaned the dataset into pairs of articles and highlights and reduced the articles' length (only version 1.0.0) and highlights' length to a maximum of 512 and 128, respectively, suitable for parsBERT. This dataset is created to achieve state-of-the-art results on some interesting NLP tasks like Text Summarization. ### Supported Tasks and Leaderboards ### Languages The text in the dataset is in Percy. Dataset Structure ----------------- ### Data Instances ### Data Fields * 'id': Article id * 'link': Article link * 'title': Title of the article * 'article': Full text content in the article * 'highlights': Summary of the article ### Data Splits Train: 45,654, Test: 5,638, Validation: 5,074 Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process No annotations. #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators The dataset was created by Mehrdad Farahani. ### Licensing Information Apache License 2.0 ### Contributions Thanks to @tanmoyio for adding this dataset.
[ "### Dataset Summary\n\n\nThe dataset extracted from Persian Wikipedia into the form of articles and highlights and cleaned the dataset into pairs of articles and highlights and reduced the articles' length (only version 1.0.0) and highlights' length to a maximum of 512 and 128, respectively, suitable for parsBERT. This dataset is created to achieve state-of-the-art results on some interesting NLP tasks like Text Summarization.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe text in the dataset is in Percy.\n\n\nDataset Structure\n-----------------", "### Data Instances", "### Data Fields\n\n\n* 'id': Article id\n* 'link': Article link\n* 'title': Title of the article\n* 'article': Full text content in the article\n* 'highlights': Summary of the article", "### Data Splits\n\n\nTrain: 45,654, Test: 5,638, Validation: 5,074\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process\n\n\nNo annotations.", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThe dataset was created by Mehrdad Farahani.", "### Licensing Information\n\n\nApache License 2.0", "### Contributions\n\n\nThanks to @tanmoyio for adding this dataset." ]
[ "TAGS\n#task_categories-text2text-generation #task_categories-translation #task_categories-question-answering #task_categories-summarization #task_ids-abstractive-qa #task_ids-explanation-generation #task_ids-extractive-qa #task_ids-open-domain-qa #task_ids-open-domain-abstractive-qa #task_ids-text-simplification #annotations_creators-no-annotation #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Persian #license-apache-2.0 #region-us \n", "### Dataset Summary\n\n\nThe dataset extracted from Persian Wikipedia into the form of articles and highlights and cleaned the dataset into pairs of articles and highlights and reduced the articles' length (only version 1.0.0) and highlights' length to a maximum of 512 and 128, respectively, suitable for parsBERT. This dataset is created to achieve state-of-the-art results on some interesting NLP tasks like Text Summarization.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nThe text in the dataset is in Percy.\n\n\nDataset Structure\n-----------------", "### Data Instances", "### Data Fields\n\n\n* 'id': Article id\n* 'link': Article link\n* 'title': Title of the article\n* 'article': Full text content in the article\n* 'highlights': Summary of the article", "### Data Splits\n\n\nTrain: 45,654, Test: 5,638, Validation: 5,074\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process\n\n\nNo annotations.", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators\n\n\nThe dataset was created by Mehrdad Farahani.", "### Licensing Information\n\n\nApache License 2.0", "### Contributions\n\n\nThanks to @tanmoyio for adding this dataset." ]
[ 188, 102, 10, 22, 6, 51, 31, 7, 4, 10, 10, 5, 10, 9, 18, 7, 8, 14, 17, 10, 18 ]
[ "passage: TAGS\n#task_categories-text2text-generation #task_categories-translation #task_categories-question-answering #task_categories-summarization #task_ids-abstractive-qa #task_ids-explanation-generation #task_ids-extractive-qa #task_ids-open-domain-qa #task_ids-open-domain-abstractive-qa #task_ids-text-simplification #annotations_creators-no-annotation #language_creators-crowdsourced #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-Persian #license-apache-2.0 #region-us \n### Dataset Summary\n\n\nThe dataset extracted from Persian Wikipedia into the form of articles and highlights and cleaned the dataset into pairs of articles and highlights and reduced the articles' length (only version 1.0.0) and highlights' length to a maximum of 512 and 128, respectively, suitable for parsBERT. This dataset is created to achieve state-of-the-art results on some interesting NLP tasks like Text Summarization.### Supported Tasks and Leaderboards### Languages\n\n\nThe text in the dataset is in Percy.\n\n\nDataset Structure\n-----------------### Data Instances### Data Fields\n\n\n* 'id': Article id\n* 'link': Article link\n* 'title': Title of the article\n* 'article': Full text content in the article\n* 'highlights': Summary of the article### Data Splits\n\n\nTrain: 45,654, Test: 5,638, Validation: 5,074\n\n\nDataset Creation\n----------------### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process\n\n\nNo annotations.#### Who are the annotators?### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------### Social Impact of Dataset### Discussion of Biases" ]
fc3acd1fa45a1305ba6e98448db885fea00bced3
# Dataset Card for WikiANN ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Massively Multilingual Transfer for NER](https://github.com/afshinrahimi/mmner) - **Repository:** [Massively Multilingual Transfer for NER](https://github.com/afshinrahimi/mmner) - **Paper:** The original datasets come from the _Cross-lingual name tagging and linking for 282 languages_ [paper](https://www.aclweb.org/anthology/P17-1178/) by Xiaoman Pan et al. (2018). This version corresponds to the balanced train, dev, and test splits of the original data from the _Massively Multilingual Transfer for NER_ [paper](https://arxiv.org/abs/1902.00193) by Afshin Rahimi et al. (2019). - **Leaderboard:** - **Point of Contact:** [Afshin Rahimi](mailto:afshinrahimi@gmail.com) or [Lewis Tunstall](mailto:lewis.c.tunstall@gmail.com) or [Albert Villanova del Moral](albert@huggingface.co) ### Dataset Summary WikiANN (sometimes called PAN-X) is a multilingual named entity recognition dataset consisting of Wikipedia articles annotated with LOC (location), PER (person), and ORG (organisation) tags in the IOB2 format. This version corresponds to the balanced train, dev, and test splits of Rahimi et al. (2019), which supports 176 of the 282 languages from the original WikiANN corpus. ### Supported Tasks and Leaderboards - `named-entity-recognition`: The dataset can be used to train a model for named entity recognition in many languages, or evaluate the zero-shot cross-lingual capabilities of multilingual models. ### Languages The dataset contains 176 languages, one in each of the configuration subsets. The corresponding BCP 47 language tags are: | | Language tag | |:-------------------|:---------------| | ace | ace | | af | af | | als | als | | am | am | | an | an | | ang | ang | | ar | ar | | arc | arc | | arz | arz | | as | as | | ast | ast | | ay | ay | | az | az | | ba | ba | | bar | bar | | be | be | | bg | bg | | bh | bh | | bn | bn | | bo | bo | | br | br | | bs | bs | | ca | ca | | cdo | cdo | | ce | ce | | ceb | ceb | | ckb | ckb | | co | co | | crh | crh | | cs | cs | | csb | csb | | cv | cv | | cy | cy | | da | da | | de | de | | diq | diq | | dv | dv | | el | el | | en | en | | eo | eo | | es | es | | et | et | | eu | eu | | ext | ext | | fa | fa | | fi | fi | | fo | fo | | fr | fr | | frr | frr | | fur | fur | | fy | fy | | ga | ga | | gan | gan | | gd | gd | | gl | gl | | gn | gn | | gu | gu | | hak | hak | | he | he | | hi | hi | | hr | hr | | hsb | hsb | | hu | hu | | hy | hy | | ia | ia | | id | id | | ig | ig | | ilo | ilo | | io | io | | is | is | | it | it | | ja | ja | | jbo | jbo | | jv | jv | | ka | ka | | kk | kk | | km | km | | kn | kn | | ko | ko | | ksh | ksh | | ku | ku | | ky | ky | | la | la | | lb | lb | | li | li | | lij | lij | | lmo | lmo | | ln | ln | | lt | lt | | lv | lv | | mg | mg | | mhr | mhr | | mi | mi | | min | min | | mk | mk | | ml | ml | | mn | mn | | mr | mr | | ms | ms | | mt | mt | | mwl | mwl | | my | my | | mzn | mzn | | nap | nap | | nds | nds | | ne | ne | | nl | nl | | nn | nn | | no | no | | nov | nov | | oc | oc | | or | or | | os | os | | other-bat-smg | sgs | | other-be-x-old | be-tarask | | other-cbk-zam | cbk | | other-eml | eml | | other-fiu-vro | vro | | other-map-bms | jv-x-bms | | other-simple | en-basiceng | | other-zh-classical | lzh | | other-zh-min-nan | nan | | other-zh-yue | yue | | pa | pa | | pdc | pdc | | pl | pl | | pms | pms | | pnb | pnb | | ps | ps | | pt | pt | | qu | qu | | rm | rm | | ro | ro | | ru | ru | | rw | rw | | sa | sa | | sah | sah | | scn | scn | | sco | sco | | sd | sd | | sh | sh | | si | si | | sk | sk | | sl | sl | | so | so | | sq | sq | | sr | sr | | su | su | | sv | sv | | sw | sw | | szl | szl | | ta | ta | | te | te | | tg | tg | | th | th | | tk | tk | | tl | tl | | tr | tr | | tt | tt | | ug | ug | | uk | uk | | ur | ur | | uz | uz | | vec | vec | | vep | vep | | vi | vi | | vls | vls | | vo | vo | | wa | wa | | war | war | | wuu | wuu | | xmf | xmf | | yi | yi | | yo | yo | | zea | zea | | zh | zh | ## Dataset Structure ### Data Instances This is an example in the "train" split of the "af" (Afrikaans language) configuration subset: ```python { 'tokens': ['Sy', 'ander', 'seun', ',', 'Swjatopolk', ',', 'was', 'die', 'resultaat', 'van', '’n', 'buite-egtelike', 'verhouding', '.'], 'ner_tags': [0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'langs': ['af', 'af', 'af', 'af', 'af', 'af', 'af', 'af', 'af', 'af', 'af', 'af', 'af', 'af'], 'spans': ['PER: Swjatopolk'] } ``` ### Data Fields - `tokens`: a `list` of `string` features. - `langs`: a `list` of `string` features that correspond to the language of each token. - `ner_tags`: a `list` of classification labels, with possible values including `O` (0), `B-PER` (1), `I-PER` (2), `B-ORG` (3), `I-ORG` (4), `B-LOC` (5), `I-LOC` (6). - `spans`: a `list` of `string` features, that is the list of named entities in the input text formatted as ``<TAG>: <mention>`` ### Data Splits For each configuration subset, the data is split into "train", "validation" and "test" sets, each containing the following number of examples: | | Train | Validation | Test | |:-------------|--------:|-------------:|-------:| | ace | 100 | 100 | 100 | | af | 5000 | 1000 | 1000 | | als | 100 | 100 | 100 | | am | 100 | 100 | 100 | | an | 1000 | 1000 | 1000 | | ang | 100 | 100 | 100 | | ar | 20000 | 10000 | 10000 | | arc | 100 | 100 | 100 | | arz | 100 | 100 | 100 | | as | 100 | 100 | 100 | | ast | 1000 | 1000 | 1000 | | ay | 100 | 100 | 100 | | az | 10000 | 1000 | 1000 | | ba | 100 | 100 | 100 | | bar | 100 | 100 | 100 | | bat-smg | 100 | 100 | 100 | | be | 15000 | 1000 | 1000 | | be-x-old | 5000 | 1000 | 1000 | | bg | 20000 | 10000 | 10000 | | bh | 100 | 100 | 100 | | bn | 10000 | 1000 | 1000 | | bo | 100 | 100 | 100 | | br | 1000 | 1000 | 1000 | | bs | 15000 | 1000 | 1000 | | ca | 20000 | 10000 | 10000 | | cbk-zam | 100 | 100 | 100 | | cdo | 100 | 100 | 100 | | ce | 100 | 100 | 100 | | ceb | 100 | 100 | 100 | | ckb | 1000 | 1000 | 1000 | | co | 100 | 100 | 100 | | crh | 100 | 100 | 100 | | cs | 20000 | 10000 | 10000 | | csb | 100 | 100 | 100 | | cv | 100 | 100 | 100 | | cy | 10000 | 1000 | 1000 | | da | 20000 | 10000 | 10000 | | de | 20000 | 10000 | 10000 | | diq | 100 | 100 | 100 | | dv | 100 | 100 | 100 | | el | 20000 | 10000 | 10000 | | eml | 100 | 100 | 100 | | en | 20000 | 10000 | 10000 | | eo | 15000 | 10000 | 10000 | | es | 20000 | 10000 | 10000 | | et | 15000 | 10000 | 10000 | | eu | 10000 | 10000 | 10000 | | ext | 100 | 100 | 100 | | fa | 20000 | 10000 | 10000 | | fi | 20000 | 10000 | 10000 | | fiu-vro | 100 | 100 | 100 | | fo | 100 | 100 | 100 | | fr | 20000 | 10000 | 10000 | | frr | 100 | 100 | 100 | | fur | 100 | 100 | 100 | | fy | 1000 | 1000 | 1000 | | ga | 1000 | 1000 | 1000 | | gan | 100 | 100 | 100 | | gd | 100 | 100 | 100 | | gl | 15000 | 10000 | 10000 | | gn | 100 | 100 | 100 | | gu | 100 | 100 | 100 | | hak | 100 | 100 | 100 | | he | 20000 | 10000 | 10000 | | hi | 5000 | 1000 | 1000 | | hr | 20000 | 10000 | 10000 | | hsb | 100 | 100 | 100 | | hu | 20000 | 10000 | 10000 | | hy | 15000 | 1000 | 1000 | | ia | 100 | 100 | 100 | | id | 20000 | 10000 | 10000 | | ig | 100 | 100 | 100 | | ilo | 100 | 100 | 100 | | io | 100 | 100 | 100 | | is | 1000 | 1000 | 1000 | | it | 20000 | 10000 | 10000 | | ja | 20000 | 10000 | 10000 | | jbo | 100 | 100 | 100 | | jv | 100 | 100 | 100 | | ka | 10000 | 10000 | 10000 | | kk | 1000 | 1000 | 1000 | | km | 100 | 100 | 100 | | kn | 100 | 100 | 100 | | ko | 20000 | 10000 | 10000 | | ksh | 100 | 100 | 100 | | ku | 100 | 100 | 100 | | ky | 100 | 100 | 100 | | la | 5000 | 1000 | 1000 | | lb | 5000 | 1000 | 1000 | | li | 100 | 100 | 100 | | lij | 100 | 100 | 100 | | lmo | 100 | 100 | 100 | | ln | 100 | 100 | 100 | | lt | 10000 | 10000 | 10000 | | lv | 10000 | 10000 | 10000 | | map-bms | 100 | 100 | 100 | | mg | 100 | 100 | 100 | | mhr | 100 | 100 | 100 | | mi | 100 | 100 | 100 | | min | 100 | 100 | 100 | | mk | 10000 | 1000 | 1000 | | ml | 10000 | 1000 | 1000 | | mn | 100 | 100 | 100 | | mr | 5000 | 1000 | 1000 | | ms | 20000 | 1000 | 1000 | | mt | 100 | 100 | 100 | | mwl | 100 | 100 | 100 | | my | 100 | 100 | 100 | | mzn | 100 | 100 | 100 | | nap | 100 | 100 | 100 | | nds | 100 | 100 | 100 | | ne | 100 | 100 | 100 | | nl | 20000 | 10000 | 10000 | | nn | 20000 | 1000 | 1000 | | no | 20000 | 10000 | 10000 | | nov | 100 | 100 | 100 | | oc | 100 | 100 | 100 | | or | 100 | 100 | 100 | | os | 100 | 100 | 100 | | pa | 100 | 100 | 100 | | pdc | 100 | 100 | 100 | | pl | 20000 | 10000 | 10000 | | pms | 100 | 100 | 100 | | pnb | 100 | 100 | 100 | | ps | 100 | 100 | 100 | | pt | 20000 | 10000 | 10000 | | qu | 100 | 100 | 100 | | rm | 100 | 100 | 100 | | ro | 20000 | 10000 | 10000 | | ru | 20000 | 10000 | 10000 | | rw | 100 | 100 | 100 | | sa | 100 | 100 | 100 | | sah | 100 | 100 | 100 | | scn | 100 | 100 | 100 | | sco | 100 | 100 | 100 | | sd | 100 | 100 | 100 | | sh | 20000 | 10000 | 10000 | | si | 100 | 100 | 100 | | simple | 20000 | 1000 | 1000 | | sk | 20000 | 10000 | 10000 | | sl | 15000 | 10000 | 10000 | | so | 100 | 100 | 100 | | sq | 5000 | 1000 | 1000 | | sr | 20000 | 10000 | 10000 | | su | 100 | 100 | 100 | | sv | 20000 | 10000 | 10000 | | sw | 1000 | 1000 | 1000 | | szl | 100 | 100 | 100 | | ta | 15000 | 1000 | 1000 | | te | 1000 | 1000 | 1000 | | tg | 100 | 100 | 100 | | th | 20000 | 10000 | 10000 | | tk | 100 | 100 | 100 | | tl | 10000 | 1000 | 1000 | | tr | 20000 | 10000 | 10000 | | tt | 1000 | 1000 | 1000 | | ug | 100 | 100 | 100 | | uk | 20000 | 10000 | 10000 | | ur | 20000 | 1000 | 1000 | | uz | 1000 | 1000 | 1000 | | vec | 100 | 100 | 100 | | vep | 100 | 100 | 100 | | vi | 20000 | 10000 | 10000 | | vls | 100 | 100 | 100 | | vo | 100 | 100 | 100 | | wa | 100 | 100 | 100 | | war | 100 | 100 | 100 | | wuu | 100 | 100 | 100 | | xmf | 100 | 100 | 100 | | yi | 100 | 100 | 100 | | yo | 100 | 100 | 100 | | zea | 100 | 100 | 100 | | zh | 20000 | 10000 | 10000 | | zh-classical | 100 | 100 | 100 | | zh-min-nan | 100 | 100 | 100 | | zh-yue | 20000 | 10000 | 10000 | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information The original 282 datasets are associated with this article ``` @inproceedings{pan-etal-2017-cross, title = "Cross-lingual Name Tagging and Linking for 282 Languages", author = "Pan, Xiaoman and Zhang, Boliang and May, Jonathan and Nothman, Joel and Knight, Kevin and Ji, Heng", booktitle = "Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = jul, year = "2017", address = "Vancouver, Canada", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P17-1178", doi = "10.18653/v1/P17-1178", pages = "1946--1958", abstract = "The ambitious goal of this work is to develop a cross-lingual name tagging and linking framework for 282 languages that exist in Wikipedia. Given a document in any of these languages, our framework is able to identify name mentions, assign a coarse-grained or fine-grained type to each mention, and link it to an English Knowledge Base (KB) if it is linkable. We achieve this goal by performing a series of new KB mining methods: generating {``}silver-standard{''} annotations by transferring annotations from English to other languages through cross-lingual links and KB properties, refining annotations through self-training and topic selection, deriving language-specific morphology features from anchor links, and mining word translation pairs from cross-lingual links. Both name tagging and linking results for 282 languages are promising on Wikipedia data and on-Wikipedia data.", } ``` while the 176 languages supported in this version are associated with the following article ``` @inproceedings{rahimi-etal-2019-massively, title = "Massively Multilingual Transfer for {NER}", author = "Rahimi, Afshin and Li, Yuan and Cohn, Trevor", booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics", month = jul, year = "2019", address = "Florence, Italy", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P19-1015", pages = "151--164", } ``` ### Contributions Thanks to [@lewtun](https://github.com/lewtun) and [@rabeehk](https://github.com/rabeehk) for adding this dataset.
wikiann
[ "task_categories:token-classification", "task_ids:named-entity-recognition", "annotations_creators:machine-generated", "language_creators:crowdsourced", "multilinguality:multilingual", "size_categories:n<1K", "source_datasets:original", "language:ace", "language:af", "language:als", "language:am", "language:an", "language:ang", "language:ar", "language:arc", "language:arz", "language:as", "language:ast", "language:ay", "language:az", "language:ba", "language:bar", "language:be", "language:bg", "language:bh", "language:bn", "language:bo", "language:br", "language:bs", "language:ca", "language:cbk", "language:cdo", "language:ce", "language:ceb", "language:ckb", "language:co", "language:crh", "language:cs", "language:csb", "language:cv", "language:cy", "language:da", "language:de", "language:diq", "language:dv", "language:el", "language:eml", "language:en", "language:eo", "language:es", "language:et", "language:eu", "language:ext", "language:fa", "language:fi", "language:fo", "language:fr", "language:frr", "language:fur", "language:fy", "language:ga", "language:gan", "language:gd", "language:gl", "language:gn", "language:gu", "language:hak", "language:he", "language:hi", "language:hr", "language:hsb", "language:hu", "language:hy", "language:ia", "language:id", "language:ig", "language:ilo", "language:io", "language:is", "language:it", "language:ja", "language:jbo", "language:jv", "language:ka", "language:kk", "language:km", "language:kn", "language:ko", "language:ksh", "language:ku", "language:ky", "language:la", "language:lb", "language:li", "language:lij", "language:lmo", "language:ln", "language:lt", "language:lv", "language:lzh", "language:mg", "language:mhr", "language:mi", "language:min", "language:mk", "language:ml", "language:mn", "language:mr", "language:ms", "language:mt", "language:mwl", "language:my", "language:mzn", "language:nan", "language:nap", "language:nds", "language:ne", "language:nl", "language:nn", "language:no", "language:nov", "language:oc", "language:or", "language:os", "language:pa", "language:pdc", "language:pl", "language:pms", "language:pnb", "language:ps", "language:pt", "language:qu", "language:rm", "language:ro", "language:ru", "language:rw", "language:sa", "language:sah", "language:scn", "language:sco", "language:sd", "language:sgs", "language:sh", "language:si", "language:sk", "language:sl", "language:so", "language:sq", "language:sr", "language:su", "language:sv", "language:sw", "language:szl", "language:ta", "language:te", "language:tg", "language:th", "language:tk", "language:tl", "language:tr", "language:tt", "language:ug", "language:uk", "language:ur", "language:uz", "language:vec", "language:vep", "language:vi", "language:vls", "language:vo", "language:vro", "language:wa", "language:war", "language:wuu", "language:xmf", "language:yi", "language:yo", "language:yue", "language:zea", "language:zh", "license:unknown", "arxiv:1902.00193", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["machine-generated"], "language_creators": ["crowdsourced"], "language": ["ace", "af", "als", "am", "an", "ang", "ar", "arc", "arz", "as", "ast", "ay", "az", "ba", "bar", "be", "bg", "bh", "bn", "bo", "br", "bs", "ca", "cbk", "cdo", "ce", "ceb", "ckb", "co", "crh", "cs", "csb", "cv", "cy", "da", "de", "diq", "dv", "el", "eml", "en", "eo", "es", "et", "eu", "ext", "fa", "fi", "fo", "fr", "frr", "fur", "fy", "ga", "gan", "gd", "gl", "gn", "gu", "hak", "he", "hi", "hr", "hsb", "hu", "hy", "ia", "id", "ig", "ilo", "io", "is", "it", "ja", "jbo", "jv", "ka", "kk", "km", "kn", "ko", "ksh", "ku", "ky", "la", "lb", "li", "lij", "lmo", "ln", "lt", "lv", "lzh", "mg", "mhr", "mi", "min", "mk", "ml", "mn", "mr", "ms", "mt", "mwl", "my", "mzn", "nan", "nap", "nds", "ne", "nl", "nn", "no", "nov", "oc", "or", "os", "pa", "pdc", "pl", "pms", "pnb", "ps", "pt", "qu", "rm", "ro", "ru", "rw", "sa", "sah", "scn", "sco", "sd", "sgs", "sh", "si", "sk", "sl", "so", "sq", "sr", "su", "sv", "sw", "szl", "ta", "te", "tg", "th", "tk", "tl", "tr", "tt", "ug", "uk", "ur", "uz", "vec", "vep", "vi", "vls", "vo", "vro", "wa", "war", "wuu", "xmf", "yi", "yo", "yue", "zea", "zh"], "license": ["unknown"], "multilinguality": ["multilingual"], "size_categories": ["n<1K"], "source_datasets": ["original"], "task_categories": ["token-classification"], "task_ids": ["named-entity-recognition"], "paperswithcode_id": "wikiann-1", "pretty_name": "WikiANN", "config_names": ["no", "ace", "af", "als", "am", "an", "ang", "ar", "arc", "arz", "as", "ast", "ay", "az", "ba", "bar", "be", "bg", "bh", "bn", "bo", "br", "bs", "ca", "cdo", "ce", "ceb", "ckb", "co", "crh", "cs", "csb", "cv", "cy", "da", "de", "diq", "dv", "el", "en", "eo", "es", "et", "eu", "ext", "fa", "fi", "fo", "fr", "frr", "fur", "fy", "ga", "gan", "gd", "gl", "gn", "gu", "hak", "he", "hi", "hr", "hsb", "hu", "hy", "ia", "id", "ig", "ilo", "io", "is", "it", "ja", "jbo", "jv", "ka", "kk", "km", "kn", "ko", "ksh", "ku", "ky", "la", "lb", "li", "lij", "lmo", "ln", "lt", "lv", "mg", "mhr", "mi", "min", "mk", "ml", "mn", "mr", "ms", "mt", "mwl", "my", "mzn", "nap", "nds", "ne", "nl", "nn", "nov", "oc", "or", "os", "other-bat-smg", "other-be-x-old", "other-cbk-zam", "other-eml", "other-fiu-vro", "other-map-bms", "other-simple", "other-zh-classical", "other-zh-min-nan", "other-zh-yue", "pa", "pdc", "pl", "pms", "pnb", "ps", "pt", "qu", "rm", "ro", "ru", "rw", "sa", "sah", "scn", "sco", "sd", "sh", "si", "sk", "sl", "so", "sq", "sr", "su", "sv", "sw", "szl", "ta", "te", "tg", "th", "tk", "tl", "tr", "tt", "ug", "uk", "ur", "uz", "vec", "vep", "vi", "vls", "vo", "wa", "war", "wuu", "xmf", "yi", "yo", "zea", "zh"], "language_bcp47": ["be-tarask", "en-basiceng", "jv-x-bms"], "dataset_info": [{"config_name": "ace", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 22453, "num_examples": 100}, {"name": "test", "num_bytes": 25752, "num_examples": 100}, {"name": "train", "num_bytes": 23231, "num_examples": 100}], "download_size": 234008884, "dataset_size": 71436}, {"config_name": "af", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 299137, "num_examples": 1000}, {"name": "test", "num_bytes": 295849, "num_examples": 1000}, {"name": "train", "num_bytes": 1521604, "num_examples": 5000}], "download_size": 234008884, "dataset_size": 2116590}, {"config_name": "als", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 34318, "num_examples": 100}, {"name": "test", "num_bytes": 36345, "num_examples": 100}, {"name": "train", "num_bytes": 34968, "num_examples": 100}], "download_size": 234008884, "dataset_size": 105631}, {"config_name": "am", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 21429, "num_examples": 100}, {"name": "test", "num_bytes": 23811, "num_examples": 100}, {"name": "train", "num_bytes": 22214, "num_examples": 100}], "download_size": 234008884, "dataset_size": 67454}, {"config_name": "an", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 180609, "num_examples": 1000}, {"name": "test", "num_bytes": 174992, "num_examples": 1000}, {"name": "train", "num_bytes": 180967, "num_examples": 1000}], "download_size": 234008884, "dataset_size": 536568}, {"config_name": "ang", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 21925, "num_examples": 100}, {"name": "test", "num_bytes": 24523, "num_examples": 100}, {"name": "train", "num_bytes": 23296, "num_examples": 100}], "download_size": 234008884, "dataset_size": 69744}, {"config_name": "ar", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2325688, "num_examples": 10000}, {"name": "test", "num_bytes": 2334664, "num_examples": 10000}, {"name": "train", "num_bytes": 4671669, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 9332021}, {"config_name": "arc", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 15726, "num_examples": 100}, {"name": "test", "num_bytes": 16641, "num_examples": 100}, {"name": "train", "num_bytes": 18536, "num_examples": 100}], "download_size": 234008884, "dataset_size": 50903}, {"config_name": "arz", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 26609, "num_examples": 100}, {"name": "test", "num_bytes": 25663, "num_examples": 100}, {"name": "train", "num_bytes": 26375, "num_examples": 100}], "download_size": 234008884, "dataset_size": 78647}, {"config_name": "as", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 25736, "num_examples": 100}, {"name": "test", "num_bytes": 23350, "num_examples": 100}, {"name": "train", "num_bytes": 24984, "num_examples": 100}], "download_size": 234008884, "dataset_size": 74070}, {"config_name": "ast", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 217477, "num_examples": 1000}, {"name": "test", "num_bytes": 220874, "num_examples": 1000}, {"name": "train", "num_bytes": 228238, "num_examples": 1000}], "download_size": 234008884, "dataset_size": 666589}, {"config_name": "ay", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 11684, "num_examples": 100}, {"name": "test", "num_bytes": 13379, "num_examples": 100}, {"name": "train", "num_bytes": 12596, "num_examples": 100}], "download_size": 234008884, "dataset_size": 37659}, {"config_name": "az", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 272066, "num_examples": 1000}, {"name": "test", "num_bytes": 267935, "num_examples": 1000}, {"name": "train", "num_bytes": 2645552, "num_examples": 10000}], "download_size": 234008884, "dataset_size": 3185553}, {"config_name": "ba", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 29262, "num_examples": 100}, {"name": "test", "num_bytes": 30502, "num_examples": 100}, {"name": "train", "num_bytes": 31123, "num_examples": 100}], "download_size": 234008884, "dataset_size": 90887}, {"config_name": "bar", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 17374, "num_examples": 100}, {"name": "test", "num_bytes": 17839, "num_examples": 100}, {"name": "train", "num_bytes": 16796, "num_examples": 100}], "download_size": 234008884, "dataset_size": 52009}, {"config_name": "bat-smg", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 26496, "num_examples": 100}, {"name": "test", "num_bytes": 26093, "num_examples": 100}, {"name": "train", "num_bytes": 24677, "num_examples": 100}], "download_size": 234008884, "dataset_size": 77266}, {"config_name": "be", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 262042, "num_examples": 1000}, {"name": "test", "num_bytes": 266104, "num_examples": 1000}, {"name": "train", "num_bytes": 3983322, "num_examples": 15000}], "download_size": 234008884, "dataset_size": 4511468}, {"config_name": "be-x-old", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 342654, "num_examples": 1000}, {"name": "test", "num_bytes": 337599, "num_examples": 1000}, {"name": "train", "num_bytes": 1704256, "num_examples": 5000}], "download_size": 234008884, "dataset_size": 2384509}, {"config_name": "bg", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2840907, "num_examples": 10000}, {"name": "test", "num_bytes": 2830213, "num_examples": 10000}, {"name": "train", "num_bytes": 5665063, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 11336183}, {"config_name": "bh", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 33682, "num_examples": 100}, {"name": "test", "num_bytes": 30692, "num_examples": 100}, {"name": "train", "num_bytes": 36374, "num_examples": 100}], "download_size": 234008884, "dataset_size": 100748}, {"config_name": "bn", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 238446, "num_examples": 1000}, {"name": "test", "num_bytes": 237218, "num_examples": 1000}, {"name": "train", "num_bytes": 2351591, "num_examples": 10000}], "download_size": 234008884, "dataset_size": 2827255}, {"config_name": "bo", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 22688, "num_examples": 100}, {"name": "test", "num_bytes": 15437, "num_examples": 100}, {"name": "train", "num_bytes": 14085, "num_examples": 100}], "download_size": 234008884, "dataset_size": 52210}, {"config_name": "br", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 206839, "num_examples": 1000}, {"name": "test", "num_bytes": 222083, "num_examples": 1000}, {"name": "train", "num_bytes": 221495, "num_examples": 1000}], "download_size": 234008884, "dataset_size": 650417}, {"config_name": "bs", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 246378, "num_examples": 1000}, {"name": "test", "num_bytes": 247331, "num_examples": 1000}, {"name": "train", "num_bytes": 3669346, "num_examples": 15000}], "download_size": 234008884, "dataset_size": 4163055}, {"config_name": "ca", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 1836319, "num_examples": 10000}, {"name": "test", "num_bytes": 1847746, "num_examples": 10000}, {"name": "train", "num_bytes": 3689342, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 7373407}, {"config_name": "cbk-zam", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 47060, "num_examples": 100}, {"name": "test", "num_bytes": 47277, "num_examples": 100}, {"name": "train", "num_bytes": 52545, "num_examples": 100}], "download_size": 234008884, "dataset_size": 146882}, {"config_name": "cdo", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 37479, "num_examples": 100}, {"name": "test", "num_bytes": 34319, "num_examples": 100}, {"name": "train", "num_bytes": 36204, "num_examples": 100}], "download_size": 234008884, "dataset_size": 108002}, {"config_name": "ce", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 40303, "num_examples": 100}, {"name": "test", "num_bytes": 38640, "num_examples": 100}, {"name": "train", "num_bytes": 38284, "num_examples": 100}], "download_size": 234008884, "dataset_size": 117227}, {"config_name": "ceb", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 22789, "num_examples": 100}, {"name": "test", "num_bytes": 23950, "num_examples": 100}, {"name": "train", "num_bytes": 21365, "num_examples": 100}], "download_size": 234008884, "dataset_size": 68104}, {"config_name": "ckb", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 214231, "num_examples": 1000}, {"name": "test", "num_bytes": 211988, "num_examples": 1000}, {"name": "train", "num_bytes": 217066, "num_examples": 1000}], "download_size": 234008884, "dataset_size": 643285}, {"config_name": "co", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 15968, "num_examples": 100}, {"name": "test", "num_bytes": 15880, "num_examples": 100}, {"name": "train", "num_bytes": 18032, "num_examples": 100}], "download_size": 234008884, "dataset_size": 49880}, {"config_name": "crh", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 20230, "num_examples": 100}, {"name": "test", "num_bytes": 23879, "num_examples": 100}, {"name": "train", "num_bytes": 23336, "num_examples": 100}], "download_size": 234008884, "dataset_size": 67445}, {"config_name": "cs", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2456654, "num_examples": 10000}, {"name": "test", "num_bytes": 2458155, "num_examples": 10000}, {"name": "train", "num_bytes": 4944758, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 9859567}, {"config_name": "csb", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 28841, "num_examples": 100}, {"name": "test", "num_bytes": 27840, "num_examples": 100}, {"name": "train", "num_bytes": 31640, "num_examples": 100}], "download_size": 234008884, "dataset_size": 88321}, {"config_name": "cv", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 24787, "num_examples": 100}, {"name": "test", "num_bytes": 26403, "num_examples": 100}, {"name": "train", "num_bytes": 26956, "num_examples": 100}], "download_size": 234008884, "dataset_size": 78146}, {"config_name": "cy", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 228586, "num_examples": 1000}, {"name": "test", "num_bytes": 233869, "num_examples": 1000}, {"name": "train", "num_bytes": 2337116, "num_examples": 10000}], "download_size": 234008884, "dataset_size": 2799571}, {"config_name": "da", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2422976, "num_examples": 10000}, {"name": "test", "num_bytes": 2432324, "num_examples": 10000}, {"name": "train", "num_bytes": 4882222, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 9737522}, {"config_name": "de", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2754550, "num_examples": 10000}, {"name": "test", "num_bytes": 2750996, "num_examples": 10000}, {"name": "train", "num_bytes": 5510641, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 11016187}, {"config_name": "diq", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 24147, "num_examples": 100}, {"name": "test", "num_bytes": 22476, "num_examples": 100}, {"name": "train", "num_bytes": 24131, "num_examples": 100}], "download_size": 234008884, "dataset_size": 70754}, {"config_name": "dv", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 30322, "num_examples": 100}, {"name": "test", "num_bytes": 27279, "num_examples": 100}, {"name": "train", "num_bytes": 31033, "num_examples": 100}], "download_size": 234008884, "dataset_size": 88634}, {"config_name": "el", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 3027962, "num_examples": 10000}, {"name": "test", "num_bytes": 3034329, "num_examples": 10000}, {"name": "train", "num_bytes": 6046638, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 12108929}, {"config_name": "eml", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 30050, "num_examples": 100}, {"name": "test", "num_bytes": 35880, "num_examples": 100}, {"name": "train", "num_bytes": 30792, "num_examples": 100}], "download_size": 234008884, "dataset_size": 96722}, {"config_name": "en", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2336353, "num_examples": 10000}, {"name": "test", "num_bytes": 2330245, "num_examples": 10000}, {"name": "train", "num_bytes": 4649601, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 9316199}, {"config_name": "eo", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 1968690, "num_examples": 10000}, {"name": "test", "num_bytes": 1961486, "num_examples": 10000}, {"name": "train", "num_bytes": 2952610, "num_examples": 15000}], "download_size": 234008884, "dataset_size": 6882786}, {"config_name": "es", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 1976935, "num_examples": 10000}, {"name": "test", "num_bytes": 1986664, "num_examples": 10000}, {"name": "train", "num_bytes": 3972292, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 7935891}, {"config_name": "et", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2403361, "num_examples": 10000}, {"name": "test", "num_bytes": 2392424, "num_examples": 10000}, {"name": "train", "num_bytes": 3579264, "num_examples": 15000}], "download_size": 234008884, "dataset_size": 8375049}, {"config_name": "eu", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2677036, "num_examples": 10000}, {"name": "test", "num_bytes": 2628951, "num_examples": 10000}, {"name": "train", "num_bytes": 2672353, "num_examples": 10000}], "download_size": 234008884, "dataset_size": 7978340}, {"config_name": "ext", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 30821, "num_examples": 100}, {"name": "test", "num_bytes": 29483, "num_examples": 100}, {"name": "train", "num_bytes": 23110, "num_examples": 100}], "download_size": 234008884, "dataset_size": 83414}, {"config_name": "fa", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2328640, "num_examples": 10000}, {"name": "test", "num_bytes": 2314687, "num_examples": 10000}, {"name": "train", "num_bytes": 4618098, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 9261425}, {"config_name": "fi", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2500586, "num_examples": 10000}, {"name": "test", "num_bytes": 2505161, "num_examples": 10000}, {"name": "train", "num_bytes": 5020655, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 10026402}, {"config_name": "fiu-vro", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 27672, "num_examples": 100}, {"name": "test", "num_bytes": 27728, "num_examples": 100}, {"name": "train", "num_bytes": 28689, "num_examples": 100}], "download_size": 234008884, "dataset_size": 84089}, {"config_name": "fo", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 26094, "num_examples": 100}, {"name": "test", "num_bytes": 23531, "num_examples": 100}, {"name": "train", "num_bytes": 26178, "num_examples": 100}], "download_size": 234008884, "dataset_size": 75803}, {"config_name": "fr", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2058004, "num_examples": 10000}, {"name": "test", "num_bytes": 2073593, "num_examples": 10000}, {"name": "train", "num_bytes": 4123995, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 8255592}, {"config_name": "frr", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 15883, "num_examples": 100}, {"name": "test", "num_bytes": 15736, "num_examples": 100}, {"name": "train", "num_bytes": 16654, "num_examples": 100}], "download_size": 234008884, "dataset_size": 48273}, {"config_name": "fur", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 25264, "num_examples": 100}, {"name": "test", "num_bytes": 30562, "num_examples": 100}, {"name": "train", "num_bytes": 33654, "num_examples": 100}], "download_size": 234008884, "dataset_size": 89480}, {"config_name": "fy", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 226436, "num_examples": 1000}, {"name": "test", "num_bytes": 229700, "num_examples": 1000}, {"name": "train", "num_bytes": 223013, "num_examples": 1000}], "download_size": 234008884, "dataset_size": 679149}, {"config_name": "ga", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 234092, "num_examples": 1000}, {"name": "test", "num_bytes": 235083, "num_examples": 1000}, {"name": "train", "num_bytes": 238047, "num_examples": 1000}], "download_size": 234008884, "dataset_size": 707222}, {"config_name": "gan", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 17533, "num_examples": 100}, {"name": "test", "num_bytes": 13879, "num_examples": 100}, {"name": "train", "num_bytes": 14398, "num_examples": 100}], "download_size": 234008884, "dataset_size": 45810}, {"config_name": "gd", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 23230, "num_examples": 100}, {"name": "test", "num_bytes": 20308, "num_examples": 100}, {"name": "train", "num_bytes": 20154, "num_examples": 100}], "download_size": 234008884, "dataset_size": 63692}, {"config_name": "gl", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2029683, "num_examples": 10000}, {"name": "test", "num_bytes": 2031150, "num_examples": 10000}, {"name": "train", "num_bytes": 3030993, "num_examples": 15000}], "download_size": 234008884, "dataset_size": 7091826}, {"config_name": "gn", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 29132, "num_examples": 100}, {"name": "test", "num_bytes": 24263, "num_examples": 100}, {"name": "train", "num_bytes": 28220, "num_examples": 100}], "download_size": 234008884, "dataset_size": 81615}, {"config_name": "gu", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 48009, "num_examples": 100}, {"name": "test", "num_bytes": 45417, "num_examples": 100}, {"name": "train", "num_bytes": 42625, "num_examples": 100}], "download_size": 234008884, "dataset_size": 136051}, {"config_name": "hak", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 17977, "num_examples": 100}, {"name": "test", "num_bytes": 18155, "num_examples": 100}, {"name": "train", "num_bytes": 16208, "num_examples": 100}], "download_size": 234008884, "dataset_size": 52340}, {"config_name": "he", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2801392, "num_examples": 10000}, {"name": "test", "num_bytes": 2785474, "num_examples": 10000}, {"name": "train", "num_bytes": 5600488, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 11187354}, {"config_name": "hi", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 261207, "num_examples": 1000}, {"name": "test", "num_bytes": 267255, "num_examples": 1000}, {"name": "train", "num_bytes": 1315829, "num_examples": 5000}], "download_size": 234008884, "dataset_size": 1844291}, {"config_name": "hr", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2417450, "num_examples": 10000}, {"name": "test", "num_bytes": 2430440, "num_examples": 10000}, {"name": "train", "num_bytes": 4877331, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 9725221}, {"config_name": "hsb", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 24695, "num_examples": 100}, {"name": "test", "num_bytes": 24348, "num_examples": 100}, {"name": "train", "num_bytes": 24228, "num_examples": 100}], "download_size": 234008884, "dataset_size": 73271}, {"config_name": "hu", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2590116, "num_examples": 10000}, {"name": "test", "num_bytes": 2626771, "num_examples": 10000}, {"name": "train", "num_bytes": 5263122, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 10480009}, {"config_name": "hy", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 237560, "num_examples": 1000}, {"name": "test", "num_bytes": 237121, "num_examples": 1000}, {"name": "train", "num_bytes": 3634065, "num_examples": 15000}], "download_size": 234008884, "dataset_size": 4108746}, {"config_name": "ia", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 32064, "num_examples": 100}, {"name": "test", "num_bytes": 37617, "num_examples": 100}, {"name": "train", "num_bytes": 32928, "num_examples": 100}], "download_size": 234008884, "dataset_size": 102609}, {"config_name": "id", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 1901625, "num_examples": 10000}, {"name": "test", "num_bytes": 1902732, "num_examples": 10000}, {"name": "train", "num_bytes": 3814047, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 7618404}, {"config_name": "ig", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 17721, "num_examples": 100}, {"name": "test", "num_bytes": 18432, "num_examples": 100}, {"name": "train", "num_bytes": 15988, "num_examples": 100}], "download_size": 234008884, "dataset_size": 52141}, {"config_name": "ilo", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 16675, "num_examples": 100}, {"name": "test", "num_bytes": 17245, "num_examples": 100}, {"name": "train", "num_bytes": 17152, "num_examples": 100}], "download_size": 234008884, "dataset_size": 51072}, {"config_name": "io", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 19026, "num_examples": 100}, {"name": "test", "num_bytes": 17231, "num_examples": 100}, {"name": "train", "num_bytes": 20781, "num_examples": 100}], "download_size": 234008884, "dataset_size": 57038}, {"config_name": "is", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 243667, "num_examples": 1000}, {"name": "test", "num_bytes": 235946, "num_examples": 1000}, {"name": "train", "num_bytes": 243465, "num_examples": 1000}], "download_size": 234008884, "dataset_size": 723078}, {"config_name": "it", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2282947, "num_examples": 10000}, {"name": "test", "num_bytes": 2307618, "num_examples": 10000}, {"name": "train", "num_bytes": 4633575, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 9224140}, {"config_name": "ja", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 6775608, "num_examples": 10000}, {"name": "test", "num_bytes": 6898538, "num_examples": 10000}, {"name": "train", "num_bytes": 13578325, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 27252471}, {"config_name": "jbo", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 15618, "num_examples": 100}, {"name": "test", "num_bytes": 19586, "num_examples": 100}, {"name": "train", "num_bytes": 15070, "num_examples": 100}], "download_size": 234008884, "dataset_size": 50274}, {"config_name": "jv", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 17691, "num_examples": 100}, {"name": "test", "num_bytes": 20203, "num_examples": 100}, {"name": "train", "num_bytes": 19409, "num_examples": 100}], "download_size": 234008884, "dataset_size": 57303}, {"config_name": "ka", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 3454381, "num_examples": 10000}, {"name": "test", "num_bytes": 3480870, "num_examples": 10000}, {"name": "train", "num_bytes": 3428008, "num_examples": 10000}], "download_size": 234008884, "dataset_size": 10363259}, {"config_name": "kk", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 286502, "num_examples": 1000}, {"name": "test", "num_bytes": 284503, "num_examples": 1000}, {"name": "train", "num_bytes": 287952, "num_examples": 1000}], "download_size": 234008884, "dataset_size": 858957}, {"config_name": "km", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 29310, "num_examples": 100}, {"name": "test", "num_bytes": 36101, "num_examples": 100}, {"name": "train", "num_bytes": 31938, "num_examples": 100}], "download_size": 234008884, "dataset_size": 97349}, {"config_name": "kn", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 36853, "num_examples": 100}, {"name": "test", "num_bytes": 32278, "num_examples": 100}, {"name": "train", "num_bytes": 34346, "num_examples": 100}], "download_size": 234008884, "dataset_size": 103477}, {"config_name": "ko", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2553068, "num_examples": 10000}, {"name": "test", "num_bytes": 2547800, "num_examples": 10000}, {"name": "train", "num_bytes": 5107090, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 10207958}, {"config_name": "ksh", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 26338, "num_examples": 100}, {"name": "test", "num_bytes": 25249, "num_examples": 100}, {"name": "train", "num_bytes": 25941, "num_examples": 100}], "download_size": 234008884, "dataset_size": 77528}, {"config_name": "ku", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 22597, "num_examples": 100}, {"name": "test", "num_bytes": 20795, "num_examples": 100}, {"name": "train", "num_bytes": 22669, "num_examples": 100}], "download_size": 234008884, "dataset_size": 66061}, {"config_name": "ky", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 31010, "num_examples": 100}, {"name": "test", "num_bytes": 31896, "num_examples": 100}, {"name": "train", "num_bytes": 32768, "num_examples": 100}], "download_size": 234008884, "dataset_size": 95674}, {"config_name": "la", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 207205, "num_examples": 1000}, {"name": "test", "num_bytes": 198910, "num_examples": 1000}, {"name": "train", "num_bytes": 999050, "num_examples": 5000}], "download_size": 234008884, "dataset_size": 1405165}, {"config_name": "lb", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 253774, "num_examples": 1000}, {"name": "test", "num_bytes": 249989, "num_examples": 1000}, {"name": "train", "num_bytes": 1260939, "num_examples": 5000}], "download_size": 234008884, "dataset_size": 1764702}, {"config_name": "li", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 20201, "num_examples": 100}, {"name": "test", "num_bytes": 18817, "num_examples": 100}, {"name": "train", "num_bytes": 20211, "num_examples": 100}], "download_size": 234008884, "dataset_size": 59229}, {"config_name": "lij", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 28005, "num_examples": 100}, {"name": "test", "num_bytes": 27882, "num_examples": 100}, {"name": "train", "num_bytes": 30581, "num_examples": 100}], "download_size": 234008884, "dataset_size": 86468}, {"config_name": "lmo", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 26575, "num_examples": 100}, {"name": "test", "num_bytes": 29453, "num_examples": 100}, {"name": "train", "num_bytes": 24161, "num_examples": 100}], "download_size": 234008884, "dataset_size": 80189}, {"config_name": "ln", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 21709, "num_examples": 100}, {"name": "test", "num_bytes": 27003, "num_examples": 100}, {"name": "train", "num_bytes": 22227, "num_examples": 100}], "download_size": 234008884, "dataset_size": 70939}, {"config_name": "lt", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2192874, "num_examples": 10000}, {"name": "test", "num_bytes": 2191269, "num_examples": 10000}, {"name": "train", "num_bytes": 2199946, "num_examples": 10000}], "download_size": 234008884, "dataset_size": 6584089}, {"config_name": "lv", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2173420, "num_examples": 10000}, {"name": "test", "num_bytes": 2190458, "num_examples": 10000}, {"name": "train", "num_bytes": 2206943, "num_examples": 10000}], "download_size": 234008884, "dataset_size": 6570821}, {"config_name": "map-bms", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 19780, "num_examples": 100}, {"name": "test", "num_bytes": 20558, "num_examples": 100}, {"name": "train", "num_bytes": 21639, "num_examples": 100}], "download_size": 234008884, "dataset_size": 61977}, {"config_name": "mg", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 24861, "num_examples": 100}, {"name": "test", "num_bytes": 22570, "num_examples": 100}, {"name": "train", "num_bytes": 25739, "num_examples": 100}], "download_size": 234008884, "dataset_size": 73170}, {"config_name": "mhr", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 23263, "num_examples": 100}, {"name": "test", "num_bytes": 23639, "num_examples": 100}, {"name": "train", "num_bytes": 18648, "num_examples": 100}], "download_size": 234008884, "dataset_size": 65550}, {"config_name": "mi", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 39399, "num_examples": 100}, {"name": "test", "num_bytes": 40147, "num_examples": 100}, {"name": "train", "num_bytes": 37896, "num_examples": 100}], "download_size": 234008884, "dataset_size": 117442}, {"config_name": "min", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 28719, "num_examples": 100}, {"name": "test", "num_bytes": 24741, "num_examples": 100}, {"name": "train", "num_bytes": 26620, "num_examples": 100}], "download_size": 234008884, "dataset_size": 80080}, {"config_name": "mk", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 333193, "num_examples": 1000}, {"name": "test", "num_bytes": 337757, "num_examples": 1000}, {"name": "train", "num_bytes": 3355936, "num_examples": 10000}], "download_size": 234008884, "dataset_size": 4026886}, {"config_name": "ml", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 363008, "num_examples": 1000}, {"name": "test", "num_bytes": 349383, "num_examples": 1000}, {"name": "train", "num_bytes": 3582066, "num_examples": 10000}], "download_size": 234008884, "dataset_size": 4294457}, {"config_name": "mn", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 22006, "num_examples": 100}, {"name": "test", "num_bytes": 23538, "num_examples": 100}, {"name": "train", "num_bytes": 23244, "num_examples": 100}], "download_size": 234008884, "dataset_size": 68788}, {"config_name": "mr", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 314858, "num_examples": 1000}, {"name": "test", "num_bytes": 326290, "num_examples": 1000}, {"name": "train", "num_bytes": 1598804, "num_examples": 5000}], "download_size": 234008884, "dataset_size": 2239952}, {"config_name": "ms", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 183944, "num_examples": 1000}, {"name": "test", "num_bytes": 183539, "num_examples": 1000}, {"name": "train", "num_bytes": 3699238, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 4066721}, {"config_name": "mt", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 24571, "num_examples": 100}, {"name": "test", "num_bytes": 24662, "num_examples": 100}, {"name": "train", "num_bytes": 24956, "num_examples": 100}], "download_size": 234008884, "dataset_size": 74189}, {"config_name": "mwl", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 51987, "num_examples": 100}, {"name": "test", "num_bytes": 43008, "num_examples": 100}, {"name": "train", "num_bytes": 44605, "num_examples": 100}], "download_size": 234008884, "dataset_size": 139600}, {"config_name": "my", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 48953, "num_examples": 100}, {"name": "test", "num_bytes": 45956, "num_examples": 100}, {"name": "train", "num_bytes": 41371, "num_examples": 100}], "download_size": 234008884, "dataset_size": 136280}, {"config_name": "mzn", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 25304, "num_examples": 100}, {"name": "test", "num_bytes": 25947, "num_examples": 100}, {"name": "train", "num_bytes": 24841, "num_examples": 100}], "download_size": 234008884, "dataset_size": 76092}, {"config_name": "nap", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 21546, "num_examples": 100}, {"name": "test", "num_bytes": 24194, "num_examples": 100}, {"name": "train", "num_bytes": 26596, "num_examples": 100}], "download_size": 234008884, "dataset_size": 72336}, {"config_name": "nds", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 28388, "num_examples": 100}, {"name": "test", "num_bytes": 26571, "num_examples": 100}, {"name": "train", "num_bytes": 24679, "num_examples": 100}], "download_size": 234008884, "dataset_size": 79638}, {"config_name": "ne", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 33932, "num_examples": 100}, {"name": "test", "num_bytes": 33227, "num_examples": 100}, {"name": "train", "num_bytes": 36173, "num_examples": 100}], "download_size": 234008884, "dataset_size": 103332}, {"config_name": "nl", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2378080, "num_examples": 10000}, {"name": "test", "num_bytes": 2403076, "num_examples": 10000}, {"name": "train", "num_bytes": 4784289, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 9565445}, {"config_name": "nn", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 274140, "num_examples": 1000}, {"name": "test", "num_bytes": 269631, "num_examples": 1000}, {"name": "train", "num_bytes": 5436185, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 5979956}, {"config_name": "no", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2576669, "num_examples": 10000}, {"name": "test", "num_bytes": 2563559, "num_examples": 10000}, {"name": "train", "num_bytes": 5139548, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 10279776}, {"config_name": "nov", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 14856, "num_examples": 100}, {"name": "test", "num_bytes": 14830, "num_examples": 100}, {"name": "train", "num_bytes": 17270, "num_examples": 100}], "download_size": 234008884, "dataset_size": 46956}, {"config_name": "oc", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 20428, "num_examples": 100}, {"name": "test", "num_bytes": 18600, "num_examples": 100}, {"name": "train", "num_bytes": 19319, "num_examples": 100}], "download_size": 234008884, "dataset_size": 58347}, {"config_name": "or", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 32131, "num_examples": 100}, {"name": "test", "num_bytes": 29508, "num_examples": 100}, {"name": "train", "num_bytes": 27822, "num_examples": 100}], "download_size": 234008884, "dataset_size": 89461}, {"config_name": "os", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 26779, "num_examples": 100}, {"name": "test", "num_bytes": 25995, "num_examples": 100}, {"name": "train", "num_bytes": 26033, "num_examples": 100}], "download_size": 234008884, "dataset_size": 78807}, {"config_name": "pa", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 25230, "num_examples": 100}, {"name": "test", "num_bytes": 23708, "num_examples": 100}, {"name": "train", "num_bytes": 24171, "num_examples": 100}], "download_size": 234008884, "dataset_size": 73109}, {"config_name": "pdc", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 24419, "num_examples": 100}, {"name": "test", "num_bytes": 24674, "num_examples": 100}, {"name": "train", "num_bytes": 23991, "num_examples": 100}], "download_size": 234008884, "dataset_size": 73084}, {"config_name": "pl", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2448324, "num_examples": 10000}, {"name": "test", "num_bytes": 2463783, "num_examples": 10000}, {"name": "train", "num_bytes": 4851527, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 9763634}, {"config_name": "pms", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 28369, "num_examples": 100}, {"name": "test", "num_bytes": 24015, "num_examples": 100}, {"name": "train", "num_bytes": 27429, "num_examples": 100}], "download_size": 234008884, "dataset_size": 79813}, {"config_name": "pnb", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 19070, "num_examples": 100}, {"name": "test", "num_bytes": 21206, "num_examples": 100}, {"name": "train", "num_bytes": 19504, "num_examples": 100}], "download_size": 234008884, "dataset_size": 59780}, {"config_name": "ps", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 49901, "num_examples": 100}, {"name": "test", "num_bytes": 43621, "num_examples": 100}, {"name": "train", "num_bytes": 63501, "num_examples": 100}], "download_size": 234008884, "dataset_size": 157023}, {"config_name": "pt", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 1962145, "num_examples": 10000}, {"name": "test", "num_bytes": 1946729, "num_examples": 10000}, {"name": "train", "num_bytes": 3917453, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 7826327}, {"config_name": "qu", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 18231, "num_examples": 100}, {"name": "test", "num_bytes": 17675, "num_examples": 100}, {"name": "train", "num_bytes": 16989, "num_examples": 100}], "download_size": 234008884, "dataset_size": 52895}, {"config_name": "rm", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 32776, "num_examples": 100}, {"name": "test", "num_bytes": 35880, "num_examples": 100}, {"name": "train", "num_bytes": 30489, "num_examples": 100}], "download_size": 234008884, "dataset_size": 99145}, {"config_name": "ro", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2063860, "num_examples": 10000}, {"name": "test", "num_bytes": 2060933, "num_examples": 10000}, {"name": "train", "num_bytes": 4179869, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 8304662}, {"config_name": "ru", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2574546, "num_examples": 10000}, {"name": "test", "num_bytes": 2597248, "num_examples": 10000}, {"name": "train", "num_bytes": 5175665, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 10347459}, {"config_name": "rw", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 17999, "num_examples": 100}, {"name": "test", "num_bytes": 14445, "num_examples": 100}, {"name": "train", "num_bytes": 16778, "num_examples": 100}], "download_size": 234008884, "dataset_size": 49222}, {"config_name": "sa", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 45721, "num_examples": 100}, {"name": "test", "num_bytes": 49209, "num_examples": 100}, {"name": "train", "num_bytes": 52504, "num_examples": 100}], "download_size": 234008884, "dataset_size": 147434}, {"config_name": "sah", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 27875, "num_examples": 100}, {"name": "test", "num_bytes": 26853, "num_examples": 100}, {"name": "train", "num_bytes": 27041, "num_examples": 100}], "download_size": 234008884, "dataset_size": 81769}, {"config_name": "scn", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 20105, "num_examples": 100}, {"name": "test", "num_bytes": 17384, "num_examples": 100}, {"name": "train", "num_bytes": 21032, "num_examples": 100}], "download_size": 234008884, "dataset_size": 58521}, {"config_name": "sco", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 22215, "num_examples": 100}, {"name": "test", "num_bytes": 21589, "num_examples": 100}, {"name": "train", "num_bytes": 20308, "num_examples": 100}], "download_size": 234008884, "dataset_size": 64112}, {"config_name": "sd", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 51555, "num_examples": 100}, {"name": "test", "num_bytes": 38534, "num_examples": 100}, {"name": "train", "num_bytes": 56925, "num_examples": 100}], "download_size": 234008884, "dataset_size": 147014}, {"config_name": "sh", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 1789918, "num_examples": 10000}, {"name": "test", "num_bytes": 1791491, "num_examples": 10000}, {"name": "train", "num_bytes": 3583633, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 7165042}, {"config_name": "si", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 30845, "num_examples": 100}, {"name": "test", "num_bytes": 29341, "num_examples": 100}, {"name": "train", "num_bytes": 31255, "num_examples": 100}], "download_size": 234008884, "dataset_size": 91441}, {"config_name": "simple", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 247147, "num_examples": 1000}, {"name": "test", "num_bytes": 245358, "num_examples": 1000}, {"name": "train", "num_bytes": 4921916, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 5414421}, {"config_name": "sk", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2342061, "num_examples": 10000}, {"name": "test", "num_bytes": 2335009, "num_examples": 10000}, {"name": "train", "num_bytes": 4701553, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 9378623}, {"config_name": "sl", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2090247, "num_examples": 10000}, {"name": "test", "num_bytes": 2133491, "num_examples": 10000}, {"name": "train", "num_bytes": 3158676, "num_examples": 15000}], "download_size": 234008884, "dataset_size": 7382414}, {"config_name": "so", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 21864, "num_examples": 100}, {"name": "test", "num_bytes": 17219, "num_examples": 100}, {"name": "train", "num_bytes": 23780, "num_examples": 100}], "download_size": 234008884, "dataset_size": 62863}, {"config_name": "sq", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 210888, "num_examples": 1000}, {"name": "test", "num_bytes": 209824, "num_examples": 1000}, {"name": "train", "num_bytes": 1052387, "num_examples": 5000}], "download_size": 234008884, "dataset_size": 1473099}, {"config_name": "sr", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2548390, "num_examples": 10000}, {"name": "test", "num_bytes": 2564831, "num_examples": 10000}, {"name": "train", "num_bytes": 5105569, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 10218790}, {"config_name": "su", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 22605, "num_examples": 100}, {"name": "test", "num_bytes": 21861, "num_examples": 100}, {"name": "train", "num_bytes": 20839, "num_examples": 100}], "download_size": 234008884, "dataset_size": 65305}, {"config_name": "sv", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2678672, "num_examples": 10000}, {"name": "test", "num_bytes": 2719077, "num_examples": 10000}, {"name": "train", "num_bytes": 5395722, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 10793471}, {"config_name": "sw", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 168819, "num_examples": 1000}, {"name": "test", "num_bytes": 172693, "num_examples": 1000}, {"name": "train", "num_bytes": 168749, "num_examples": 1000}], "download_size": 234008884, "dataset_size": 510261}, {"config_name": "szl", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 19397, "num_examples": 100}, {"name": "test", "num_bytes": 18967, "num_examples": 100}, {"name": "train", "num_bytes": 17646, "num_examples": 100}], "download_size": 234008884, "dataset_size": 56010}, {"config_name": "ta", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 354957, "num_examples": 1000}, {"name": "test", "num_bytes": 357667, "num_examples": 1000}, {"name": "train", "num_bytes": 5275759, "num_examples": 15000}], "download_size": 234008884, "dataset_size": 5988383}, {"config_name": "te", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 356189, "num_examples": 1000}, {"name": "test", "num_bytes": 359780, "num_examples": 1000}, {"name": "train", "num_bytes": 358792, "num_examples": 1000}], "download_size": 234008884, "dataset_size": 1074761}, {"config_name": "tg", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 27130, "num_examples": 100}, {"name": "test", "num_bytes": 28821, "num_examples": 100}, {"name": "train", "num_bytes": 27200, "num_examples": 100}], "download_size": 234008884, "dataset_size": 83151}, {"config_name": "th", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 14189743, "num_examples": 10000}, {"name": "test", "num_bytes": 14505054, "num_examples": 10000}, {"name": "train", "num_bytes": 28968916, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 57663713}, {"config_name": "tk", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 21611, "num_examples": 100}, {"name": "test", "num_bytes": 20302, "num_examples": 100}, {"name": "train", "num_bytes": 19521, "num_examples": 100}], "download_size": 234008884, "dataset_size": 61434}, {"config_name": "tl", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 148682, "num_examples": 1000}, {"name": "test", "num_bytes": 152964, "num_examples": 1000}, {"name": "train", "num_bytes": 1518784, "num_examples": 10000}], "download_size": 234008884, "dataset_size": 1820430}, {"config_name": "tr", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2280517, "num_examples": 10000}, {"name": "test", "num_bytes": 2276920, "num_examples": 10000}, {"name": "train", "num_bytes": 4501912, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 9059349}, {"config_name": "tt", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 282535, "num_examples": 1000}, {"name": "test", "num_bytes": 282691, "num_examples": 1000}, {"name": "train", "num_bytes": 283392, "num_examples": 1000}], "download_size": 234008884, "dataset_size": 848618}, {"config_name": "ug", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 35219, "num_examples": 100}, {"name": "test", "num_bytes": 31129, "num_examples": 100}, {"name": "train", "num_bytes": 26620, "num_examples": 100}], "download_size": 234008884, "dataset_size": 92968}, {"config_name": "uk", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 2934897, "num_examples": 10000}, {"name": "test", "num_bytes": 2928200, "num_examples": 10000}, {"name": "train", "num_bytes": 5928026, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 11791123}, {"config_name": "ur", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 203747, "num_examples": 1000}, {"name": "test", "num_bytes": 203138, "num_examples": 1000}, {"name": "train", "num_bytes": 4108707, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 4515592}, {"config_name": "uz", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 184625, "num_examples": 1000}, {"name": "test", "num_bytes": 184713, "num_examples": 1000}, {"name": "train", "num_bytes": 186105, "num_examples": 1000}], "download_size": 234008884, "dataset_size": 555443}, {"config_name": "vec", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 19335, "num_examples": 100}, {"name": "test", "num_bytes": 20254, "num_examples": 100}, {"name": "train", "num_bytes": 20437, "num_examples": 100}], "download_size": 234008884, "dataset_size": 60026}, {"config_name": "vep", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 22306, "num_examples": 100}, {"name": "test", "num_bytes": 21371, "num_examples": 100}, {"name": "train", "num_bytes": 21387, "num_examples": 100}], "download_size": 234008884, "dataset_size": 65064}, {"config_name": "vi", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 1944856, "num_examples": 10000}, {"name": "test", "num_bytes": 1960024, "num_examples": 10000}, {"name": "train", "num_bytes": 3915944, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 7820824}, {"config_name": "vls", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 27895, "num_examples": 100}, {"name": "test", "num_bytes": 26778, "num_examples": 100}, {"name": "train", "num_bytes": 26183, "num_examples": 100}], "download_size": 234008884, "dataset_size": 80856}, {"config_name": "vo", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 14385, "num_examples": 100}, {"name": "test", "num_bytes": 14001, "num_examples": 100}, {"name": "train", "num_bytes": 14442, "num_examples": 100}], "download_size": 234008884, "dataset_size": 42828}, {"config_name": "wa", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 22493, "num_examples": 100}, {"name": "test", "num_bytes": 21581, "num_examples": 100}, {"name": "train", "num_bytes": 23072, "num_examples": 100}], "download_size": 234008884, "dataset_size": 67146}, {"config_name": "war", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 16834, "num_examples": 100}, {"name": "test", "num_bytes": 19912, "num_examples": 100}, {"name": "train", "num_bytes": 18829, "num_examples": 100}], "download_size": 234008884, "dataset_size": 55575}, {"config_name": "wuu", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 15123, "num_examples": 100}, {"name": "test", "num_bytes": 15067, "num_examples": 100}, {"name": "train", "num_bytes": 17016, "num_examples": 100}], "download_size": 234008884, "dataset_size": 47206}, {"config_name": "xmf", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 39979, "num_examples": 100}, {"name": "test", "num_bytes": 36081, "num_examples": 100}, {"name": "train", "num_bytes": 31796, "num_examples": 100}], "download_size": 234008884, "dataset_size": 107856}, {"config_name": "yi", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 25269, "num_examples": 100}, {"name": "test", "num_bytes": 25005, "num_examples": 100}, {"name": "train", "num_bytes": 27303, "num_examples": 100}], "download_size": 234008884, "dataset_size": 77577}, {"config_name": "yo", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 17738, "num_examples": 100}, {"name": "test", "num_bytes": 17996, "num_examples": 100}, {"name": "train", "num_bytes": 18984, "num_examples": 100}], "download_size": 234008884, "dataset_size": 54718}, {"config_name": "zea", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 24916, "num_examples": 100}, {"name": "test", "num_bytes": 22997, "num_examples": 100}, {"name": "train", "num_bytes": 21252, "num_examples": 100}], "download_size": 234008884, "dataset_size": 69165}, {"config_name": "zh", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 4839728, "num_examples": 10000}, {"name": "test", "num_bytes": 4709458, "num_examples": 10000}, {"name": "train", "num_bytes": 9524981, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 19074167}, {"config_name": "zh-classical", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 59980, "num_examples": 100}, {"name": "test", "num_bytes": 65885, "num_examples": 100}, {"name": "train", "num_bytes": 56238, "num_examples": 100}], "download_size": 234008884, "dataset_size": 182103}, {"config_name": "zh-min-nan", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 24533, "num_examples": 100}, {"name": "test", "num_bytes": 24326, "num_examples": 100}, {"name": "train", "num_bytes": 19358, "num_examples": 100}], "download_size": 234008884, "dataset_size": 68217}, {"config_name": "zh-yue", "features": [{"name": "tokens", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "O", "1": "B-PER", "2": "I-PER", "3": "B-ORG", "4": "I-ORG", "5": "B-LOC", "6": "I-LOC"}}}}, {"name": "langs", "sequence": "string"}, {"name": "spans", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 4934158, "num_examples": 10000}, {"name": "test", "num_bytes": 4964029, "num_examples": 10000}, {"name": "train", "num_bytes": 9950629, "num_examples": 20000}], "download_size": 234008884, "dataset_size": 19848816}]}
2024-01-18T11:18:13+00:00
[ "1902.00193" ]
[ "ace", "af", "als", "am", "an", "ang", "ar", "arc", "arz", "as", "ast", "ay", "az", "ba", "bar", "be", "bg", "bh", "bn", "bo", "br", "bs", "ca", "cbk", "cdo", "ce", "ceb", "ckb", "co", "crh", "cs", "csb", "cv", "cy", "da", "de", "diq", "dv", "el", "eml", "en", "eo", "es", "et", "eu", "ext", "fa", "fi", "fo", "fr", "frr", "fur", "fy", "ga", "gan", "gd", "gl", "gn", "gu", "hak", "he", "hi", "hr", "hsb", "hu", "hy", "ia", "id", "ig", "ilo", "io", "is", "it", "ja", "jbo", "jv", "ka", "kk", "km", "kn", "ko", "ksh", "ku", "ky", "la", "lb", "li", "lij", "lmo", "ln", "lt", "lv", "lzh", "mg", "mhr", "mi", "min", "mk", "ml", "mn", "mr", "ms", "mt", "mwl", "my", "mzn", "nan", "nap", "nds", "ne", "nl", "nn", "no", "nov", "oc", "or", "os", "pa", "pdc", "pl", "pms", "pnb", "ps", "pt", "qu", "rm", "ro", "ru", "rw", "sa", "sah", "scn", "sco", "sd", "sgs", "sh", "si", "sk", "sl", "so", "sq", "sr", "su", "sv", "sw", "szl", "ta", "te", "tg", "th", "tk", "tl", "tr", "tt", "ug", "uk", "ur", "uz", "vec", "vep", "vi", "vls", "vo", "vro", "wa", "war", "wuu", "xmf", "yi", "yo", "yue", "zea", "zh" ]
TAGS #task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-machine-generated #language_creators-crowdsourced #multilinguality-multilingual #size_categories-n<1K #source_datasets-original #language-Achinese #language-Afrikaans #language-Tosk Albanian #language-Amharic #language-Aragonese #language-Old English (ca. 450-1100) #language-Arabic #language-Official Aramaic (700-300 BCE) #language-Egyptian Arabic #language-Assamese #language-Asturian #language-Aymara #language-Azerbaijani #language-Bashkir #language-Bavarian #language-Belarusian #language-Bulgarian #language-bh #language-Bengali #language-Tibetan #language-Breton #language-Bosnian #language-Catalan #language-Chavacano #language-Min Dong Chinese #language-Chechen #language-Cebuano #language-Central Kurdish #language-Corsican #language-Crimean Tatar #language-Czech #language-Kashubian #language-Chuvash #language-Welsh #language-Danish #language-German #language-Dimli (individual language) #language-Dhivehi #language-Modern Greek (1453-) #language-Emiliano-Romagnolo #language-English #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Extremaduran #language-Persian #language-Finnish #language-Faroese #language-French #language-Northern Frisian #language-Friulian #language-Western Frisian #language-Irish #language-Gan Chinese #language-Scottish Gaelic #language-Galician #language-Guarani #language-Gujarati #language-Hakka Chinese #language-Hebrew #language-Hindi #language-Croatian #language-Upper Sorbian #language-Hungarian #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Indonesian #language-Igbo #language-Iloko #language-Ido #language-Icelandic #language-Italian #language-Japanese #language-Lojban #language-Javanese #language-Georgian #language-Kazakh #language-Khmer #language-Kannada #language-Korean #language-Kölsch #language-Kurdish #language-Kirghiz #language-Latin #language-Luxembourgish #language-Limburgan #language-Ligurian #language-Lombard #language-Lingala #language-Lithuanian #language-Latvian #language-Literary Chinese #language-Malagasy #language-Eastern Mari #language-Maori #language-Minangkabau #language-Macedonian #language-Malayalam #language-Mongolian #language-Marathi #language-Malay (macrolanguage) #language-Maltese #language-Mirandese #language-Burmese #language-Mazanderani #language-Min Nan Chinese #language-Neapolitan #language-Low German #language-Nepali (macrolanguage) #language-Dutch #language-Norwegian Nynorsk #language-Norwegian #language-Novial #language-Occitan (post 1500) #language-Oriya (macrolanguage) #language-Ossetian #language-Panjabi #language-Pennsylvania German #language-Polish #language-Piemontese #language-Western Panjabi #language-Pushto #language-Portuguese #language-Quechua #language-Romansh #language-Romanian #language-Russian #language-Kinyarwanda #language-Sanskrit #language-Yakut #language-Sicilian #language-Scots #language-Sindhi #language-Samogitian #language-Serbo-Croatian #language-Sinhala #language-Slovak #language-Slovenian #language-Somali #language-Albanian #language-Serbian #language-Sundanese #language-Swedish #language-Swahili (macrolanguage) #language-Silesian #language-Tamil #language-Telugu #language-Tajik #language-Thai #language-Turkmen #language-Tagalog #language-Turkish #language-Tatar #language-Uighur #language-Ukrainian #language-Urdu #language-Uzbek #language-Venetian #language-Veps #language-Vietnamese #language-Vlaams #language-Volapük #language-Võro #language-Walloon #language-Waray (Philippines) #language-Wu Chinese #language-Mingrelian #language-Yiddish #language-Yoruba #language-Yue Chinese #language-Zeeuws #language-Chinese #license-unknown #arxiv-1902.00193 #region-us
Dataset Card for WikiANN ======================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: Massively Multilingual Transfer for NER * Repository: Massively Multilingual Transfer for NER * Paper: The original datasets come from the *Cross-lingual name tagging and linking for 282 languages* paper by Xiaoman Pan et al. (2018). This version corresponds to the balanced train, dev, and test splits of the original data from the *Massively Multilingual Transfer for NER* paper by Afshin Rahimi et al. (2019). * Leaderboard: * Point of Contact: Afshin Rahimi or Lewis Tunstall or Albert Villanova del Moral ### Dataset Summary WikiANN (sometimes called PAN-X) is a multilingual named entity recognition dataset consisting of Wikipedia articles annotated with LOC (location), PER (person), and ORG (organisation) tags in the IOB2 format. This version corresponds to the balanced train, dev, and test splits of Rahimi et al. (2019), which supports 176 of the 282 languages from the original WikiANN corpus. ### Supported Tasks and Leaderboards * 'named-entity-recognition': The dataset can be used to train a model for named entity recognition in many languages, or evaluate the zero-shot cross-lingual capabilities of multilingual models. ### Languages The dataset contains 176 languages, one in each of the configuration subsets. The corresponding BCP 47 language tags are: Dataset Structure ----------------- ### Data Instances This is an example in the "train" split of the "af" (Afrikaans language) configuration subset: ### Data Fields * 'tokens': a 'list' of 'string' features. * 'langs': a 'list' of 'string' features that correspond to the language of each token. * 'ner\_tags': a 'list' of classification labels, with possible values including 'O' (0), 'B-PER' (1), 'I-PER' (2), 'B-ORG' (3), 'I-ORG' (4), 'B-LOC' (5), 'I-LOC' (6). * 'spans': a 'list' of 'string' features, that is the list of named entities in the input text formatted as '': '' ### Data Splits For each configuration subset, the data is split into "train", "validation" and "test" sets, each containing the following number of examples: Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information The original 282 datasets are associated with this article while the 176 languages supported in this version are associated with the following article ### Contributions Thanks to @lewtun and @rabeehk for adding this dataset.
[ "### Dataset Summary\n\n\nWikiANN (sometimes called PAN-X) is a multilingual named entity recognition dataset consisting of Wikipedia articles annotated with LOC (location), PER (person), and ORG (organisation) tags in the IOB2 format. This version corresponds to the balanced train, dev, and test splits of Rahimi et al. (2019), which supports 176 of the 282 languages from the original WikiANN corpus.", "### Supported Tasks and Leaderboards\n\n\n* 'named-entity-recognition': The dataset can be used to train a model for named entity recognition in many languages, or evaluate the zero-shot cross-lingual capabilities of multilingual models.", "### Languages\n\n\nThe dataset contains 176 languages, one in each of the configuration subsets. The corresponding BCP 47 language tags\nare:\n\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nThis is an example in the \"train\" split of the \"af\" (Afrikaans language) configuration subset:", "### Data Fields\n\n\n* 'tokens': a 'list' of 'string' features.\n* 'langs': a 'list' of 'string' features that correspond to the language of each token.\n* 'ner\\_tags': a 'list' of classification labels, with possible values including 'O' (0), 'B-PER' (1), 'I-PER' (2), 'B-ORG' (3), 'I-ORG' (4), 'B-LOC' (5), 'I-LOC' (6).\n* 'spans': a 'list' of 'string' features, that is the list of named entities in the input text formatted as '': ''", "### Data Splits\n\n\nFor each configuration subset, the data is split into \"train\", \"validation\" and \"test\" sets, each containing the\nfollowing number of examples:\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nThe original 282 datasets are associated with this article\n\n\nwhile the 176 languages supported in this version are associated with the following article", "### Contributions\n\n\nThanks to @lewtun and @rabeehk for adding this dataset." ]
[ "TAGS\n#task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-machine-generated #language_creators-crowdsourced #multilinguality-multilingual #size_categories-n<1K #source_datasets-original #language-Achinese #language-Afrikaans #language-Tosk Albanian #language-Amharic #language-Aragonese #language-Old English (ca. 450-1100) #language-Arabic #language-Official Aramaic (700-300 BCE) #language-Egyptian Arabic #language-Assamese #language-Asturian #language-Aymara #language-Azerbaijani #language-Bashkir #language-Bavarian #language-Belarusian #language-Bulgarian #language-bh #language-Bengali #language-Tibetan #language-Breton #language-Bosnian #language-Catalan #language-Chavacano #language-Min Dong Chinese #language-Chechen #language-Cebuano #language-Central Kurdish #language-Corsican #language-Crimean Tatar #language-Czech #language-Kashubian #language-Chuvash #language-Welsh #language-Danish #language-German #language-Dimli (individual language) #language-Dhivehi #language-Modern Greek (1453-) #language-Emiliano-Romagnolo #language-English #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Extremaduran #language-Persian #language-Finnish #language-Faroese #language-French #language-Northern Frisian #language-Friulian #language-Western Frisian #language-Irish #language-Gan Chinese #language-Scottish Gaelic #language-Galician #language-Guarani #language-Gujarati #language-Hakka Chinese #language-Hebrew #language-Hindi #language-Croatian #language-Upper Sorbian #language-Hungarian #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Indonesian #language-Igbo #language-Iloko #language-Ido #language-Icelandic #language-Italian #language-Japanese #language-Lojban #language-Javanese #language-Georgian #language-Kazakh #language-Khmer #language-Kannada #language-Korean #language-Kölsch #language-Kurdish #language-Kirghiz #language-Latin #language-Luxembourgish #language-Limburgan #language-Ligurian #language-Lombard #language-Lingala #language-Lithuanian #language-Latvian #language-Literary Chinese #language-Malagasy #language-Eastern Mari #language-Maori #language-Minangkabau #language-Macedonian #language-Malayalam #language-Mongolian #language-Marathi #language-Malay (macrolanguage) #language-Maltese #language-Mirandese #language-Burmese #language-Mazanderani #language-Min Nan Chinese #language-Neapolitan #language-Low German #language-Nepali (macrolanguage) #language-Dutch #language-Norwegian Nynorsk #language-Norwegian #language-Novial #language-Occitan (post 1500) #language-Oriya (macrolanguage) #language-Ossetian #language-Panjabi #language-Pennsylvania German #language-Polish #language-Piemontese #language-Western Panjabi #language-Pushto #language-Portuguese #language-Quechua #language-Romansh #language-Romanian #language-Russian #language-Kinyarwanda #language-Sanskrit #language-Yakut #language-Sicilian #language-Scots #language-Sindhi #language-Samogitian #language-Serbo-Croatian #language-Sinhala #language-Slovak #language-Slovenian #language-Somali #language-Albanian #language-Serbian #language-Sundanese #language-Swedish #language-Swahili (macrolanguage) #language-Silesian #language-Tamil #language-Telugu #language-Tajik #language-Thai #language-Turkmen #language-Tagalog #language-Turkish #language-Tatar #language-Uighur #language-Ukrainian #language-Urdu #language-Uzbek #language-Venetian #language-Veps #language-Vietnamese #language-Vlaams #language-Volapük #language-Võro #language-Walloon #language-Waray (Philippines) #language-Wu Chinese #language-Mingrelian #language-Yiddish #language-Yoruba #language-Yue Chinese #language-Zeeuws #language-Chinese #license-unknown #arxiv-1902.00193 #region-us \n", "### Dataset Summary\n\n\nWikiANN (sometimes called PAN-X) is a multilingual named entity recognition dataset consisting of Wikipedia articles annotated with LOC (location), PER (person), and ORG (organisation) tags in the IOB2 format. This version corresponds to the balanced train, dev, and test splits of Rahimi et al. (2019), which supports 176 of the 282 languages from the original WikiANN corpus.", "### Supported Tasks and Leaderboards\n\n\n* 'named-entity-recognition': The dataset can be used to train a model for named entity recognition in many languages, or evaluate the zero-shot cross-lingual capabilities of multilingual models.", "### Languages\n\n\nThe dataset contains 176 languages, one in each of the configuration subsets. The corresponding BCP 47 language tags\nare:\n\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nThis is an example in the \"train\" split of the \"af\" (Afrikaans language) configuration subset:", "### Data Fields\n\n\n* 'tokens': a 'list' of 'string' features.\n* 'langs': a 'list' of 'string' features that correspond to the language of each token.\n* 'ner\\_tags': a 'list' of classification labels, with possible values including 'O' (0), 'B-PER' (1), 'I-PER' (2), 'B-ORG' (3), 'I-ORG' (4), 'B-LOC' (5), 'I-LOC' (6).\n* 'spans': a 'list' of 'string' features, that is the list of named entities in the input text formatted as '': ''", "### Data Splits\n\n\nFor each configuration subset, the data is split into \"train\", \"validation\" and \"test\" sets, each containing the\nfollowing number of examples:\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nThe original 282 datasets are associated with this article\n\n\nwhile the 176 languages supported in this version are associated with the following article", "### Contributions\n\n\nThanks to @lewtun and @rabeehk for adding this dataset." ]
[ 1169, 104, 63, 40, 31, 150, 48, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 34, 22 ]
[ "passage: ", "passage: TAGS\n#task_categories-token-classification #task_ids-named-entity-recognition #annotations_creators-machine-generated #language_creators-crowdsourced #multilinguality-multilingual #size_categories-n<1K #source_datasets-original #language-Achinese #language-Afrikaans #language-Tosk Albanian #language-Amharic #language-Aragonese #language-Old English (ca. 450-1100) #language-Arabic #language-Official Aramaic (700-300 BCE) #language-Egyptian Arabic #language-Assamese #language-Asturian #language-Aymara #language-Azerbaijani #language-Bashkir #language-Bavarian #language-Belarusian #language-Bulgarian #language-bh #language-Bengali #language-Tibetan #language-Breton #language-Bosnian #language-Catalan #language-Chavacano #language-Min Dong Chinese #language-Chechen #language-Cebuano #language-Central Kurdish #language-Corsican #language-Crimean Tatar #language-Czech #language-Kashubian #language-Chuvash #language-Welsh #language-Danish #language-German #language-Dimli (individual language) #language-Dhivehi #language-Modern Greek (1453-) #language-Emiliano-Romagnolo #language-English #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Extremaduran #language-Persian #language-Finnish #language-Faroese #language-French #language-Northern Frisian #language-Friulian #language-Western Frisian #language-Irish #language-Gan Chinese #language-Scottish Gaelic #language-Galician #language-Guarani #language-Gujarati #language-Hakka Chinese #language-Hebrew #language-Hindi #language-Croatian #language-Upper Sorbian #language-Hungarian #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Indonesian #language-Igbo #language-Iloko #language-Ido #language-Icelandic #language-Italian #language-Japanese #language-Lojban #language-Javanese #language-Georgian #language-Kazakh #language-Khmer #language-Kannada #language-Korean #language-Kölsch #language-Kurdish #language-Kirghiz #language-Latin #language-Luxembourgish #language-Limburgan #language-Ligurian #language-Lombard #language-Lingala #language-Lithuanian #language-Latvian #language-Literary Chinese #language-Malagasy #language-Eastern Mari #language-Maori #language-Minangkabau #language-Macedonian #language-Malayalam #language-Mongolian #language-Marathi #language-Malay (macrolanguage) #language-Maltese #language-Mirandese #language-Burmese #language-Mazanderani #language-Min Nan Chinese #language-Neapolitan #language-Low German #language-Nepali (macrolanguage) #language-Dutch #language-Norwegian Nynorsk #language-Norwegian #language-Novial #language-Occitan (post 1500) #language-Oriya (macrolanguage) #language-Ossetian #language-Panjabi #language-Pennsylvania German #language-Polish #language-Piemontese #language-Western Panjabi #language-Pushto #language-Portuguese #language-Quechua #language-Romansh #language-Romanian #language-Russian #language-Kinyarwanda #language-Sanskrit #language-Yakut #language-Sicilian #language-Scots #language-Sindhi #language-Samogitian #language-Serbo-Croatian #language-Sinhala #language-Slovak #language-Slovenian #language-Somali #language-Albanian #language-Serbian #language-Sundanese #language-Swedish #language-Swahili (macrolanguage) #language-Silesian #language-Tamil #language-Telugu #language-Tajik #language-Thai #language-Turkmen #language-Tagalog #language-Turkish #language-Tatar #language-Uighur #language-Ukrainian #language-Urdu #language-Uzbek #language-Venetian #language-Veps #language-Vietnamese #language-Vlaams #language-Volapük #language-Võro #language-Walloon #language-Waray (Philippines) #language-Wu Chinese #language-Mingrelian #language-Yiddish #language-Yoruba #language-Yue Chinese #language-Zeeuws #language-Chinese #license-unknown #arxiv-1902.00193 #region-us \n### Dataset Summary\n\n\nWikiANN (sometimes called PAN-X) is a multilingual named entity recognition dataset consisting of Wikipedia articles annotated with LOC (location), PER (person), and ORG (organisation) tags in the IOB2 format. This version corresponds to the balanced train, dev, and test splits of Rahimi et al. (2019), which supports 176 of the 282 languages from the original WikiANN corpus.### Supported Tasks and Leaderboards\n\n\n* 'named-entity-recognition': The dataset can be used to train a model for named entity recognition in many languages, or evaluate the zero-shot cross-lingual capabilities of multilingual models.### Languages\n\n\nThe dataset contains 176 languages, one in each of the configuration subsets. The corresponding BCP 47 language tags\nare:\n\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nThis is an example in the \"train\" split of the \"af\" (Afrikaans language) configuration subset:### Data Fields\n\n\n* 'tokens': a 'list' of 'string' features.\n* 'langs': a 'list' of 'string' features that correspond to the language of each token.\n* 'ner\\_tags': a 'list' of classification labels, with possible values including 'O' (0), 'B-PER' (1), 'I-PER' (2), 'B-ORG' (3), 'I-ORG' (4), 'B-LOC' (5), 'I-LOC' (6).\n* 'spans': a 'list' of 'string' features, that is the list of named entities in the input text formatted as '': ''### Data Splits\n\n\nFor each configuration subset, the data is split into \"train\", \"validation\" and \"test\" sets, each containing the\nfollowing number of examples:\n\n\n\nDataset Creation\n----------------### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------" ]
5eef3045e6e2eb495f5d3006122a46cfb6bd843c
# Dataset Card for Wikicorpus ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.cs.upc.edu/~nlp/wikicorpus/ - **Repository:** - **Paper:** https://www.cs.upc.edu/~nlp/papers/reese10.pdf - **Leaderboard:** - **Point of Contact:** ### Dataset Summary The Wikicorpus is a trilingual corpus (Catalan, Spanish, English) that contains large portions of the Wikipedia (based on a 2006 dump) and has been automatically enriched with linguistic information. In its present version, it contains over 750 million words. The corpora have been annotated with lemma and part of speech information using the open source library FreeLing. Also, they have been sense annotated with the state of the art Word Sense Disambiguation algorithm UKB. As UKB assigns WordNet senses, and WordNet has been aligned across languages via the InterLingual Index, this sort of annotation opens the way to massive explorations in lexical semantics that were not possible before. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages Each sub-dataset is monolingual in the languages: - ca: Catalan - en: English - es: Spanish ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information The WikiCorpus is licensed under the same license as Wikipedia, that is, the [GNU Free Documentation License](http://www.fsf.org/licensing/licenses/fdl.html) ### Citation Information ``` @inproceedings{reese-etal-2010-wikicorpus, title = "{W}ikicorpus: A Word-Sense Disambiguated Multilingual {W}ikipedia Corpus", author = "Reese, Samuel and Boleda, Gemma and Cuadros, Montse and Padr{\'o}, Llu{\'i}s and Rigau, German", booktitle = "Proceedings of the Seventh International Conference on Language Resources and Evaluation ({LREC}'10)", month = may, year = "2010", address = "Valletta, Malta", publisher = "European Language Resources Association (ELRA)", url = "http://www.lrec-conf.org/proceedings/lrec2010/pdf/222_Paper.pdf", abstract = "This article presents a new freely available trilingual corpus (Catalan, Spanish, English) that contains large portions of the Wikipedia and has been automatically enriched with linguistic information. To our knowledge, this is the largest such corpus that is freely available to the community: In its present version, it contains over 750 million words. The corpora have been annotated with lemma and part of speech information using the open source library FreeLing. Also, they have been sense annotated with the state of the art Word Sense Disambiguation algorithm UKB. As UKB assigns WordNet senses, and WordNet has been aligned across languages via the InterLingual Index, this sort of annotation opens the way to massive explorations in lexical semantics that were not possible before. We present a first attempt at creating a trilingual lexical resource from the sense-tagged Wikipedia corpora, namely, WikiNet. Moreover, we present two by-products of the project that are of use for the NLP community: An open source Java-based parser for Wikipedia pages developed for the construction of the corpus, and the integration of the WSD algorithm UKB in FreeLing.", } ``` ### Contributions Thanks to [@albertvillanova](https://github.com/albertvillanova) for adding this dataset.
wikicorpus
[ "task_categories:fill-mask", "task_categories:text-classification", "task_categories:text-generation", "task_categories:token-classification", "task_ids:language-modeling", "task_ids:masked-language-modeling", "task_ids:part-of-speech", "annotations_creators:machine-generated", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "size_categories:10M<n<100M", "size_categories:1M<n<10M", "source_datasets:original", "language:ca", "language:en", "language:es", "license:gfdl", "word-sense-disambiguation", "lemmatization", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["machine-generated", "no-annotation"], "language_creators": ["found"], "language": ["ca", "en", "es"], "license": ["gfdl"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M", "10M<n<100M", "1M<n<10M"], "source_datasets": ["original"], "task_categories": ["fill-mask", "text-classification", "text-generation", "token-classification"], "task_ids": ["language-modeling", "masked-language-modeling", "part-of-speech"], "pretty_name": "Wikicorpus", "config_names": ["raw_ca", "raw_en", "raw_es", "tagged_ca", "tagged_en", "tagged_es"], "tags": ["word-sense-disambiguation", "lemmatization"], "dataset_info": [{"config_name": "raw_ca", "features": [{"name": "id", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 263170192, "num_examples": 143883}], "download_size": 96437841, "dataset_size": 263170192}, {"config_name": "raw_es", "features": [{"name": "id", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 671295359, "num_examples": 259409}], "download_size": 252926918, "dataset_size": 671295359}, {"config_name": "raw_en", "features": [{"name": "id", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 3388801074, "num_examples": 1359146}], "download_size": 1346378932, "dataset_size": 3388801074}, {"config_name": "tagged_ca", "features": [{"name": "id", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "sentence", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "pos_tags", "sequence": "string"}, {"name": "wordnet_senses", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 1666129919, "num_examples": 2016221}], "download_size": 226390380, "dataset_size": 1666129919}, {"config_name": "tagged_es", "features": [{"name": "id", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "sentence", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "pos_tags", "sequence": "string"}, {"name": "wordnet_senses", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 4100040390, "num_examples": 5039367}], "download_size": 604910899, "dataset_size": 4100040390}, {"config_name": "tagged_en", "features": [{"name": "id", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "sentence", "sequence": "string"}, {"name": "lemmas", "sequence": "string"}, {"name": "pos_tags", "sequence": "string"}, {"name": "wordnet_senses", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 18077275300, "num_examples": 26350272}], "download_size": 2477450893, "dataset_size": 18077275300}]}
2024-01-18T11:18:14+00:00
[]
[ "ca", "en", "es" ]
TAGS #task_categories-fill-mask #task_categories-text-classification #task_categories-text-generation #task_categories-token-classification #task_ids-language-modeling #task_ids-masked-language-modeling #task_ids-part-of-speech #annotations_creators-machine-generated #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #size_categories-10M<n<100M #size_categories-1M<n<10M #source_datasets-original #language-Catalan #language-English #language-Spanish #license-gfdl #word-sense-disambiguation #lemmatization #region-us
# Dataset Card for Wikicorpus ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: URL - Leaderboard: - Point of Contact: ### Dataset Summary The Wikicorpus is a trilingual corpus (Catalan, Spanish, English) that contains large portions of the Wikipedia (based on a 2006 dump) and has been automatically enriched with linguistic information. In its present version, it contains over 750 million words. The corpora have been annotated with lemma and part of speech information using the open source library FreeLing. Also, they have been sense annotated with the state of the art Word Sense Disambiguation algorithm UKB. As UKB assigns WordNet senses, and WordNet has been aligned across languages via the InterLingual Index, this sort of annotation opens the way to massive explorations in lexical semantics that were not possible before. ### Supported Tasks and Leaderboards ### Languages Each sub-dataset is monolingual in the languages: - ca: Catalan - en: English - es: Spanish ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information The WikiCorpus is licensed under the same license as Wikipedia, that is, the GNU Free Documentation License ### Contributions Thanks to @albertvillanova for adding this dataset.
[ "# Dataset Card for Wikicorpus", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper: URL\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nThe Wikicorpus is a trilingual corpus (Catalan, Spanish, English) that contains large portions of the Wikipedia (based on a 2006 dump) and has been automatically enriched with linguistic information. In its present version, it contains over 750 million words.\n\nThe corpora have been annotated with lemma and part of speech information using the open source library FreeLing. Also, they have been sense annotated with the state of the art Word Sense Disambiguation algorithm UKB. As UKB assigns WordNet senses, and WordNet has been aligned across languages via the InterLingual Index, this sort of annotation opens the way to massive explorations in lexical semantics that were not possible before.", "### Supported Tasks and Leaderboards", "### Languages\n\nEach sub-dataset is monolingual in the languages:\n- ca: Catalan\n- en: English\n- es: Spanish", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nThe WikiCorpus is licensed under the same license as Wikipedia, that is, the GNU Free Documentation License", "### Contributions\n\nThanks to @albertvillanova for adding this dataset." ]
[ "TAGS\n#task_categories-fill-mask #task_categories-text-classification #task_categories-text-generation #task_categories-token-classification #task_ids-language-modeling #task_ids-masked-language-modeling #task_ids-part-of-speech #annotations_creators-machine-generated #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #size_categories-10M<n<100M #size_categories-1M<n<10M #source_datasets-original #language-Catalan #language-English #language-Spanish #license-gfdl #word-sense-disambiguation #lemmatization #region-us \n", "# Dataset Card for Wikicorpus", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper: URL\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nThe Wikicorpus is a trilingual corpus (Catalan, Spanish, English) that contains large portions of the Wikipedia (based on a 2006 dump) and has been automatically enriched with linguistic information. In its present version, it contains over 750 million words.\n\nThe corpora have been annotated with lemma and part of speech information using the open source library FreeLing. Also, they have been sense annotated with the state of the art Word Sense Disambiguation algorithm UKB. As UKB assigns WordNet senses, and WordNet has been aligned across languages via the InterLingual Index, this sort of annotation opens the way to massive explorations in lexical semantics that were not possible before.", "### Supported Tasks and Leaderboards", "### Languages\n\nEach sub-dataset is monolingual in the languages:\n- ca: Catalan\n- en: English\n- es: Spanish", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators", "### Licensing Information\n\nThe WikiCorpus is licensed under the same license as Wikipedia, that is, the GNU Free Documentation License", "### Contributions\n\nThanks to @albertvillanova for adding this dataset." ]
[ 206, 8, 120, 26, 169, 10, 30, 6, 6, 5, 5, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 6, 29, 18 ]
[ "passage: TAGS\n#task_categories-fill-mask #task_categories-text-classification #task_categories-text-generation #task_categories-token-classification #task_ids-language-modeling #task_ids-masked-language-modeling #task_ids-part-of-speech #annotations_creators-machine-generated #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-100K<n<1M #size_categories-10M<n<100M #size_categories-1M<n<10M #source_datasets-original #language-Catalan #language-English #language-Spanish #license-gfdl #word-sense-disambiguation #lemmatization #region-us \n# Dataset Card for Wikicorpus## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository:\n- Paper: URL\n- Leaderboard:\n- Point of Contact:" ]
b236551f9954e110be00be867c56fbb874e09f09
### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@lewtun](https://github.com/lewtun), [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset.
wikihow
[ "region:us" ]
2022-03-02T23:29:22+00:00
{"paperswithcode_id": "wikihow", "pretty_name": "WikiHow", "dataset_info": [{"config_name": "all", "features": [{"name": "text", "dtype": "string"}, {"name": "headline", "dtype": "string"}, {"name": "title", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 513238309, "num_examples": 157252}, {"name": "validation", "num_bytes": 18246897, "num_examples": 5599}, {"name": "test", "num_bytes": 18276023, "num_examples": 5577}], "download_size": 5460385, "dataset_size": 549761229}, {"config_name": "sep", "features": [{"name": "text", "dtype": "string"}, {"name": "headline", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "overview", "dtype": "string"}, {"name": "sectionLabel", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 990499776, "num_examples": 1060732}, {"name": "validation", "num_bytes": 35173966, "num_examples": 37932}, {"name": "test", "num_bytes": 35271826, "num_examples": 37800}], "download_size": 5460385, "dataset_size": 1060945568}]}
2024-01-18T11:18:15+00:00
[]
[]
TAGS #region-us
### Contributions Thanks to @thomwolf, @lewtun, @patrickvonplaten for adding this dataset.
[ "### Contributions\n\nThanks to @thomwolf, @lewtun, @patrickvonplaten for adding this dataset." ]
[ "TAGS\n#region-us \n", "### Contributions\n\nThanks to @thomwolf, @lewtun, @patrickvonplaten for adding this dataset." ]
[ 6, 28 ]
[ "passage: TAGS\n#region-us \n### Contributions\n\nThanks to @thomwolf, @lewtun, @patrickvonplaten for adding this dataset." ]
4d013bdd32c475c8536aae00a56efc774f061649
# Dataset Card for Wikipedia ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://dumps.wikimedia.org](https://dumps.wikimedia.org) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Dataset Summary Wikipedia dataset containing cleaned articles of all languages. The datasets are built from the Wikipedia dump (https://dumps.wikimedia.org/) with one split per language. Each example contains the content of one full Wikipedia article with cleaning to strip markdown and unwanted sections (references, etc.). The articles are parsed using the ``mwparserfromhell`` tool. To load this dataset you need to install Apache Beam and ``mwparserfromhell`` first: ``` pip install apache_beam mwparserfromhell ``` Then, you can load any subset of Wikipedia per language and per date this way: ```python from datasets import load_dataset load_dataset("wikipedia", language="sw", date="20220120", beam_runner=...) ``` where you can pass as `beam_runner` any Apache Beam supported runner for (distributed) data processing (see [here](https://beam.apache.org/documentation/runners/capability-matrix/)). Pass "DirectRunner" to run it on your machine. You can find the full list of languages and dates [here](https://dumps.wikimedia.org/backup-index.html). Some subsets of Wikipedia have already been processed by HuggingFace, and you can load them just with: ```python from datasets import load_dataset load_dataset("wikipedia", "20220301.en") ``` The list of pre-processed subsets is: - "20220301.de" - "20220301.en" - "20220301.fr" - "20220301.frr" - "20220301.it" - "20220301.simple" ### Supported Tasks and Leaderboards The dataset is generally used for Language Modeling. ### Languages You can find the list of languages [here](https://meta.wikimedia.org/wiki/List_of_Wikipedias). ## Dataset Structure ### Data Instances An example looks as follows: ``` {'id': '1', 'url': 'https://simple.wikipedia.org/wiki/April', 'title': 'April', 'text': 'April is the fourth month...' } ``` Some subsets of Wikipedia have already been processed by HuggingFace, as you can see below: #### 20220301.de - **Size of downloaded dataset files:** 6.84 GB - **Size of the generated dataset:** 9.34 GB - **Total amount of disk used:** 16.18 GB #### 20220301.en - **Size of downloaded dataset files:** 21.60 GB - **Size of the generated dataset:** 21.26 GB - **Total amount of disk used:** 42.86 GB #### 20220301.fr - **Size of downloaded dataset files:** 5.87 GB - **Size of the generated dataset:** 7.73 GB - **Total amount of disk used:** 13.61 GB #### 20220301.frr - **Size of downloaded dataset files:** 13.04 MB - **Size of the generated dataset:** 9.57 MB - **Total amount of disk used:** 22.62 MB #### 20220301.it - **Size of downloaded dataset files:** 3.69 GB - **Size of the generated dataset:** 4.76 GB - **Total amount of disk used:** 8.45 GB #### 20220301.simple - **Size of downloaded dataset files:** 251.32 MB - **Size of the generated dataset:** 246.49 MB - **Total amount of disk used:** 497.82 MB ### Data Fields The data fields are the same among all configurations: - `id` (`str`): ID of the article. - `url` (`str`): URL of the article. - `title` (`str`): Title of the article. - `text` (`str`): Text content of the article. ### Data Splits Here are the number of examples for several configurations: | name | train | |-----------------|--------:| | 20220301.de | 2665357 | | 20220301.en | 6458670 | | 20220301.fr | 2402095 | | 20220301.frr | 15199 | | 20220301.it | 1743035 | | 20220301.simple | 205328 | ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information Most of Wikipedia's text and many of its images are co-licensed under the [Creative Commons Attribution-ShareAlike 3.0 Unported License](https://en.wikipedia.org/wiki/Wikipedia:Text_of_Creative_Commons_Attribution-ShareAlike_3.0_Unported_License) (CC BY-SA) and the [GNU Free Documentation License](https://en.wikipedia.org/wiki/Wikipedia:Text_of_the_GNU_Free_Documentation_License) (GFDL) (unversioned, with no invariant sections, front-cover texts, or back-cover texts). Some text has been imported only under CC BY-SA and CC BY-SA-compatible license and cannot be reused under GFDL; such text will be identified on the page footer, in the page history, or on the discussion page of the article that utilizes the text. ### Citation Information ``` @ONLINE{wikidump, author = "Wikimedia Foundation", title = "Wikimedia Downloads", url = "https://dumps.wikimedia.org" } ``` ### Contributions Thanks to [@lewtun](https://github.com/lewtun), [@mariamabarham](https://github.com/mariamabarham), [@thomwolf](https://github.com/thomwolf), [@lhoestq](https://github.com/lhoestq), [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset.
wikipedia
[ "task_categories:text-generation", "task_categories:fill-mask", "task_ids:language-modeling", "task_ids:masked-language-modeling", "annotations_creators:no-annotation", "language_creators:crowdsourced", "multilinguality:multilingual", "size_categories:n<1K", "size_categories:1K<n<10K", "size_categories:10K<n<100K", "size_categories:100K<n<1M", "size_categories:1M<n<10M", "source_datasets:original", "language:aa", "language:ab", "language:ace", "language:af", "language:ak", "language:als", "language:am", "language:an", "language:ang", "language:ar", "language:arc", "language:arz", "language:as", "language:ast", "language:atj", "language:av", "language:ay", "language:az", "language:azb", "language:ba", "language:bar", "language:bcl", "language:be", "language:bg", "language:bh", "language:bi", "language:bjn", "language:bm", "language:bn", "language:bo", "language:bpy", "language:br", "language:bs", "language:bug", "language:bxr", "language:ca", "language:cbk", "language:cdo", "language:ce", "language:ceb", "language:ch", "language:cho", "language:chr", "language:chy", "language:ckb", "language:co", "language:cr", "language:crh", "language:cs", "language:csb", "language:cu", "language:cv", "language:cy", "language:da", "language:de", "language:din", "language:diq", "language:dsb", "language:dty", "language:dv", "language:dz", "language:ee", "language:el", "language:eml", "language:en", "language:eo", "language:es", "language:et", "language:eu", "language:ext", "language:fa", "language:ff", "language:fi", "language:fj", "language:fo", "language:fr", "language:frp", "language:frr", "language:fur", "language:fy", "language:ga", "language:gag", "language:gan", "language:gd", "language:gl", "language:glk", "language:gn", "language:gom", "language:gor", "language:got", "language:gu", "language:gv", "language:ha", "language:hak", "language:haw", "language:he", "language:hi", "language:hif", "language:ho", "language:hr", "language:hsb", "language:ht", "language:hu", "language:hy", "language:ia", "language:id", "language:ie", "language:ig", "language:ii", "language:ik", "language:ilo", "language:inh", "language:io", "language:is", "language:it", "language:iu", "language:ja", "language:jam", "language:jbo", "language:jv", "language:ka", "language:kaa", "language:kab", "language:kbd", "language:kbp", "language:kg", "language:ki", "language:kj", "language:kk", "language:kl", "language:km", "language:kn", "language:ko", "language:koi", "language:krc", "language:ks", "language:ksh", "language:ku", "language:kv", "language:kw", "language:ky", "language:la", "language:lad", "language:lb", "language:lbe", "language:lez", "language:lfn", "language:lg", "language:li", "language:lij", "language:lmo", "language:ln", "language:lo", "language:lrc", "language:lt", "language:ltg", "language:lv", "language:lzh", "language:mai", "language:mdf", "language:mg", "language:mh", "language:mhr", "language:mi", "language:min", "language:mk", "language:ml", "language:mn", "language:mr", "language:mrj", "language:ms", "language:mt", "language:mus", "language:mwl", "language:my", "language:myv", "language:mzn", "language:na", "language:nah", "language:nan", "language:nap", "language:nds", "language:ne", "language:new", "language:ng", "language:nl", "language:nn", "language:no", "language:nov", "language:nrf", "language:nso", "language:nv", "language:ny", "language:oc", "language:olo", "language:om", "language:or", "language:os", "language:pa", "language:pag", "language:pam", "language:pap", "language:pcd", "language:pdc", "language:pfl", "language:pi", "language:pih", "language:pl", "language:pms", "language:pnb", "language:pnt", "language:ps", "language:pt", "language:qu", "language:rm", "language:rmy", "language:rn", "language:ro", "language:ru", "language:rue", "language:rup", "language:rw", "language:sa", "language:sah", "language:sat", "language:sc", "language:scn", "language:sco", "language:sd", "language:se", "language:sg", "language:sgs", "language:sh", "language:si", "language:sk", "language:sl", "language:sm", "language:sn", "language:so", "language:sq", "language:sr", "language:srn", "language:ss", "language:st", "language:stq", "language:su", "language:sv", "language:sw", "language:szl", "language:ta", "language:tcy", "language:tdt", "language:te", "language:tg", "language:th", "language:ti", "language:tk", "language:tl", "language:tn", "language:to", "language:tpi", "language:tr", "language:ts", "language:tt", "language:tum", "language:tw", "language:ty", "language:tyv", "language:udm", "language:ug", "language:uk", "language:ur", "language:uz", "language:ve", "language:vec", "language:vep", "language:vi", "language:vls", "language:vo", "language:vro", "language:wa", "language:war", "language:wo", "language:wuu", "language:xal", "language:xh", "language:xmf", "language:yi", "language:yo", "language:yue", "language:za", "language:zea", "language:zh", "language:zu", "license:cc-by-sa-3.0", "license:gfdl", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["crowdsourced"], "language": ["aa", "ab", "ace", "af", "ak", "als", "am", "an", "ang", "ar", "arc", "arz", "as", "ast", "atj", "av", "ay", "az", "azb", "ba", "bar", "bcl", "be", "bg", "bh", "bi", "bjn", "bm", "bn", "bo", "bpy", "br", "bs", "bug", "bxr", "ca", "cbk", "cdo", "ce", "ceb", "ch", "cho", "chr", "chy", "ckb", "co", "cr", "crh", "cs", "csb", "cu", "cv", "cy", "da", "de", "din", "diq", "dsb", "dty", "dv", "dz", "ee", "el", "eml", "en", "eo", "es", "et", "eu", "ext", "fa", "ff", "fi", "fj", "fo", "fr", "frp", "frr", "fur", "fy", "ga", "gag", "gan", "gd", "gl", "glk", "gn", "gom", "gor", "got", "gu", "gv", "ha", "hak", "haw", "he", "hi", "hif", "ho", "hr", "hsb", "ht", "hu", "hy", "ia", "id", "ie", "ig", "ii", "ik", "ilo", "inh", "io", "is", "it", "iu", "ja", "jam", "jbo", "jv", "ka", "kaa", "kab", "kbd", "kbp", "kg", "ki", "kj", "kk", "kl", "km", "kn", "ko", "koi", "krc", "ks", "ksh", "ku", "kv", "kw", "ky", "la", "lad", "lb", "lbe", "lez", "lfn", "lg", "li", "lij", "lmo", "ln", "lo", "lrc", "lt", "ltg", "lv", "lzh", "mai", "mdf", "mg", "mh", "mhr", "mi", "min", "mk", "ml", "mn", "mr", "mrj", "ms", "mt", "mus", "mwl", "my", "myv", "mzn", "na", "nah", "nan", "nap", "nds", "ne", "new", "ng", "nl", "nn", "no", "nov", "nrf", "nso", "nv", "ny", "oc", "olo", "om", "or", "os", "pa", "pag", "pam", "pap", "pcd", "pdc", "pfl", "pi", "pih", "pl", "pms", "pnb", "pnt", "ps", "pt", "qu", "rm", "rmy", "rn", "ro", "ru", "rue", "rup", "rw", "sa", "sah", "sat", "sc", "scn", "sco", "sd", "se", "sg", "sgs", "sh", "si", "sk", "sl", "sm", "sn", "so", "sq", "sr", "srn", "ss", "st", "stq", "su", "sv", "sw", "szl", "ta", "tcy", "tdt", "te", "tg", "th", "ti", "tk", "tl", "tn", "to", "tpi", "tr", "ts", "tt", "tum", "tw", "ty", "tyv", "udm", "ug", "uk", "ur", "uz", "ve", "vec", "vep", "vi", "vls", "vo", "vro", "wa", "war", "wo", "wuu", "xal", "xh", "xmf", "yi", "yo", "yue", "za", "zea", "zh", "zu"], "license": ["cc-by-sa-3.0", "gfdl"], "multilinguality": ["multilingual"], "size_categories": ["n<1K", "1K<n<10K", "10K<n<100K", "100K<n<1M", "1M<n<10M"], "source_datasets": ["original"], "task_categories": ["text-generation", "fill-mask"], "task_ids": ["language-modeling", "masked-language-modeling"], "pretty_name": "Wikipedia", "config_names": ["20220301.aa", "20220301.ab", "20220301.ace", "20220301.ady", "20220301.af", "20220301.ak", "20220301.als", "20220301.am", "20220301.an", "20220301.ang", "20220301.ar", "20220301.arc", "20220301.arz", "20220301.as", "20220301.ast", "20220301.atj", "20220301.av", "20220301.ay", "20220301.az", "20220301.azb", "20220301.ba", "20220301.bar", "20220301.bat-smg", "20220301.bcl", "20220301.be", "20220301.be-x-old", "20220301.bg", "20220301.bh", "20220301.bi", "20220301.bjn", "20220301.bm", "20220301.bn", "20220301.bo", "20220301.bpy", "20220301.br", "20220301.bs", "20220301.bug", "20220301.bxr", "20220301.ca", "20220301.cbk-zam", "20220301.cdo", "20220301.ce", "20220301.ceb", "20220301.ch", "20220301.cho", "20220301.chr", "20220301.chy", "20220301.ckb", "20220301.co", "20220301.cr", "20220301.crh", "20220301.cs", "20220301.csb", "20220301.cu", "20220301.cv", "20220301.cy", "20220301.da", "20220301.de", "20220301.din", "20220301.diq", "20220301.dsb", "20220301.dty", "20220301.dv", "20220301.dz", "20220301.ee", "20220301.el", "20220301.eml", "20220301.en", "20220301.eo", "20220301.es", "20220301.et", "20220301.eu", "20220301.ext", "20220301.fa", "20220301.ff", "20220301.fi", "20220301.fiu-vro", "20220301.fj", "20220301.fo", "20220301.fr", "20220301.frp", "20220301.frr", "20220301.fur", "20220301.fy", "20220301.ga", "20220301.gag", "20220301.gan", "20220301.gd", "20220301.gl", "20220301.glk", "20220301.gn", "20220301.gom", "20220301.gor", "20220301.got", "20220301.gu", "20220301.gv", "20220301.ha", "20220301.hak", "20220301.haw", "20220301.he", "20220301.hi", "20220301.hif", "20220301.ho", "20220301.hr", "20220301.hsb", "20220301.ht", "20220301.hu", "20220301.hy", "20220301.ia", "20220301.id", "20220301.ie", "20220301.ig", "20220301.ii", "20220301.ik", "20220301.ilo", "20220301.inh", "20220301.io", "20220301.is", "20220301.it", "20220301.iu", "20220301.ja", "20220301.jam", "20220301.jbo", "20220301.jv", "20220301.ka", "20220301.kaa", "20220301.kab", "20220301.kbd", "20220301.kbp", "20220301.kg", "20220301.ki", "20220301.kj", "20220301.kk", "20220301.kl", "20220301.km", "20220301.kn", "20220301.ko", "20220301.koi", "20220301.krc", "20220301.ks", "20220301.ksh", "20220301.ku", "20220301.kv", "20220301.kw", "20220301.ky", "20220301.la", "20220301.lad", "20220301.lb", "20220301.lbe", "20220301.lez", "20220301.lfn", "20220301.lg", "20220301.li", "20220301.lij", "20220301.lmo", "20220301.ln", "20220301.lo", "20220301.lrc", "20220301.lt", "20220301.ltg", "20220301.lv", "20220301.mai", "20220301.map-bms", "20220301.mdf", "20220301.mg", "20220301.mh", "20220301.mhr", "20220301.mi", "20220301.min", "20220301.mk", "20220301.ml", "20220301.mn", "20220301.mr", "20220301.mrj", "20220301.ms", "20220301.mt", "20220301.mus", "20220301.mwl", "20220301.my", "20220301.myv", "20220301.mzn", "20220301.na", "20220301.nah", "20220301.nap", "20220301.nds", "20220301.nds-nl", "20220301.ne", "20220301.new", "20220301.ng", "20220301.nl", "20220301.nn", "20220301.no", "20220301.nov", "20220301.nrm", "20220301.nso", "20220301.nv", "20220301.ny", "20220301.oc", "20220301.olo", "20220301.om", "20220301.or", "20220301.os", "20220301.pa", "20220301.pag", "20220301.pam", "20220301.pap", "20220301.pcd", "20220301.pdc", "20220301.pfl", "20220301.pi", "20220301.pih", "20220301.pl", "20220301.pms", "20220301.pnb", "20220301.pnt", "20220301.ps", "20220301.pt", "20220301.qu", "20220301.rm", "20220301.rmy", "20220301.rn", "20220301.ro", "20220301.roa-rup", "20220301.roa-tara", "20220301.ru", "20220301.rue", "20220301.rw", "20220301.sa", "20220301.sah", "20220301.sat", "20220301.sc", "20220301.scn", "20220301.sco", "20220301.sd", "20220301.se", "20220301.sg", "20220301.sh", "20220301.si", "20220301.simple", "20220301.sk", "20220301.sl", "20220301.sm", "20220301.sn", "20220301.so", "20220301.sq", "20220301.sr", "20220301.srn", "20220301.ss", "20220301.st", "20220301.stq", "20220301.su", "20220301.sv", "20220301.sw", "20220301.szl", "20220301.ta", "20220301.tcy", "20220301.te", "20220301.tet", "20220301.tg", "20220301.th", "20220301.ti", "20220301.tk", "20220301.tl", "20220301.tn", "20220301.to", "20220301.tpi", "20220301.tr", "20220301.ts", "20220301.tt", "20220301.tum", "20220301.tw", "20220301.ty", "20220301.tyv", "20220301.udm", "20220301.ug", "20220301.uk", "20220301.ur", "20220301.uz", "20220301.ve", "20220301.vec", "20220301.vep", "20220301.vi", "20220301.vls", "20220301.vo", "20220301.wa", "20220301.war", "20220301.wo", "20220301.wuu", "20220301.xal", "20220301.xh", "20220301.xmf", "20220301.yi", "20220301.yo", "20220301.za", "20220301.zea", "20220301.zh", "20220301.zh-classical", "20220301.zh-min-nan", "20220301.zh-yue", "20220301.zu"], "language_bcp47": ["nds-nl"], "dataset_info": [{"config_name": "20220301.de", "features": [{"name": "id", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 8905282792, "num_examples": 2665357}], "download_size": 6523215105, "dataset_size": 8905282792}, {"config_name": "20220301.en", "features": [{"name": "id", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 20275516160, "num_examples": 6458670}], "download_size": 20598313936, "dataset_size": 20275516160}, {"config_name": "20220301.fr", "features": [{"name": "id", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 7375920768, "num_examples": 2402095}], "download_size": 5602565274, "dataset_size": 7375920768}, {"config_name": "20220301.frr", "features": [{"name": "id", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 9129760, "num_examples": 15199}], "download_size": 12438017, "dataset_size": 9129760}, {"config_name": "20220301.it", "features": [{"name": "id", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 4539944448, "num_examples": 1743035}], "download_size": 3516441239, "dataset_size": 4539944448}, {"config_name": "20220301.simple", "features": [{"name": "id", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 235072360, "num_examples": 205328}], "download_size": 239682796, "dataset_size": 235072360}]}
2024-01-18T11:18:16+00:00
[]
[ "aa", "ab", "ace", "af", "ak", "als", "am", "an", "ang", "ar", "arc", "arz", "as", "ast", "atj", "av", "ay", "az", "azb", "ba", "bar", "bcl", "be", "bg", "bh", "bi", "bjn", "bm", "bn", "bo", "bpy", "br", "bs", "bug", "bxr", "ca", "cbk", "cdo", "ce", "ceb", "ch", "cho", "chr", "chy", "ckb", "co", "cr", "crh", "cs", "csb", "cu", "cv", "cy", "da", "de", "din", "diq", "dsb", "dty", "dv", "dz", "ee", "el", "eml", "en", "eo", "es", "et", "eu", "ext", "fa", "ff", "fi", "fj", "fo", "fr", "frp", "frr", "fur", "fy", "ga", "gag", "gan", "gd", "gl", "glk", "gn", "gom", "gor", "got", "gu", "gv", "ha", "hak", "haw", "he", "hi", "hif", "ho", "hr", "hsb", "ht", "hu", "hy", "ia", "id", "ie", "ig", "ii", "ik", "ilo", "inh", "io", "is", "it", "iu", "ja", "jam", "jbo", "jv", "ka", "kaa", "kab", "kbd", "kbp", "kg", "ki", "kj", "kk", "kl", "km", "kn", "ko", "koi", "krc", "ks", "ksh", "ku", "kv", "kw", "ky", "la", "lad", "lb", "lbe", "lez", "lfn", "lg", "li", "lij", "lmo", "ln", "lo", "lrc", "lt", "ltg", "lv", "lzh", "mai", "mdf", "mg", "mh", "mhr", "mi", "min", "mk", "ml", "mn", "mr", "mrj", "ms", "mt", "mus", "mwl", "my", "myv", "mzn", "na", "nah", "nan", "nap", "nds", "ne", "new", "ng", "nl", "nn", "no", "nov", "nrf", "nso", "nv", "ny", "oc", "olo", "om", "or", "os", "pa", "pag", "pam", "pap", "pcd", "pdc", "pfl", "pi", "pih", "pl", "pms", "pnb", "pnt", "ps", "pt", "qu", "rm", "rmy", "rn", "ro", "ru", "rue", "rup", "rw", "sa", "sah", "sat", "sc", "scn", "sco", "sd", "se", "sg", "sgs", "sh", "si", "sk", "sl", "sm", "sn", "so", "sq", "sr", "srn", "ss", "st", "stq", "su", "sv", "sw", "szl", "ta", "tcy", "tdt", "te", "tg", "th", "ti", "tk", "tl", "tn", "to", "tpi", "tr", "ts", "tt", "tum", "tw", "ty", "tyv", "udm", "ug", "uk", "ur", "uz", "ve", "vec", "vep", "vi", "vls", "vo", "vro", "wa", "war", "wo", "wuu", "xal", "xh", "xmf", "yi", "yo", "yue", "za", "zea", "zh", "zu" ]
TAGS #task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-crowdsourced #multilinguality-multilingual #size_categories-n<1K #size_categories-1K<n<10K #size_categories-10K<n<100K #size_categories-100K<n<1M #size_categories-1M<n<10M #source_datasets-original #language-Afar #language-Abkhazian #language-Achinese #language-Afrikaans #language-Akan #language-Tosk Albanian #language-Amharic #language-Aragonese #language-Old English (ca. 450-1100) #language-Arabic #language-Official Aramaic (700-300 BCE) #language-Egyptian Arabic #language-Assamese #language-Asturian #language-Atikamekw #language-Avaric #language-Aymara #language-Azerbaijani #language-South Azerbaijani #language-Bashkir #language-Bavarian #language-Central Bikol #language-Belarusian #language-Bulgarian #language-bh #language-Bislama #language-Banjar #language-Bambara #language-Bengali #language-Tibetan #language-Bishnupriya #language-Breton #language-Bosnian #language-Buginese #language-Russia Buriat #language-Catalan #language-Chavacano #language-Min Dong Chinese #language-Chechen #language-Cebuano #language-Chamorro #language-Choctaw #language-Cherokee #language-Cheyenne #language-Central Kurdish #language-Corsican #language-Cree #language-Crimean Tatar #language-Czech #language-Kashubian #language-Church Slavic #language-Chuvash #language-Welsh #language-Danish #language-German #language-Dinka #language-Dimli (individual language) #language-Lower Sorbian #language-Dotyali #language-Dhivehi #language-Dzongkha #language-Ewe #language-Modern Greek (1453-) #language-Emiliano-Romagnolo #language-English #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Extremaduran #language-Persian #language-Fulah #language-Finnish #language-Fijian #language-Faroese #language-French #language-Arpitan #language-Northern Frisian #language-Friulian #language-Western Frisian #language-Irish #language-Gagauz #language-Gan Chinese #language-Scottish Gaelic #language-Galician #language-Gilaki #language-Guarani #language-Goan Konkani #language-Gorontalo #language-Gothic #language-Gujarati #language-Manx #language-Hausa #language-Hakka Chinese #language-Hawaiian #language-Hebrew #language-Hindi #language-Fiji Hindi #language-Hiri Motu #language-Croatian #language-Upper Sorbian #language-Haitian #language-Hungarian #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Indonesian #language-Interlingue #language-Igbo #language-Sichuan Yi #language-Inupiaq #language-Iloko #language-Ingush #language-Ido #language-Icelandic #language-Italian #language-Inuktitut #language-Japanese #language-Jamaican Creole English #language-Lojban #language-Javanese #language-Georgian #language-Kara-Kalpak #language-Kabyle #language-Kabardian #language-Kabiyè #language-Kongo #language-Kikuyu #language-Kuanyama #language-Kazakh #language-Kalaallisut #language-Khmer #language-Kannada #language-Korean #language-Komi-Permyak #language-Karachay-Balkar #language-Kashmiri #language-Kölsch #language-Kurdish #language-Komi #language-Cornish #language-Kirghiz #language-Latin #language-Ladino #language-Luxembourgish #language-Lak #language-Lezghian #language-Lingua Franca Nova #language-Ganda #language-Limburgan #language-Ligurian #language-Lombard #language-Lingala #language-Lao #language-Northern Luri #language-Lithuanian #language-Latgalian #language-Latvian #language-Literary Chinese #language-Maithili #language-Moksha #language-Malagasy #language-Marshallese #language-Eastern Mari #language-Maori #language-Minangkabau #language-Macedonian #language-Malayalam #language-Mongolian #language-Marathi #language-Western Mari #language-Malay (macrolanguage) #language-Maltese #language-Creek #language-Mirandese #language-Burmese #language-Erzya #language-Mazanderani #language-Nauru #language-nah #language-Min Nan Chinese #language-Neapolitan #language-Low German #language-Nepali (macrolanguage) #language-Newari #language-Ndonga #language-Dutch #language-Norwegian Nynorsk #language-Norwegian #language-Novial #language-Jèrriais #language-Pedi #language-Navajo #language-Nyanja #language-Occitan (post 1500) #language-Livvi #language-Oromo #language-Oriya (macrolanguage) #language-Ossetian #language-Panjabi #language-Pangasinan #language-Pampanga #language-Papiamento #language-Picard #language-Pennsylvania German #language-Pfaelzisch #language-Pali #language-Pitcairn-Norfolk #language-Polish #language-Piemontese #language-Western Panjabi #language-Pontic #language-Pushto #language-Portuguese #language-Quechua #language-Romansh #language-Vlax Romani #language-Rundi #language-Romanian #language-Russian #language-Rusyn #language-Macedo-Romanian #language-Kinyarwanda #language-Sanskrit #language-Yakut #language-Santali #language-Sardinian #language-Sicilian #language-Scots #language-Sindhi #language-Northern Sami #language-Sango #language-Samogitian #language-Serbo-Croatian #language-Sinhala #language-Slovak #language-Slovenian #language-Samoan #language-Shona #language-Somali #language-Albanian #language-Serbian #language-Sranan Tongo #language-Swati #language-Southern Sotho #language-Saterfriesisch #language-Sundanese #language-Swedish #language-Swahili (macrolanguage) #language-Silesian #language-Tamil #language-Tulu #language-Tetun Dili #language-Telugu #language-Tajik #language-Thai #language-Tigrinya #language-Turkmen #language-Tagalog #language-Tswana #language-Tonga (Tonga Islands) #language-Tok Pisin #language-Turkish #language-Tsonga #language-Tatar #language-Tumbuka #language-Twi #language-Tahitian #language-Tuvinian #language-Udmurt #language-Uighur #language-Ukrainian #language-Urdu #language-Uzbek #language-Venda #language-Venetian #language-Veps #language-Vietnamese #language-Vlaams #language-Volapük #language-Võro #language-Walloon #language-Waray (Philippines) #language-Wolof #language-Wu Chinese #language-Kalmyk #language-Xhosa #language-Mingrelian #language-Yiddish #language-Yoruba #language-Yue Chinese #language-Zhuang #language-Zeeuws #language-Chinese #language-Zulu #license-cc-by-sa-3.0 #license-gfdl #region-us
Dataset Card for Wikipedia ========================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: * Point of Contact: ### Dataset Summary Wikipedia dataset containing cleaned articles of all languages. The datasets are built from the Wikipedia dump (URL with one split per language. Each example contains the content of one full Wikipedia article with cleaning to strip markdown and unwanted sections (references, etc.). The articles are parsed using the ''mwparserfromhell'' tool. To load this dataset you need to install Apache Beam and ''mwparserfromhell'' first: Then, you can load any subset of Wikipedia per language and per date this way: where you can pass as 'beam\_runner' any Apache Beam supported runner for (distributed) data processing (see here). Pass "DirectRunner" to run it on your machine. You can find the full list of languages and dates here. Some subsets of Wikipedia have already been processed by HuggingFace, and you can load them just with: The list of pre-processed subsets is: * "URL" * "URL" * "URL" * "URL" * "URL" * "URL" ### Supported Tasks and Leaderboards The dataset is generally used for Language Modeling. ### Languages You can find the list of languages here. Dataset Structure ----------------- ### Data Instances An example looks as follows: Some subsets of Wikipedia have already been processed by HuggingFace, as you can see below: #### URL * Size of downloaded dataset files: 6.84 GB * Size of the generated dataset: 9.34 GB * Total amount of disk used: 16.18 GB #### URL * Size of downloaded dataset files: 21.60 GB * Size of the generated dataset: 21.26 GB * Total amount of disk used: 42.86 GB #### URL * Size of downloaded dataset files: 5.87 GB * Size of the generated dataset: 7.73 GB * Total amount of disk used: 13.61 GB #### URL * Size of downloaded dataset files: 13.04 MB * Size of the generated dataset: 9.57 MB * Total amount of disk used: 22.62 MB #### URL * Size of downloaded dataset files: 3.69 GB * Size of the generated dataset: 4.76 GB * Total amount of disk used: 8.45 GB #### URL * Size of downloaded dataset files: 251.32 MB * Size of the generated dataset: 246.49 MB * Total amount of disk used: 497.82 MB ### Data Fields The data fields are the same among all configurations: * 'id' ('str'): ID of the article. * 'url' ('str'): URL of the article. * 'title' ('str'): Title of the article. * 'text' ('str'): Text content of the article. ### Data Splits Here are the number of examples for several configurations: Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information Most of Wikipedia's text and many of its images are co-licensed under the Creative Commons Attribution-ShareAlike 3.0 Unported License (CC BY-SA) and the GNU Free Documentation License (GFDL) (unversioned, with no invariant sections, front-cover texts, or back-cover texts). Some text has been imported only under CC BY-SA and CC BY-SA-compatible license and cannot be reused under GFDL; such text will be identified on the page footer, in the page history, or on the discussion page of the article that utilizes the text. ### Contributions Thanks to @lewtun, @mariamabarham, @thomwolf, @lhoestq, @patrickvonplaten for adding this dataset.
[ "### Dataset Summary\n\n\nWikipedia dataset containing cleaned articles of all languages.\nThe datasets are built from the Wikipedia dump\n(URL with one split per language. Each example\ncontains the content of one full Wikipedia article with cleaning to strip\nmarkdown and unwanted sections (references, etc.).\n\n\nThe articles are parsed using the ''mwparserfromhell'' tool.\n\n\nTo load this dataset you need to install Apache Beam and ''mwparserfromhell'' first:\n\n\nThen, you can load any subset of Wikipedia per language and per date this way:\n\n\nwhere you can pass as 'beam\\_runner' any Apache Beam supported runner for (distributed) data processing\n(see here).\nPass \"DirectRunner\" to run it on your machine.\n\n\nYou can find the full list of languages and dates here.\n\n\nSome subsets of Wikipedia have already been processed by HuggingFace, and you can load them just with:\n\n\nThe list of pre-processed subsets is:\n\n\n* \"URL\"\n* \"URL\"\n* \"URL\"\n* \"URL\"\n* \"URL\"\n* \"URL\"", "### Supported Tasks and Leaderboards\n\n\nThe dataset is generally used for Language Modeling.", "### Languages\n\n\nYou can find the list of languages here.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nAn example looks as follows:\n\n\nSome subsets of Wikipedia have already been processed by HuggingFace, as you can see below:", "#### URL\n\n\n* Size of downloaded dataset files: 6.84 GB\n* Size of the generated dataset: 9.34 GB\n* Total amount of disk used: 16.18 GB", "#### URL\n\n\n* Size of downloaded dataset files: 21.60 GB\n* Size of the generated dataset: 21.26 GB\n* Total amount of disk used: 42.86 GB", "#### URL\n\n\n* Size of downloaded dataset files: 5.87 GB\n* Size of the generated dataset: 7.73 GB\n* Total amount of disk used: 13.61 GB", "#### URL\n\n\n* Size of downloaded dataset files: 13.04 MB\n* Size of the generated dataset: 9.57 MB\n* Total amount of disk used: 22.62 MB", "#### URL\n\n\n* Size of downloaded dataset files: 3.69 GB\n* Size of the generated dataset: 4.76 GB\n* Total amount of disk used: 8.45 GB", "#### URL\n\n\n* Size of downloaded dataset files: 251.32 MB\n* Size of the generated dataset: 246.49 MB\n* Total amount of disk used: 497.82 MB", "### Data Fields\n\n\nThe data fields are the same among all configurations:\n\n\n* 'id' ('str'): ID of the article.\n* 'url' ('str'): URL of the article.\n* 'title' ('str'): Title of the article.\n* 'text' ('str'): Text content of the article.", "### Data Splits\n\n\nHere are the number of examples for several configurations:\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nMost of Wikipedia's text and many of its images are co-licensed under the\nCreative Commons Attribution-ShareAlike 3.0 Unported License\n(CC BY-SA) and the GNU Free Documentation License\n(GFDL) (unversioned, with no invariant sections, front-cover texts, or back-cover texts).\n\n\nSome text has been imported only under CC BY-SA and CC BY-SA-compatible license and cannot be reused under GFDL; such\ntext will be identified on the page footer, in the page history, or on the discussion page of the article that utilizes\nthe text.", "### Contributions\n\n\nThanks to @lewtun, @mariamabarham, @thomwolf, @lhoestq, @patrickvonplaten for adding this dataset." ]
[ "TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-crowdsourced #multilinguality-multilingual #size_categories-n<1K #size_categories-1K<n<10K #size_categories-10K<n<100K #size_categories-100K<n<1M #size_categories-1M<n<10M #source_datasets-original #language-Afar #language-Abkhazian #language-Achinese #language-Afrikaans #language-Akan #language-Tosk Albanian #language-Amharic #language-Aragonese #language-Old English (ca. 450-1100) #language-Arabic #language-Official Aramaic (700-300 BCE) #language-Egyptian Arabic #language-Assamese #language-Asturian #language-Atikamekw #language-Avaric #language-Aymara #language-Azerbaijani #language-South Azerbaijani #language-Bashkir #language-Bavarian #language-Central Bikol #language-Belarusian #language-Bulgarian #language-bh #language-Bislama #language-Banjar #language-Bambara #language-Bengali #language-Tibetan #language-Bishnupriya #language-Breton #language-Bosnian #language-Buginese #language-Russia Buriat #language-Catalan #language-Chavacano #language-Min Dong Chinese #language-Chechen #language-Cebuano #language-Chamorro #language-Choctaw #language-Cherokee #language-Cheyenne #language-Central Kurdish #language-Corsican #language-Cree #language-Crimean Tatar #language-Czech #language-Kashubian #language-Church Slavic #language-Chuvash #language-Welsh #language-Danish #language-German #language-Dinka #language-Dimli (individual language) #language-Lower Sorbian #language-Dotyali #language-Dhivehi #language-Dzongkha #language-Ewe #language-Modern Greek (1453-) #language-Emiliano-Romagnolo #language-English #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Extremaduran #language-Persian #language-Fulah #language-Finnish #language-Fijian #language-Faroese #language-French #language-Arpitan #language-Northern Frisian #language-Friulian #language-Western Frisian #language-Irish #language-Gagauz #language-Gan Chinese #language-Scottish Gaelic #language-Galician #language-Gilaki #language-Guarani #language-Goan Konkani #language-Gorontalo #language-Gothic #language-Gujarati #language-Manx #language-Hausa #language-Hakka Chinese #language-Hawaiian #language-Hebrew #language-Hindi #language-Fiji Hindi #language-Hiri Motu #language-Croatian #language-Upper Sorbian #language-Haitian #language-Hungarian #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Indonesian #language-Interlingue #language-Igbo #language-Sichuan Yi #language-Inupiaq #language-Iloko #language-Ingush #language-Ido #language-Icelandic #language-Italian #language-Inuktitut #language-Japanese #language-Jamaican Creole English #language-Lojban #language-Javanese #language-Georgian #language-Kara-Kalpak #language-Kabyle #language-Kabardian #language-Kabiyè #language-Kongo #language-Kikuyu #language-Kuanyama #language-Kazakh #language-Kalaallisut #language-Khmer #language-Kannada #language-Korean #language-Komi-Permyak #language-Karachay-Balkar #language-Kashmiri #language-Kölsch #language-Kurdish #language-Komi #language-Cornish #language-Kirghiz #language-Latin #language-Ladino #language-Luxembourgish #language-Lak #language-Lezghian #language-Lingua Franca Nova #language-Ganda #language-Limburgan #language-Ligurian #language-Lombard #language-Lingala #language-Lao #language-Northern Luri #language-Lithuanian #language-Latgalian #language-Latvian #language-Literary Chinese #language-Maithili #language-Moksha #language-Malagasy #language-Marshallese #language-Eastern Mari #language-Maori #language-Minangkabau #language-Macedonian #language-Malayalam #language-Mongolian #language-Marathi #language-Western Mari #language-Malay (macrolanguage) #language-Maltese #language-Creek #language-Mirandese #language-Burmese #language-Erzya #language-Mazanderani #language-Nauru #language-nah #language-Min Nan Chinese #language-Neapolitan #language-Low German #language-Nepali (macrolanguage) #language-Newari #language-Ndonga #language-Dutch #language-Norwegian Nynorsk #language-Norwegian #language-Novial #language-Jèrriais #language-Pedi #language-Navajo #language-Nyanja #language-Occitan (post 1500) #language-Livvi #language-Oromo #language-Oriya (macrolanguage) #language-Ossetian #language-Panjabi #language-Pangasinan #language-Pampanga #language-Papiamento #language-Picard #language-Pennsylvania German #language-Pfaelzisch #language-Pali #language-Pitcairn-Norfolk #language-Polish #language-Piemontese #language-Western Panjabi #language-Pontic #language-Pushto #language-Portuguese #language-Quechua #language-Romansh #language-Vlax Romani #language-Rundi #language-Romanian #language-Russian #language-Rusyn #language-Macedo-Romanian #language-Kinyarwanda #language-Sanskrit #language-Yakut #language-Santali #language-Sardinian #language-Sicilian #language-Scots #language-Sindhi #language-Northern Sami #language-Sango #language-Samogitian #language-Serbo-Croatian #language-Sinhala #language-Slovak #language-Slovenian #language-Samoan #language-Shona #language-Somali #language-Albanian #language-Serbian #language-Sranan Tongo #language-Swati #language-Southern Sotho #language-Saterfriesisch #language-Sundanese #language-Swedish #language-Swahili (macrolanguage) #language-Silesian #language-Tamil #language-Tulu #language-Tetun Dili #language-Telugu #language-Tajik #language-Thai #language-Tigrinya #language-Turkmen #language-Tagalog #language-Tswana #language-Tonga (Tonga Islands) #language-Tok Pisin #language-Turkish #language-Tsonga #language-Tatar #language-Tumbuka #language-Twi #language-Tahitian #language-Tuvinian #language-Udmurt #language-Uighur #language-Ukrainian #language-Urdu #language-Uzbek #language-Venda #language-Venetian #language-Veps #language-Vietnamese #language-Vlaams #language-Volapük #language-Võro #language-Walloon #language-Waray (Philippines) #language-Wolof #language-Wu Chinese #language-Kalmyk #language-Xhosa #language-Mingrelian #language-Yiddish #language-Yoruba #language-Yue Chinese #language-Zhuang #language-Zeeuws #language-Chinese #language-Zulu #license-cc-by-sa-3.0 #license-gfdl #region-us \n", "### Dataset Summary\n\n\nWikipedia dataset containing cleaned articles of all languages.\nThe datasets are built from the Wikipedia dump\n(URL with one split per language. Each example\ncontains the content of one full Wikipedia article with cleaning to strip\nmarkdown and unwanted sections (references, etc.).\n\n\nThe articles are parsed using the ''mwparserfromhell'' tool.\n\n\nTo load this dataset you need to install Apache Beam and ''mwparserfromhell'' first:\n\n\nThen, you can load any subset of Wikipedia per language and per date this way:\n\n\nwhere you can pass as 'beam\\_runner' any Apache Beam supported runner for (distributed) data processing\n(see here).\nPass \"DirectRunner\" to run it on your machine.\n\n\nYou can find the full list of languages and dates here.\n\n\nSome subsets of Wikipedia have already been processed by HuggingFace, and you can load them just with:\n\n\nThe list of pre-processed subsets is:\n\n\n* \"URL\"\n* \"URL\"\n* \"URL\"\n* \"URL\"\n* \"URL\"\n* \"URL\"", "### Supported Tasks and Leaderboards\n\n\nThe dataset is generally used for Language Modeling.", "### Languages\n\n\nYou can find the list of languages here.\n\n\nDataset Structure\n-----------------", "### Data Instances\n\n\nAn example looks as follows:\n\n\nSome subsets of Wikipedia have already been processed by HuggingFace, as you can see below:", "#### URL\n\n\n* Size of downloaded dataset files: 6.84 GB\n* Size of the generated dataset: 9.34 GB\n* Total amount of disk used: 16.18 GB", "#### URL\n\n\n* Size of downloaded dataset files: 21.60 GB\n* Size of the generated dataset: 21.26 GB\n* Total amount of disk used: 42.86 GB", "#### URL\n\n\n* Size of downloaded dataset files: 5.87 GB\n* Size of the generated dataset: 7.73 GB\n* Total amount of disk used: 13.61 GB", "#### URL\n\n\n* Size of downloaded dataset files: 13.04 MB\n* Size of the generated dataset: 9.57 MB\n* Total amount of disk used: 22.62 MB", "#### URL\n\n\n* Size of downloaded dataset files: 3.69 GB\n* Size of the generated dataset: 4.76 GB\n* Total amount of disk used: 8.45 GB", "#### URL\n\n\n* Size of downloaded dataset files: 251.32 MB\n* Size of the generated dataset: 246.49 MB\n* Total amount of disk used: 497.82 MB", "### Data Fields\n\n\nThe data fields are the same among all configurations:\n\n\n* 'id' ('str'): ID of the article.\n* 'url' ('str'): URL of the article.\n* 'title' ('str'): Title of the article.\n* 'text' ('str'): Text content of the article.", "### Data Splits\n\n\nHere are the number of examples for several configurations:\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nMost of Wikipedia's text and many of its images are co-licensed under the\nCreative Commons Attribution-ShareAlike 3.0 Unported License\n(CC BY-SA) and the GNU Free Documentation License\n(GFDL) (unversioned, with no invariant sections, front-cover texts, or back-cover texts).\n\n\nSome text has been imported only under CC BY-SA and CC BY-SA-compatible license and cannot be reused under GFDL; such\ntext will be identified on the page footer, in the page history, or on the discussion page of the article that utilizes\nthe text.", "### Contributions\n\n\nThanks to @lewtun, @mariamabarham, @thomwolf, @lhoestq, @patrickvonplaten for adding this dataset." ]
[ 1978, 248, 21, 21, 36, 37, 38, 37, 37, 37, 42, 74, 23, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 137, 39 ]
[ "passage: ", "passage: TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-crowdsourced #multilinguality-multilingual #size_categories-n<1K #size_categories-1K<n<10K #size_categories-10K<n<100K #size_categories-100K<n<1M #size_categories-1M<n<10M #source_datasets-original #language-Afar #language-Abkhazian #language-Achinese #language-Afrikaans #language-Akan #language-Tosk Albanian #language-Amharic #language-Aragonese #language-Old English (ca. 450-1100) #language-Arabic #language-Official Aramaic (700-300 BCE) #language-Egyptian Arabic #language-Assamese #language-Asturian #language-Atikamekw #language-Avaric #language-Aymara #language-Azerbaijani #language-South Azerbaijani #language-Bashkir #language-Bavarian #language-Central Bikol #language-Belarusian #language-Bulgarian #language-bh #language-Bislama #language-Banjar #language-Bambara #language-Bengali #language-Tibetan #language-Bishnupriya #language-Breton #language-Bosnian #language-Buginese #language-Russia Buriat #language-Catalan #language-Chavacano #language-Min Dong Chinese #language-Chechen #language-Cebuano #language-Chamorro #language-Choctaw #language-Cherokee #language-Cheyenne #language-Central Kurdish #language-Corsican #language-Cree #language-Crimean Tatar #language-Czech #language-Kashubian #language-Church Slavic #language-Chuvash #language-Welsh #language-Danish #language-German #language-Dinka #language-Dimli (individual language) #language-Lower Sorbian #language-Dotyali #language-Dhivehi #language-Dzongkha #language-Ewe #language-Modern Greek (1453-) #language-Emiliano-Romagnolo #language-English #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Extremaduran #language-Persian #language-Fulah #language-Finnish #language-Fijian #language-Faroese #language-French #language-Arpitan #language-Northern Frisian #language-Friulian #language-Western Frisian #language-Irish #language-Gagauz #language-Gan Chinese #language-Scottish Gaelic #language-Galician #language-Gilaki #language-Guarani #language-Goan Konkani #language-Gorontalo #language-Gothic #language-Gujarati #language-Manx #language-Hausa #language-Hakka Chinese #language-Hawaiian #language-Hebrew #language-Hindi #language-Fiji Hindi #language-Hiri Motu #language-Croatian #language-Upper Sorbian #language-Haitian #language-Hungarian #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Indonesian #language-Interlingue #language-Igbo #language-Sichuan Yi #language-Inupiaq #language-Iloko #language-Ingush #language-Ido #language-Icelandic #language-Italian #language-Inuktitut #language-Japanese #language-Jamaican Creole English #language-Lojban #language-Javanese #language-Georgian #language-Kara-Kalpak #language-Kabyle #language-Kabardian #language-Kabiyè #language-Kongo #language-Kikuyu #language-Kuanyama #language-Kazakh #language-Kalaallisut #language-Khmer #language-Kannada #language-Korean #language-Komi-Permyak #language-Karachay-Balkar #language-Kashmiri #language-Kölsch #language-Kurdish #language-Komi #language-Cornish #language-Kirghiz #language-Latin #language-Ladino #language-Luxembourgish #language-Lak #language-Lezghian #language-Lingua Franca Nova #language-Ganda #language-Limburgan #language-Ligurian #language-Lombard #language-Lingala #language-Lao #language-Northern Luri #language-Lithuanian #language-Latgalian #language-Latvian #language-Literary Chinese #language-Maithili #language-Moksha #language-Malagasy #language-Marshallese #language-Eastern Mari #language-Maori #language-Minangkabau #language-Macedonian #language-Malayalam #language-Mongolian #language-Marathi #language-Western Mari #language-Malay (macrolanguage) #language-Maltese #language-Creek #language-Mirandese #language-Burmese #language-Erzya #language-Mazanderani #language-Nauru #language-nah #language-Min Nan Chinese #language-Neapolitan #language-Low German #language-Nepali (macrolanguage) #language-Newari #language-Ndonga #language-Dutch #language-Norwegian Nynorsk #language-Norwegian #language-Novial #language-Jèrriais #language-Pedi #language-Navajo #language-Nyanja #language-Occitan (post 1500) #language-Livvi #language-Oromo #language-Oriya (macrolanguage) #language-Ossetian #language-Panjabi #language-Pangasinan #language-Pampanga #language-Papiamento #language-Picard #language-Pennsylvania German #language-Pfaelzisch #language-Pali #language-Pitcairn-Norfolk #language-Polish #language-Piemontese #language-Western Panjabi #language-Pontic #language-Pushto #language-Portuguese #language-Quechua #language-Romansh #language-Vlax Romani #language-Rundi #language-Romanian #language-Russian #language-Rusyn #language-Macedo-Romanian #language-Kinyarwanda #language-Sanskrit #language-Yakut #language-Santali #language-Sardinian #language-Sicilian #language-Scots #language-Sindhi #language-Northern Sami #language-Sango #language-Samogitian #language-Serbo-Croatian #language-Sinhala #language-Slovak #language-Slovenian #language-Samoan #language-Shona #language-Somali #language-Albanian #language-Serbian #language-Sranan Tongo #language-Swati #language-Southern Sotho #language-Saterfriesisch #language-Sundanese #language-Swedish #language-Swahili (macrolanguage) #language-Silesian #language-Tamil #language-Tulu #language-Tetun Dili #language-Telugu #language-Tajik #language-Thai #language-Tigrinya #language-Turkmen #language-Tagalog #language-Tswana #language-Tonga (Tonga Islands) #language-Tok Pisin #language-Turkish #language-Tsonga #language-Tatar #language-Tumbuka #language-Twi #language-Tahitian #language-Tuvinian #language-Udmurt #language-Uighur #language-Ukrainian #language-Urdu #language-Uzbek #language-Venda #language-Venetian #language-Veps #language-Vietnamese #language-Vlaams #language-Volapük #language-Võro #language-Walloon #language-Waray (Philippines) #language-Wolof #language-Wu Chinese #language-Kalmyk #language-Xhosa #language-Mingrelian #language-Yiddish #language-Yoruba #language-Yue Chinese #language-Zhuang #language-Zeeuws #language-Chinese #language-Zulu #license-cc-by-sa-3.0 #license-gfdl #region-us \n### Dataset Summary\n\n\nWikipedia dataset containing cleaned articles of all languages.\nThe datasets are built from the Wikipedia dump\n(URL with one split per language. Each example\ncontains the content of one full Wikipedia article with cleaning to strip\nmarkdown and unwanted sections (references, etc.).\n\n\nThe articles are parsed using the ''mwparserfromhell'' tool.\n\n\nTo load this dataset you need to install Apache Beam and ''mwparserfromhell'' first:\n\n\nThen, you can load any subset of Wikipedia per language and per date this way:\n\n\nwhere you can pass as 'beam\\_runner' any Apache Beam supported runner for (distributed) data processing\n(see here).\nPass \"DirectRunner\" to run it on your machine.\n\n\nYou can find the full list of languages and dates here.\n\n\nSome subsets of Wikipedia have already been processed by HuggingFace, and you can load them just with:\n\n\nThe list of pre-processed subsets is:\n\n\n* \"URL\"\n* \"URL\"\n* \"URL\"\n* \"URL\"\n* \"URL\"\n* \"URL\"### Supported Tasks and Leaderboards\n\n\nThe dataset is generally used for Language Modeling.### Languages\n\n\nYou can find the list of languages here.\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nAn example looks as follows:\n\n\nSome subsets of Wikipedia have already been processed by HuggingFace, as you can see below:#### URL\n\n\n* Size of downloaded dataset files: 6.84 GB\n* Size of the generated dataset: 9.34 GB\n* Total amount of disk used: 16.18 GB#### URL\n\n\n* Size of downloaded dataset files: 21.60 GB\n* Size of the generated dataset: 21.26 GB\n* Total amount of disk used: 42.86 GB#### URL\n\n\n* Size of downloaded dataset files: 5.87 GB\n* Size of the generated dataset: 7.73 GB\n* Total amount of disk used: 13.61 GB#### URL\n\n\n* Size of downloaded dataset files: 13.04 MB\n* Size of the generated dataset: 9.57 MB\n* Total amount of disk used: 22.62 MB" ]
5de818eb69c5385e763201d9b9abc36df69a81dc
# Dataset Card for "wikisql" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Repository:** https://github.com/salesforce/WikiSQL - **Paper:** [Seq2SQL: Generating Structured Queries from Natural Language using Reinforcement Learning](https://arxiv.org/abs/1709.00103) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 26.16 MB - **Size of the generated dataset:** 154.74 MB - **Total amount of disk used:** 180.90 MB ### Dataset Summary A large crowd-sourced dataset for developing natural language interfaces for relational databases. WikiSQL is a dataset of 80654 hand-annotated examples of questions and SQL queries distributed across 24241 tables from Wikipedia. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### default - **Size of downloaded dataset files:** 26.16 MB - **Size of the generated dataset:** 154.74 MB - **Total amount of disk used:** 180.90 MB An example of 'validation' looks as follows. ``` This example was too long and was cropped: { "phase": 1, "question": "How would you answer a second test question?", "sql": { "agg": 0, "conds": { "column_index": [2], "condition": ["Some Entity"], "operator_index": [0] }, "human_readable": "SELECT Header1 FROM table WHERE Another Header = Some Entity", "sel": 0 }, "table": "{\"caption\": \"L\", \"header\": [\"Header1\", \"Header 2\", \"Another Header\"], \"id\": \"1-10015132-9\", \"name\": \"table_10015132_11\", \"page_i..." } ``` ### Data Fields The data fields are the same among all splits. #### default - `phase`: a `int32` feature. - `question`: a `string` feature. - `header`: a `list` of `string` features. - `page_title`: a `string` feature. - `page_id`: a `string` feature. - `types`: a `list` of `string` features. - `id`: a `string` feature. - `section_title`: a `string` feature. - `caption`: a `string` feature. - `rows`: a dictionary feature containing: - `feature`: a `string` feature. - `name`: a `string` feature. - `human_readable`: a `string` feature. - `sel`: a `int32` feature. - `agg`: a `int32` feature. - `conds`: a dictionary feature containing: - `column_index`: a `int32` feature. - `operator_index`: a `int32` feature. - `condition`: a `string` feature. ### Data Splits | name |train|validation|test | |-------|----:|---------:|----:| |default|56355| 8421|15878| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @article{zhongSeq2SQL2017, author = {Victor Zhong and Caiming Xiong and Richard Socher}, title = {Seq2SQL: Generating Structured Queries from Natural Language using Reinforcement Learning}, journal = {CoRR}, volume = {abs/1709.00103}, year = {2017} } ``` ### Contributions Thanks to [@lewtun](https://github.com/lewtun), [@ghomasHudson](https://github.com/ghomasHudson), [@thomwolf](https://github.com/thomwolf) for adding this dataset.
wikisql
[ "task_categories:text2text-generation", "annotations_creators:crowdsourced", "language_creators:found", "language_creators:machine-generated", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:unknown", "text-to-sql", "arxiv:1709.00103", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["found", "machine-generated"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text2text-generation"], "task_ids": [], "paperswithcode_id": "wikisql", "pretty_name": "WikiSQL", "tags": ["text-to-sql"], "dataset_info": {"features": [{"name": "phase", "dtype": "int32"}, {"name": "question", "dtype": "string"}, {"name": "table", "struct": [{"name": "header", "sequence": "string"}, {"name": "page_title", "dtype": "string"}, {"name": "page_id", "dtype": "string"}, {"name": "types", "sequence": "string"}, {"name": "id", "dtype": "string"}, {"name": "section_title", "dtype": "string"}, {"name": "caption", "dtype": "string"}, {"name": "rows", "sequence": {"sequence": "string"}}, {"name": "name", "dtype": "string"}]}, {"name": "sql", "struct": [{"name": "human_readable", "dtype": "string"}, {"name": "sel", "dtype": "int32"}, {"name": "agg", "dtype": "int32"}, {"name": "conds", "sequence": [{"name": "column_index", "dtype": "int32"}, {"name": "operator_index", "dtype": "int32"}, {"name": "condition", "dtype": "string"}]}]}], "splits": [{"name": "test", "num_bytes": 32234761, "num_examples": 15878}, {"name": "validation", "num_bytes": 15159314, "num_examples": 8421}, {"name": "train", "num_bytes": 107345917, "num_examples": 56355}], "download_size": 26164664, "dataset_size": 154739992}}
2024-01-18T11:18:17+00:00
[ "1709.00103" ]
[ "en" ]
TAGS #task_categories-text2text-generation #annotations_creators-crowdsourced #language_creators-found #language_creators-machine-generated #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-unknown #text-to-sql #arxiv-1709.00103 #region-us
Dataset Card for "wikisql" ========================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Repository: URL * Paper: Seq2SQL: Generating Structured Queries from Natural Language using Reinforcement Learning * Point of Contact: * Size of downloaded dataset files: 26.16 MB * Size of the generated dataset: 154.74 MB * Total amount of disk used: 180.90 MB ### Dataset Summary A large crowd-sourced dataset for developing natural language interfaces for relational databases. WikiSQL is a dataset of 80654 hand-annotated examples of questions and SQL queries distributed across 24241 tables from Wikipedia. ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- ### Data Instances #### default * Size of downloaded dataset files: 26.16 MB * Size of the generated dataset: 154.74 MB * Total amount of disk used: 180.90 MB An example of 'validation' looks as follows. ### Data Fields The data fields are the same among all splits. #### default * 'phase': a 'int32' feature. * 'question': a 'string' feature. * 'header': a 'list' of 'string' features. * 'page\_title': a 'string' feature. * 'page\_id': a 'string' feature. * 'types': a 'list' of 'string' features. * 'id': a 'string' feature. * 'section\_title': a 'string' feature. * 'caption': a 'string' feature. * 'rows': a dictionary feature containing: + 'feature': a 'string' feature. * 'name': a 'string' feature. * 'human\_readable': a 'string' feature. * 'sel': a 'int32' feature. * 'agg': a 'int32' feature. * 'conds': a dictionary feature containing: + 'column\_index': a 'int32' feature. + 'operator\_index': a 'int32' feature. + 'condition': a 'string' feature. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @lewtun, @ghomasHudson, @thomwolf for adding this dataset.
[ "### Dataset Summary\n\n\nA large crowd-sourced dataset for developing natural language interfaces for relational databases.\n\n\nWikiSQL is a dataset of 80654 hand-annotated examples\nof questions and SQL queries distributed across 24241 tables from Wikipedia.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### default\n\n\n* Size of downloaded dataset files: 26.16 MB\n* Size of the generated dataset: 154.74 MB\n* Total amount of disk used: 180.90 MB\n\n\nAn example of 'validation' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### default\n\n\n* 'phase': a 'int32' feature.\n* 'question': a 'string' feature.\n* 'header': a 'list' of 'string' features.\n* 'page\\_title': a 'string' feature.\n* 'page\\_id': a 'string' feature.\n* 'types': a 'list' of 'string' features.\n* 'id': a 'string' feature.\n* 'section\\_title': a 'string' feature.\n* 'caption': a 'string' feature.\n* 'rows': a dictionary feature containing:\n\t+ 'feature': a 'string' feature.\n* 'name': a 'string' feature.\n* 'human\\_readable': a 'string' feature.\n* 'sel': a 'int32' feature.\n* 'agg': a 'int32' feature.\n* 'conds': a dictionary feature containing:\n\t+ 'column\\_index': a 'int32' feature.\n\t+ 'operator\\_index': a 'int32' feature.\n\t+ 'condition': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @lewtun, @ghomasHudson, @thomwolf for adding this dataset." ]
[ "TAGS\n#task_categories-text2text-generation #annotations_creators-crowdsourced #language_creators-found #language_creators-machine-generated #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-unknown #text-to-sql #arxiv-1709.00103 #region-us \n", "### Dataset Summary\n\n\nA large crowd-sourced dataset for developing natural language interfaces for relational databases.\n\n\nWikiSQL is a dataset of 80654 hand-annotated examples\nof questions and SQL queries distributed across 24241 tables from Wikipedia.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### default\n\n\n* Size of downloaded dataset files: 26.16 MB\n* Size of the generated dataset: 154.74 MB\n* Total amount of disk used: 180.90 MB\n\n\nAn example of 'validation' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### default\n\n\n* 'phase': a 'int32' feature.\n* 'question': a 'string' feature.\n* 'header': a 'list' of 'string' features.\n* 'page\\_title': a 'string' feature.\n* 'page\\_id': a 'string' feature.\n* 'types': a 'list' of 'string' features.\n* 'id': a 'string' feature.\n* 'section\\_title': a 'string' feature.\n* 'caption': a 'string' feature.\n* 'rows': a dictionary feature containing:\n\t+ 'feature': a 'string' feature.\n* 'name': a 'string' feature.\n* 'human\\_readable': a 'string' feature.\n* 'sel': a 'int32' feature.\n* 'agg': a 'int32' feature.\n* 'conds': a dictionary feature containing:\n\t+ 'column\\_index': a 'int32' feature.\n\t+ 'operator\\_index': a 'int32' feature.\n\t+ 'condition': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @lewtun, @ghomasHudson, @thomwolf for adding this dataset." ]
[ 105, 59, 10, 11, 6, 52, 17, 259, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 6, 29 ]
[ "passage: TAGS\n#task_categories-text2text-generation #annotations_creators-crowdsourced #language_creators-found #language_creators-machine-generated #multilinguality-monolingual #size_categories-10K<n<100K #source_datasets-original #language-English #license-unknown #text-to-sql #arxiv-1709.00103 #region-us \n### Dataset Summary\n\n\nA large crowd-sourced dataset for developing natural language interfaces for relational databases.\n\n\nWikiSQL is a dataset of 80654 hand-annotated examples\nof questions and SQL queries distributed across 24241 tables from Wikipedia.### Supported Tasks and Leaderboards### Languages\n\n\nDataset Structure\n-----------------### Data Instances#### default\n\n\n* Size of downloaded dataset files: 26.16 MB\n* Size of the generated dataset: 154.74 MB\n* Total amount of disk used: 180.90 MB\n\n\nAn example of 'validation' looks as follows.### Data Fields\n\n\nThe data fields are the same among all splits." ]
b08601e04326c79dfdd32d625aee71d232d685c3
# Dataset Card for "wikitext" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://blog.einstein.ai/the-wikitext-long-term-dependency-language-modeling-dataset/](https://blog.einstein.ai/the-wikitext-long-term-dependency-language-modeling-dataset/) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [Pointer Sentinel Mixture Models](https://arxiv.org/abs/1609.07843) - **Point of Contact:** [Stephen Merity](mailto:smerity@salesforce.com) - **Size of downloaded dataset files:** 391.41 MB - **Size of the generated dataset:** 1.12 GB - **Total amount of disk used:** 1.52 GB ### Dataset Summary The WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified Good and Featured articles on Wikipedia. The dataset is available under the Creative Commons Attribution-ShareAlike License. Compared to the preprocessed version of Penn Treebank (PTB), WikiText-2 is over 2 times larger and WikiText-103 is over 110 times larger. The WikiText dataset also features a far larger vocabulary and retains the original case, punctuation and numbers - all of which are removed in PTB. As it is composed of full articles, the dataset is well suited for models that can take advantage of long term dependencies. Each subset comes in two different variants: - Raw (for character level work) contain the raw tokens, before the addition of the <unk> (unknown) tokens. - Non-raw (for word level work) contain only the tokens in their vocabulary (wiki.train.tokens, wiki.valid.tokens, and wiki.test.tokens). The out-of-vocabulary tokens have been replaced with the the <unk> token. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### wikitext-103-raw-v1 - **Size of downloaded dataset files:** 191.98 MB - **Size of the generated dataset:** 549.42 MB - **Total amount of disk used:** 741.41 MB An example of 'validation' looks as follows. ``` This example was too long and was cropped: { "text": "\" The gold dollar or gold one @-@ dollar piece was a coin struck as a regular issue by the United States Bureau of the Mint from..." } ``` #### wikitext-103-v1 - **Size of downloaded dataset files:** 190.23 MB - **Size of the generated dataset:** 548.05 MB - **Total amount of disk used:** 738.27 MB An example of 'train' looks as follows. ``` This example was too long and was cropped: { "text": "\" Senjō no Valkyria 3 : <unk> Chronicles ( Japanese : 戦場のヴァルキュリア3 , lit . Valkyria of the Battlefield 3 ) , commonly referred to..." } ``` #### wikitext-2-raw-v1 - **Size of downloaded dataset files:** 4.72 MB - **Size of the generated dataset:** 13.54 MB - **Total amount of disk used:** 18.26 MB An example of 'train' looks as follows. ``` This example was too long and was cropped: { "text": "\" The Sinclair Scientific Programmable was introduced in 1975 , with the same case as the Sinclair Oxford . It was larger than t..." } ``` #### wikitext-2-v1 - **Size of downloaded dataset files:** 4.48 MB - **Size of the generated dataset:** 13.34 MB - **Total amount of disk used:** 17.82 MB An example of 'train' looks as follows. ``` This example was too long and was cropped: { "text": "\" Senjō no Valkyria 3 : <unk> Chronicles ( Japanese : 戦場のヴァルキュリア3 , lit . Valkyria of the Battlefield 3 ) , commonly referred to..." } ``` ### Data Fields The data fields are the same among all splits. #### wikitext-103-raw-v1 - `text`: a `string` feature. #### wikitext-103-v1 - `text`: a `string` feature. #### wikitext-2-raw-v1 - `text`: a `string` feature. #### wikitext-2-v1 - `text`: a `string` feature. ### Data Splits | name | train |validation|test| |-------------------|------:|---------:|---:| |wikitext-103-raw-v1|1801350| 3760|4358| |wikitext-103-v1 |1801350| 3760|4358| |wikitext-2-raw-v1 | 36718| 3760|4358| |wikitext-2-v1 | 36718| 3760|4358| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information The dataset is available under the [Creative Commons Attribution-ShareAlike License (CC BY-SA 4.0)](https://creativecommons.org/licenses/by-sa/4.0/). ### Citation Information ``` @misc{merity2016pointer, title={Pointer Sentinel Mixture Models}, author={Stephen Merity and Caiming Xiong and James Bradbury and Richard Socher}, year={2016}, eprint={1609.07843}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@lewtun](https://github.com/lewtun), [@patrickvonplaten](https://github.com/patrickvonplaten), [@mariamabarham](https://github.com/mariamabarham) for adding this dataset.
wikitext
[ "task_categories:text-generation", "task_categories:fill-mask", "task_ids:language-modeling", "task_ids:masked-language-modeling", "annotations_creators:no-annotation", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:1M<n<10M", "source_datasets:original", "language:en", "license:cc-by-sa-3.0", "license:gfdl", "arxiv:1609.07843", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["cc-by-sa-3.0", "gfdl"], "multilinguality": ["monolingual"], "size_categories": ["1M<n<10M"], "source_datasets": ["original"], "task_categories": ["text-generation", "fill-mask"], "task_ids": ["language-modeling", "masked-language-modeling"], "paperswithcode_id": "wikitext-2", "pretty_name": "WikiText", "dataset_info": [{"config_name": "wikitext-103-raw-v1", "features": [{"name": "text", "dtype": "string"}], "splits": [{"name": "test", "num_bytes": 1305088, "num_examples": 4358}, {"name": "train", "num_bytes": 546500949, "num_examples": 1801350}, {"name": "validation", "num_bytes": 1159288, "num_examples": 3760}], "download_size": 315466397, "dataset_size": 548965325}, {"config_name": "wikitext-103-v1", "features": [{"name": "text", "dtype": "string"}], "splits": [{"name": "test", "num_bytes": 1295575, "num_examples": 4358}, {"name": "train", "num_bytes": 545141915, "num_examples": 1801350}, {"name": "validation", "num_bytes": 1154751, "num_examples": 3760}], "download_size": 313093838, "dataset_size": 547592241}, {"config_name": "wikitext-2-raw-v1", "features": [{"name": "text", "dtype": "string"}], "splits": [{"name": "test", "num_bytes": 1305088, "num_examples": 4358}, {"name": "train", "num_bytes": 11061717, "num_examples": 36718}, {"name": "validation", "num_bytes": 1159288, "num_examples": 3760}], "download_size": 7747362, "dataset_size": 13526093}, {"config_name": "wikitext-2-v1", "features": [{"name": "text", "dtype": "string"}], "splits": [{"name": "test", "num_bytes": 1270947, "num_examples": 4358}, {"name": "train", "num_bytes": 10918118, "num_examples": 36718}, {"name": "validation", "num_bytes": 1134123, "num_examples": 3760}], "download_size": 7371282, "dataset_size": 13323188}], "configs": [{"config_name": "wikitext-103-raw-v1", "data_files": [{"split": "test", "path": "wikitext-103-raw-v1/test-*"}, {"split": "train", "path": "wikitext-103-raw-v1/train-*"}, {"split": "validation", "path": "wikitext-103-raw-v1/validation-*"}]}, {"config_name": "wikitext-103-v1", "data_files": [{"split": "test", "path": "wikitext-103-v1/test-*"}, {"split": "train", "path": "wikitext-103-v1/train-*"}, {"split": "validation", "path": "wikitext-103-v1/validation-*"}]}, {"config_name": "wikitext-2-raw-v1", "data_files": [{"split": "test", "path": "wikitext-2-raw-v1/test-*"}, {"split": "train", "path": "wikitext-2-raw-v1/train-*"}, {"split": "validation", "path": "wikitext-2-raw-v1/validation-*"}]}, {"config_name": "wikitext-2-v1", "data_files": [{"split": "test", "path": "wikitext-2-v1/test-*"}, {"split": "train", "path": "wikitext-2-v1/train-*"}, {"split": "validation", "path": "wikitext-2-v1/validation-*"}]}]}
2024-01-04T16:49:18+00:00
[ "1609.07843" ]
[ "en" ]
TAGS #task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-crowdsourced #multilinguality-monolingual #size_categories-1M<n<10M #source_datasets-original #language-English #license-cc-by-sa-3.0 #license-gfdl #arxiv-1609.07843 #region-us
Dataset Card for "wikitext" =========================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: Pointer Sentinel Mixture Models * Point of Contact: Stephen Merity * Size of downloaded dataset files: 391.41 MB * Size of the generated dataset: 1.12 GB * Total amount of disk used: 1.52 GB ### Dataset Summary The WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified Good and Featured articles on Wikipedia. The dataset is available under the Creative Commons Attribution-ShareAlike License. Compared to the preprocessed version of Penn Treebank (PTB), WikiText-2 is over 2 times larger and WikiText-103 is over 110 times larger. The WikiText dataset also features a far larger vocabulary and retains the original case, punctuation and numbers - all of which are removed in PTB. As it is composed of full articles, the dataset is well suited for models that can take advantage of long term dependencies. Each subset comes in two different variants: * Raw (for character level work) contain the raw tokens, before the addition of the (unknown) tokens. * Non-raw (for word level work) contain only the tokens in their vocabulary (URL, URL, and URL). The out-of-vocabulary tokens have been replaced with the the token. ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- ### Data Instances #### wikitext-103-raw-v1 * Size of downloaded dataset files: 191.98 MB * Size of the generated dataset: 549.42 MB * Total amount of disk used: 741.41 MB An example of 'validation' looks as follows. #### wikitext-103-v1 * Size of downloaded dataset files: 190.23 MB * Size of the generated dataset: 548.05 MB * Total amount of disk used: 738.27 MB An example of 'train' looks as follows. #### wikitext-2-raw-v1 * Size of downloaded dataset files: 4.72 MB * Size of the generated dataset: 13.54 MB * Total amount of disk used: 18.26 MB An example of 'train' looks as follows. #### wikitext-2-v1 * Size of downloaded dataset files: 4.48 MB * Size of the generated dataset: 13.34 MB * Total amount of disk used: 17.82 MB An example of 'train' looks as follows. ### Data Fields The data fields are the same among all splits. #### wikitext-103-raw-v1 * 'text': a 'string' feature. #### wikitext-103-v1 * 'text': a 'string' feature. #### wikitext-2-raw-v1 * 'text': a 'string' feature. #### wikitext-2-v1 * 'text': a 'string' feature. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information The dataset is available under the Creative Commons Attribution-ShareAlike License (CC BY-SA 4.0). ### Contributions Thanks to @thomwolf, @lewtun, @patrickvonplaten, @mariamabarham for adding this dataset.
[ "### Dataset Summary\n\n\nThe WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified\nGood and Featured articles on Wikipedia. The dataset is available under the Creative Commons Attribution-ShareAlike License.\n\n\nCompared to the preprocessed version of Penn Treebank (PTB), WikiText-2 is over 2 times larger and WikiText-103 is over\n110 times larger. The WikiText dataset also features a far larger vocabulary and retains the original case, punctuation\nand numbers - all of which are removed in PTB. As it is composed of full articles, the dataset is well suited for models\nthat can take advantage of long term dependencies.\n\n\nEach subset comes in two different variants:\n\n\n* Raw (for character level work) contain the raw tokens, before the addition of the (unknown) tokens.\n* Non-raw (for word level work) contain only the tokens in their vocabulary (URL, URL, and URL).\nThe out-of-vocabulary tokens have been replaced with the the token.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### wikitext-103-raw-v1\n\n\n* Size of downloaded dataset files: 191.98 MB\n* Size of the generated dataset: 549.42 MB\n* Total amount of disk used: 741.41 MB\n\n\nAn example of 'validation' looks as follows.", "#### wikitext-103-v1\n\n\n* Size of downloaded dataset files: 190.23 MB\n* Size of the generated dataset: 548.05 MB\n* Total amount of disk used: 738.27 MB\n\n\nAn example of 'train' looks as follows.", "#### wikitext-2-raw-v1\n\n\n* Size of downloaded dataset files: 4.72 MB\n* Size of the generated dataset: 13.54 MB\n* Total amount of disk used: 18.26 MB\n\n\nAn example of 'train' looks as follows.", "#### wikitext-2-v1\n\n\n* Size of downloaded dataset files: 4.48 MB\n* Size of the generated dataset: 13.34 MB\n* Total amount of disk used: 17.82 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### wikitext-103-raw-v1\n\n\n* 'text': a 'string' feature.", "#### wikitext-103-v1\n\n\n* 'text': a 'string' feature.", "#### wikitext-2-raw-v1\n\n\n* 'text': a 'string' feature.", "#### wikitext-2-v1\n\n\n* 'text': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nThe dataset is available under the Creative Commons Attribution-ShareAlike License (CC BY-SA 4.0).", "### Contributions\n\n\nThanks to @thomwolf, @lewtun, @patrickvonplaten, @mariamabarham for adding this dataset." ]
[ "TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-crowdsourced #multilinguality-monolingual #size_categories-1M<n<10M #source_datasets-original #language-English #license-cc-by-sa-3.0 #license-gfdl #arxiv-1609.07843 #region-us \n", "### Dataset Summary\n\n\nThe WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified\nGood and Featured articles on Wikipedia. The dataset is available under the Creative Commons Attribution-ShareAlike License.\n\n\nCompared to the preprocessed version of Penn Treebank (PTB), WikiText-2 is over 2 times larger and WikiText-103 is over\n110 times larger. The WikiText dataset also features a far larger vocabulary and retains the original case, punctuation\nand numbers - all of which are removed in PTB. As it is composed of full articles, the dataset is well suited for models\nthat can take advantage of long term dependencies.\n\n\nEach subset comes in two different variants:\n\n\n* Raw (for character level work) contain the raw tokens, before the addition of the (unknown) tokens.\n* Non-raw (for word level work) contain only the tokens in their vocabulary (URL, URL, and URL).\nThe out-of-vocabulary tokens have been replaced with the the token.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### wikitext-103-raw-v1\n\n\n* Size of downloaded dataset files: 191.98 MB\n* Size of the generated dataset: 549.42 MB\n* Total amount of disk used: 741.41 MB\n\n\nAn example of 'validation' looks as follows.", "#### wikitext-103-v1\n\n\n* Size of downloaded dataset files: 190.23 MB\n* Size of the generated dataset: 548.05 MB\n* Total amount of disk used: 738.27 MB\n\n\nAn example of 'train' looks as follows.", "#### wikitext-2-raw-v1\n\n\n* Size of downloaded dataset files: 4.72 MB\n* Size of the generated dataset: 13.54 MB\n* Total amount of disk used: 18.26 MB\n\n\nAn example of 'train' looks as follows.", "#### wikitext-2-v1\n\n\n* Size of downloaded dataset files: 4.48 MB\n* Size of the generated dataset: 13.34 MB\n* Total amount of disk used: 17.82 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### wikitext-103-raw-v1\n\n\n* 'text': a 'string' feature.", "#### wikitext-103-v1\n\n\n* 'text': a 'string' feature.", "#### wikitext-2-raw-v1\n\n\n* 'text': a 'string' feature.", "#### wikitext-2-v1\n\n\n* 'text': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information\n\n\nThe dataset is available under the Creative Commons Attribution-ShareAlike License (CC BY-SA 4.0).", "### Contributions\n\n\nThanks to @thomwolf, @lewtun, @patrickvonplaten, @mariamabarham for adding this dataset." ]
[ 134, 235, 10, 11, 6, 61, 59, 56, 54, 17, 21, 19, 21, 19, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 26, 34 ]
[ "passage: TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-crowdsourced #multilinguality-monolingual #size_categories-1M<n<10M #source_datasets-original #language-English #license-cc-by-sa-3.0 #license-gfdl #arxiv-1609.07843 #region-us \n### Dataset Summary\n\n\nThe WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified\nGood and Featured articles on Wikipedia. The dataset is available under the Creative Commons Attribution-ShareAlike License.\n\n\nCompared to the preprocessed version of Penn Treebank (PTB), WikiText-2 is over 2 times larger and WikiText-103 is over\n110 times larger. The WikiText dataset also features a far larger vocabulary and retains the original case, punctuation\nand numbers - all of which are removed in PTB. As it is composed of full articles, the dataset is well suited for models\nthat can take advantage of long term dependencies.\n\n\nEach subset comes in two different variants:\n\n\n* Raw (for character level work) contain the raw tokens, before the addition of the (unknown) tokens.\n* Non-raw (for word level work) contain only the tokens in their vocabulary (URL, URL, and URL).\nThe out-of-vocabulary tokens have been replaced with the the token.### Supported Tasks and Leaderboards### Languages\n\n\nDataset Structure\n-----------------### Data Instances#### wikitext-103-raw-v1\n\n\n* Size of downloaded dataset files: 191.98 MB\n* Size of the generated dataset: 549.42 MB\n* Total amount of disk used: 741.41 MB\n\n\nAn example of 'validation' looks as follows." ]
64b34b51e7a6551b5fa3acab48ba4cf9798b3f3b
# Dataset Card for WikiText-TL-39 ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Filipino Text Benchmarks](https://github.com/jcblaisecruz02/Filipino-Text-Benchmarks) - **Repository:** - **Paper:** [Evaluating language model finetuning techniques for low-resource languages](https://arxiv.org/abs/1907.00409) - **Leaderboard:** - **Point of Contact:** Jan Christian Blaise Cruz (jan_christian_cruz@dlsu.edu.ph) ### Dataset Summary Large scale, unlabeled text dataset with 39 Million tokens in the training set. Inspired by the original WikiText Long Term Dependency dataset (Merity et al., 2016). TL means "Tagalog." Published in Cruz & Cheng (2019). ### Supported Tasks and Leaderboards [More Information Needed] ### Languages Filipino/Tagalog ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields - `text` (`str`) The dataset is in plaintext and only has one field ("text") as it is compiled for language modeling. ### Data Splits Split | Documents | Tokens ------|-----------|------- Train | 120,975 | 39M Valid | 25,919 | 8M Test | 25,921 | 8M Please see the paper for more details on the dataset splits ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data Tagalog Wikipedia #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@jcblaisecruz02](https://github.com/jcblaisecruz02) for adding this dataset.
wikitext_tl39
[ "task_categories:text-generation", "task_categories:fill-mask", "task_ids:language-modeling", "task_ids:masked-language-modeling", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:monolingual", "size_categories:1M<n<10M", "source_datasets:original", "language:fil", "language:tl", "license:gpl-3.0", "arxiv:1907.00409", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["found"], "language": ["fil", "tl"], "license": ["gpl-3.0"], "multilinguality": ["monolingual"], "size_categories": ["1M<n<10M"], "source_datasets": ["original"], "task_categories": ["text-generation", "fill-mask"], "task_ids": ["language-modeling", "masked-language-modeling"], "paperswithcode_id": "wikitext-tl-39", "pretty_name": "WikiText-TL-39", "dataset_info": {"features": [{"name": "text", "dtype": "string"}], "config_name": "wikitext-tl-39", "splits": [{"name": "test", "num_bytes": 46182996, "num_examples": 376737}, {"name": "train", "num_bytes": 217182748, "num_examples": 1766072}, {"name": "validation", "num_bytes": 46256674, "num_examples": 381763}], "download_size": 116335234, "dataset_size": 309622418}}
2024-01-18T11:18:19+00:00
[ "1907.00409" ]
[ "fil", "tl" ]
TAGS #task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-1M<n<10M #source_datasets-original #language-Filipino #language-Tagalog #license-gpl-3.0 #arxiv-1907.00409 #region-us
Dataset Card for WikiText-TL-39 =============================== Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: Filipino Text Benchmarks * Repository: * Paper: Evaluating language model finetuning techniques for low-resource languages * Leaderboard: * Point of Contact: Jan Christian Blaise Cruz (jan\_christian\_cruz@URL) ### Dataset Summary Large scale, unlabeled text dataset with 39 Million tokens in the training set. Inspired by the original WikiText Long Term Dependency dataset (Merity et al., 2016). TL means "Tagalog." Published in Cruz & Cheng (2019). ### Supported Tasks and Leaderboards ### Languages Filipino/Tagalog Dataset Structure ----------------- ### Data Instances ### Data Fields * 'text' ('str') The dataset is in plaintext and only has one field ("text") as it is compiled for language modeling. ### Data Splits Split: Train, Documents: 120,975, Tokens: 39M Split: Valid, Documents: 25,919, Tokens: 8M Split: Test, Documents: 25,921, Tokens: 8M Please see the paper for more details on the dataset splits Dataset Creation ---------------- ### Curation Rationale ### Source Data Tagalog Wikipedia #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @jcblaisecruz02 for adding this dataset.
[ "### Dataset Summary\n\n\nLarge scale, unlabeled text dataset with 39 Million tokens in the training set. Inspired by the original WikiText Long Term Dependency dataset (Merity et al., 2016). TL means \"Tagalog.\" Published in Cruz & Cheng (2019).", "### Supported Tasks and Leaderboards", "### Languages\n\n\nFilipino/Tagalog\n\n\nDataset Structure\n-----------------", "### Data Instances", "### Data Fields\n\n\n* 'text' ('str')\n\n\nThe dataset is in plaintext and only has one field (\"text\") as it is compiled for language modeling.", "### Data Splits\n\n\nSplit: Train, Documents: 120,975, Tokens: 39M\nSplit: Valid, Documents: 25,919, Tokens: 8M\nSplit: Test, Documents: 25,921, Tokens: 8M\n\n\nPlease see the paper for more details on the dataset splits\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data\n\n\nTagalog Wikipedia", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @jcblaisecruz02 for adding this dataset." ]
[ "TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-1M<n<10M #source_datasets-original #language-Filipino #language-Tagalog #license-gpl-3.0 #arxiv-1907.00409 #region-us \n", "### Dataset Summary\n\n\nLarge scale, unlabeled text dataset with 39 Million tokens in the training set. Inspired by the original WikiText Long Term Dependency dataset (Merity et al., 2016). TL means \"Tagalog.\" Published in Cruz & Cheng (2019).", "### Supported Tasks and Leaderboards", "### Languages\n\n\nFilipino/Tagalog\n\n\nDataset Structure\n-----------------", "### Data Instances", "### Data Fields\n\n\n* 'text' ('str')\n\n\nThe dataset is in plaintext and only has one field (\"text\") as it is compiled for language modeling.", "### Data Splits\n\n\nSplit: Train, Documents: 120,975, Tokens: 39M\nSplit: Valid, Documents: 25,919, Tokens: 8M\nSplit: Test, Documents: 25,921, Tokens: 8M\n\n\nPlease see the paper for more details on the dataset splits\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data\n\n\nTagalog Wikipedia", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @jcblaisecruz02 for adding this dataset." ]
[ 128, 64, 10, 16, 6, 40, 74, 7, 6, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 6, 21 ]
[ "passage: TAGS\n#task_categories-text-generation #task_categories-fill-mask #task_ids-language-modeling #task_ids-masked-language-modeling #annotations_creators-no-annotation #language_creators-found #multilinguality-monolingual #size_categories-1M<n<10M #source_datasets-original #language-Filipino #language-Tagalog #license-gpl-3.0 #arxiv-1907.00409 #region-us \n### Dataset Summary\n\n\nLarge scale, unlabeled text dataset with 39 Million tokens in the training set. Inspired by the original WikiText Long Term Dependency dataset (Merity et al., 2016). TL means \"Tagalog.\" Published in Cruz & Cheng (2019).### Supported Tasks and Leaderboards### Languages\n\n\nFilipino/Tagalog\n\n\nDataset Structure\n-----------------### Data Instances### Data Fields\n\n\n* 'text' ('str')\n\n\nThe dataset is in plaintext and only has one field (\"text\") as it is compiled for language modeling.### Data Splits\n\n\nSplit: Train, Documents: 120,975, Tokens: 39M\nSplit: Valid, Documents: 25,919, Tokens: 8M\nSplit: Test, Documents: 25,921, Tokens: 8M\n\n\nPlease see the paper for more details on the dataset splits\n\n\nDataset Creation\n----------------### Curation Rationale### Source Data\n\n\nTagalog Wikipedia#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------### Social Impact of Dataset### Discussion of Biases### Other Known Limitations\n\n\nAdditional Information\n----------------------### Dataset Curators### Licensing Information### Contributions\n\n\nThanks to @jcblaisecruz02 for adding this dataset." ]
ead2d968bcdbb21b7f0a1cf32ccfe33f0e409bb0
# Dataset Card for wili_2018 ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://zenodo.org/record/841984 - **Repository:** [Needs More Information] - **Paper:** https://arxiv.org/pdf/1801.07779 - **Leaderboard:** [Needs More Information] - **Point of Contact:** Thoma, Martin (Email: info@martin-thoma.de) ### Dataset Summary WiLI-2018, the Wikipedia language identification benchmark dataset, contains 235000 paragraphs of 235 languages. The dataset is balanced and a train-test split is provided. ### Supported Tasks and Leaderboards [Needs More Information] ### Languages 235 Different Languages ## Dataset Structure ### Data Instances ``` { 'label': 207, 'sentence': 'Ti Turkia ket maysa a demokrata, sekular, unitario, batay-linteg a republika nga addaan ti taga-ugma a tinawtawid a kultura. Ti Turkia ket umadadu a naipatipon iti Laud babaen ti panagkameng kadagiti organisasion a kas ti Konsilo iti Europa, NATO, OECD, OSCE ken ti G-20 a dagiti kangrunaan nga ekonomia. Ti Turkia ket nangrugi a nakitulag ti napno a panagkameng iti Kappon ti Europa idi 2005, nga isu ket maysa idin a kumaduaan a kameng iti Europeano a Komunidad ti Ekonomia manipud idi 1963 ken nakadanon ti maysa a tulagan ti kappon ti aduana idi 1995. Ti Turkia ket nagtaraken iti asideg a kultural, politikal, ekonomiko ken industria a panakibiang iti Tengnga a Daya, dagiti Turko nga estado iti Tengnga nga Asia ken dagiti pagilian ti Aprika babaen ti panagkameng kadagiti organisasion a kas ti Turko a Konsilo, Nagsaupan nga Administrasion iti Turko nga Arte ken Kultura, Organisasion iti Islamiko a Panagtitinnulong ken ti Organisasion ti Ekonomiko a Panagtitinnulong.' } ``` ### Data Fields [Needs More Information] ### Data Splits 175000 lines of text each for train and test data. ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators The dataset was initially created by Thomas Martin ### Licensing Information ODC Open Database License v1.0 ### Citation Information ``` @dataset{thoma_martin_2018_841984, author = {Thoma, Martin}, title = {{WiLI-2018 - Wikipedia Language Identification database}}, month = jan, year = 2018, publisher = {Zenodo}, version = {1.0.0}, doi = {10.5281/zenodo.841984}, url = {https://doi.org/10.5281/zenodo.841984} } ``` ### Contributions Thanks to [@Shubhambindal2017](https://github.com/Shubhambindal2017) for adding this dataset.
wili_2018
[ "task_categories:text-classification", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:multilingual", "size_categories:100K<n<1M", "source_datasets:original", "language:ace", "language:af", "language:als", "language:am", "language:an", "language:ang", "language:ar", "language:arz", "language:as", "language:ast", "language:av", "language:ay", "language:az", "language:azb", "language:ba", "language:bar", "language:bcl", "language:be", "language:bg", "language:bho", "language:bjn", "language:bn", "language:bo", "language:bpy", "language:br", "language:bs", "language:bxr", "language:ca", "language:cbk", "language:cdo", "language:ce", "language:ceb", "language:chr", "language:ckb", "language:co", "language:crh", "language:cs", "language:csb", "language:cv", "language:cy", "language:da", "language:de", "language:diq", "language:dsb", "language:dty", "language:dv", "language:egl", "language:el", "language:en", "language:eo", "language:es", "language:et", "language:eu", "language:ext", "language:fa", "language:fi", "language:fo", "language:fr", "language:frp", "language:fur", "language:fy", "language:ga", "language:gag", "language:gd", "language:gl", "language:glk", "language:gn", "language:gu", "language:gv", "language:ha", "language:hak", "language:he", "language:hi", "language:hif", "language:hr", "language:hsb", "language:ht", "language:hu", "language:hy", "language:ia", "language:id", "language:ie", "language:ig", "language:ilo", "language:io", "language:is", "language:it", "language:ja", "language:jam", "language:jbo", "language:jv", "language:ka", "language:kaa", "language:kab", "language:kbd", "language:kk", "language:km", "language:kn", "language:ko", "language:koi", "language:kok", "language:krc", "language:ksh", "language:ku", "language:kv", "language:kw", "language:ky", "language:la", "language:lad", "language:lb", "language:lez", "language:lg", "language:li", "language:lij", "language:lmo", "language:ln", "language:lo", "language:lrc", "language:lt", "language:ltg", "language:lv", "language:lzh", "language:mai", "language:map", "language:mdf", "language:mg", "language:mhr", "language:mi", "language:min", "language:mk", "language:ml", "language:mn", "language:mr", "language:mrj", "language:ms", "language:mt", "language:mwl", "language:my", "language:myv", "language:mzn", "language:nan", "language:nap", "language:nb", "language:nci", "language:nds", "language:ne", "language:new", "language:nl", "language:nn", "language:nrm", "language:nso", "language:nv", "language:oc", "language:olo", "language:om", "language:or", "language:os", "language:pa", "language:pag", "language:pam", "language:pap", "language:pcd", "language:pdc", "language:pfl", "language:pl", "language:pnb", "language:ps", "language:pt", "language:qu", "language:rm", "language:ro", "language:roa", "language:ru", "language:rue", "language:rup", "language:rw", "language:sa", "language:sah", "language:sc", "language:scn", "language:sco", "language:sd", "language:sgs", "language:sh", "language:si", "language:sk", "language:sl", "language:sme", "language:sn", "language:so", "language:sq", "language:sr", "language:srn", "language:stq", "language:su", "language:sv", "language:sw", "language:szl", "language:ta", "language:tcy", "language:te", "language:tet", "language:tg", "language:th", "language:tk", "language:tl", "language:tn", "language:to", "language:tr", "language:tt", "language:tyv", "language:udm", "language:ug", "language:uk", "language:ur", "language:uz", "language:vec", "language:vep", "language:vi", "language:vls", "language:vo", "language:vro", "language:wa", "language:war", "language:wo", "language:wuu", "language:xh", "language:xmf", "language:yi", "language:yo", "language:zea", "language:zh", "license:odbl", "language-identification", "arxiv:1801.07779", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["found"], "language": ["ace", "af", "als", "am", "an", "ang", "ar", "arz", "as", "ast", "av", "ay", "az", "azb", "ba", "bar", "bcl", "be", "bg", "bho", "bjn", "bn", "bo", "bpy", "br", "bs", "bxr", "ca", "cbk", "cdo", "ce", "ceb", "chr", "ckb", "co", "crh", "cs", "csb", "cv", "cy", "da", "de", "diq", "dsb", "dty", "dv", "egl", "el", "en", "eo", "es", "et", "eu", "ext", "fa", "fi", "fo", "fr", "frp", "fur", "fy", "ga", "gag", "gd", "gl", "glk", "gn", "gu", "gv", "ha", "hak", "he", "hi", "hif", "hr", "hsb", "ht", "hu", "hy", "ia", "id", "ie", "ig", "ilo", "io", "is", "it", "ja", "jam", "jbo", "jv", "ka", "kaa", "kab", "kbd", "kk", "km", "kn", "ko", "koi", "kok", "krc", "ksh", "ku", "kv", "kw", "ky", "la", "lad", "lb", "lez", "lg", "li", "lij", "lmo", "ln", "lo", "lrc", "lt", "ltg", "lv", "lzh", "mai", "map", "mdf", "mg", "mhr", "mi", "min", "mk", "ml", "mn", "mr", "mrj", "ms", "mt", "mwl", "my", "myv", "mzn", "nan", "nap", "nb", "nci", "nds", "ne", "new", "nl", "nn", "nrm", "nso", "nv", "oc", "olo", "om", "or", "os", "pa", "pag", "pam", "pap", "pcd", "pdc", "pfl", "pl", "pnb", "ps", "pt", "qu", "rm", "ro", "roa", "ru", "rue", "rup", "rw", "sa", "sah", "sc", "scn", "sco", "sd", "sgs", "sh", "si", "sk", "sl", "sme", "sn", "so", "sq", "sr", "srn", "stq", "su", "sv", "sw", "szl", "ta", "tcy", "te", "tet", "tg", "th", "tk", "tl", "tn", "to", "tr", "tt", "tyv", "udm", "ug", "uk", "ur", "uz", "vec", "vep", "vi", "vls", "vo", "vro", "wa", "war", "wo", "wuu", "xh", "xmf", "yi", "yo", "zea", "zh"], "license": ["odbl"], "multilinguality": ["multilingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": [], "paperswithcode_id": "wili-2018", "pretty_name": "Wili2018", "language_bcp47": ["be-tarask", "map-bms", "nds-nl", "roa-tara", "zh-yue"], "tags": ["language-identification"], "dataset_info": {"features": [{"name": "sentence", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "cdo", "1": "glk", "2": "jam", "3": "lug", "4": "san", "5": "rue", "6": "wol", "7": "new", "8": "mwl", "9": "bre", "10": "ara", "11": "hye", "12": "xmf", "13": "ext", "14": "cor", "15": "yor", "16": "div", "17": "asm", "18": "lat", "19": "cym", "20": "hif", "21": "ace", "22": "kbd", "23": "tgk", "24": "rus", "25": "nso", "26": "mya", "27": "msa", "28": "ava", "29": "cbk", "30": "urd", "31": "deu", "32": "swa", "33": "pus", "34": "bxr", "35": "udm", "36": "csb", "37": "yid", "38": "vro", "39": "por", "40": "pdc", "41": "eng", "42": "tha", "43": "hat", "44": "lmo", "45": "pag", "46": "jav", "47": "chv", "48": "nan", "49": "sco", "50": "kat", "51": "bho", "52": "bos", "53": "kok", "54": "oss", "55": "mri", "56": "fry", "57": "cat", "58": "azb", "59": "kin", "60": "hin", "61": "sna", "62": "dan", "63": "egl", "64": "mkd", "65": "ron", "66": "bul", "67": "hrv", "68": "som", "69": "pam", "70": "nav", "71": "ksh", "72": "nci", "73": "khm", "74": "sgs", "75": "srn", "76": "bar", "77": "cos", "78": "ckb", "79": "pfl", "80": "arz", "81": "roa-tara", "82": "fra", "83": "mai", "84": "zh-yue", "85": "guj", "86": "fin", "87": "kir", "88": "vol", "89": "hau", "90": "afr", "91": "uig", "92": "lao", "93": "swe", "94": "slv", "95": "kor", "96": "szl", "97": "srp", "98": "dty", "99": "nrm", "100": "dsb", "101": "ind", "102": "wln", "103": "pnb", "104": "ukr", "105": "bpy", "106": "vie", "107": "tur", "108": "aym", "109": "lit", "110": "zea", "111": "pol", "112": "est", "113": "scn", "114": "vls", "115": "stq", "116": "gag", "117": "grn", "118": "kaz", "119": "ben", "120": "pcd", "121": "bjn", "122": "krc", "123": "amh", "124": "diq", "125": "ltz", "126": "ita", "127": "kab", "128": "bel", "129": "ang", "130": "mhr", "131": "che", "132": "koi", "133": "glv", "134": "ido", "135": "fao", "136": "bak", "137": "isl", "138": "bcl", "139": "tet", "140": "jpn", "141": "kur", "142": "map-bms", "143": "tyv", "144": "olo", "145": "arg", "146": "ori", "147": "lim", "148": "tel", "149": "lin", "150": "roh", "151": "sqi", "152": "xho", "153": "mlg", "154": "fas", "155": "hbs", "156": "tam", "157": "aze", "158": "lad", "159": "nob", "160": "sin", "161": "gla", "162": "nap", "163": "snd", "164": "ast", "165": "mal", "166": "mdf", "167": "tsn", "168": "nds", "169": "tgl", "170": "nno", "171": "sun", "172": "lzh", "173": "jbo", "174": "crh", "175": "pap", "176": "oci", "177": "hak", "178": "uzb", "179": "zho", "180": "hsb", "181": "sme", "182": "mlt", "183": "vep", "184": "lez", "185": "nld", "186": "nds-nl", "187": "mrj", "188": "spa", "189": "ceb", "190": "ina", "191": "heb", "192": "hun", "193": "que", "194": "kaa", "195": "mar", "196": "vec", "197": "frp", "198": "ell", "199": "sah", "200": "eus", "201": "ces", "202": "slk", "203": "chr", "204": "lij", "205": "nep", "206": "srd", "207": "ilo", "208": "be-tarask", "209": "bod", "210": "orm", "211": "war", "212": "glg", "213": "mon", "214": "gle", "215": "min", "216": "ibo", "217": "ile", "218": "epo", "219": "lav", "220": "lrc", "221": "als", "222": "mzn", "223": "rup", "224": "fur", "225": "tat", "226": "myv", "227": "pan", "228": "ton", "229": "kom", "230": "wuu", "231": "tcy", "232": "tuk", "233": "kan", "234": "ltg"}}}}], "config_name": "WiLI-2018 dataset", "splits": [{"name": "train", "num_bytes": 65408201, "num_examples": 117500}, {"name": "test", "num_bytes": 66491260, "num_examples": 117500}], "download_size": 130516351, "dataset_size": 131899461}}
2024-01-18T11:18:20+00:00
[ "1801.07779" ]
[ "ace", "af", "als", "am", "an", "ang", "ar", "arz", "as", "ast", "av", "ay", "az", "azb", "ba", "bar", "bcl", "be", "bg", "bho", "bjn", "bn", "bo", "bpy", "br", "bs", "bxr", "ca", "cbk", "cdo", "ce", "ceb", "chr", "ckb", "co", "crh", "cs", "csb", "cv", "cy", "da", "de", "diq", "dsb", "dty", "dv", "egl", "el", "en", "eo", "es", "et", "eu", "ext", "fa", "fi", "fo", "fr", "frp", "fur", "fy", "ga", "gag", "gd", "gl", "glk", "gn", "gu", "gv", "ha", "hak", "he", "hi", "hif", "hr", "hsb", "ht", "hu", "hy", "ia", "id", "ie", "ig", "ilo", "io", "is", "it", "ja", "jam", "jbo", "jv", "ka", "kaa", "kab", "kbd", "kk", "km", "kn", "ko", "koi", "kok", "krc", "ksh", "ku", "kv", "kw", "ky", "la", "lad", "lb", "lez", "lg", "li", "lij", "lmo", "ln", "lo", "lrc", "lt", "ltg", "lv", "lzh", "mai", "map", "mdf", "mg", "mhr", "mi", "min", "mk", "ml", "mn", "mr", "mrj", "ms", "mt", "mwl", "my", "myv", "mzn", "nan", "nap", "nb", "nci", "nds", "ne", "new", "nl", "nn", "nrm", "nso", "nv", "oc", "olo", "om", "or", "os", "pa", "pag", "pam", "pap", "pcd", "pdc", "pfl", "pl", "pnb", "ps", "pt", "qu", "rm", "ro", "roa", "ru", "rue", "rup", "rw", "sa", "sah", "sc", "scn", "sco", "sd", "sgs", "sh", "si", "sk", "sl", "sme", "sn", "so", "sq", "sr", "srn", "stq", "su", "sv", "sw", "szl", "ta", "tcy", "te", "tet", "tg", "th", "tk", "tl", "tn", "to", "tr", "tt", "tyv", "udm", "ug", "uk", "ur", "uz", "vec", "vep", "vi", "vls", "vo", "vro", "wa", "war", "wo", "wuu", "xh", "xmf", "yi", "yo", "zea", "zh" ]
TAGS #task_categories-text-classification #annotations_creators-no-annotation #language_creators-found #multilinguality-multilingual #size_categories-100K<n<1M #source_datasets-original #language-Achinese #language-Afrikaans #language-Tosk Albanian #language-Amharic #language-Aragonese #language-Old English (ca. 450-1100) #language-Arabic #language-Egyptian Arabic #language-Assamese #language-Asturian #language-Avaric #language-Aymara #language-Azerbaijani #language-South Azerbaijani #language-Bashkir #language-Bavarian #language-Central Bikol #language-Belarusian #language-Bulgarian #language-Bhojpuri #language-Banjar #language-Bengali #language-Tibetan #language-Bishnupriya #language-Breton #language-Bosnian #language-Russia Buriat #language-Catalan #language-Chavacano #language-Min Dong Chinese #language-Chechen #language-Cebuano #language-Cherokee #language-Central Kurdish #language-Corsican #language-Crimean Tatar #language-Czech #language-Kashubian #language-Chuvash #language-Welsh #language-Danish #language-German #language-Dimli (individual language) #language-Lower Sorbian #language-Dotyali #language-Dhivehi #language-Emilian #language-Modern Greek (1453-) #language-English #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Extremaduran #language-Persian #language-Finnish #language-Faroese #language-French #language-Arpitan #language-Friulian #language-Western Frisian #language-Irish #language-Gagauz #language-Scottish Gaelic #language-Galician #language-Gilaki #language-Guarani #language-Gujarati #language-Manx #language-Hausa #language-Hakka Chinese #language-Hebrew #language-Hindi #language-Fiji Hindi #language-Croatian #language-Upper Sorbian #language-Haitian #language-Hungarian #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Indonesian #language-Interlingue #language-Igbo #language-Iloko #language-Ido #language-Icelandic #language-Italian #language-Japanese #language-Jamaican Creole English #language-Lojban #language-Javanese #language-Georgian #language-Kara-Kalpak #language-Kabyle #language-Kabardian #language-Kazakh #language-Khmer #language-Kannada #language-Korean #language-Komi-Permyak #language-Konkani (macrolanguage) #language-Karachay-Balkar #language-Kölsch #language-Kurdish #language-Komi #language-Cornish #language-Kirghiz #language-Latin #language-Ladino #language-Luxembourgish #language-Lezghian #language-Ganda #language-Limburgan #language-Ligurian #language-Lombard #language-Lingala #language-Lao #language-Northern Luri #language-Lithuanian #language-Latgalian #language-Latvian #language-Literary Chinese #language-Maithili #language-map #language-Moksha #language-Malagasy #language-Eastern Mari #language-Maori #language-Minangkabau #language-Macedonian #language-Malayalam #language-Mongolian #language-Marathi #language-Western Mari #language-Malay (macrolanguage) #language-Maltese #language-Mirandese #language-Burmese #language-Erzya #language-Mazanderani #language-Min Nan Chinese #language-Neapolitan #language-Norwegian Bokmål #language-Classical Nahuatl #language-Low German #language-Nepali (macrolanguage) #language-Newari #language-Dutch #language-Norwegian Nynorsk #language-Narom #language-Pedi #language-Navajo #language-Occitan (post 1500) #language-Livvi #language-Oromo #language-Oriya (macrolanguage) #language-Ossetian #language-Panjabi #language-Pangasinan #language-Pampanga #language-Papiamento #language-Picard #language-Pennsylvania German #language-Pfaelzisch #language-Polish #language-Western Panjabi #language-Pushto #language-Portuguese #language-Quechua #language-Romansh #language-Romanian #language-roa #language-Russian #language-Rusyn #language-Macedo-Romanian #language-Kinyarwanda #language-Sanskrit #language-Yakut #language-Sardinian #language-Sicilian #language-Scots #language-Sindhi #language-Samogitian #language-Serbo-Croatian #language-Sinhala #language-Slovak #language-Slovenian #language-Northern Sami #language-Shona #language-Somali #language-Albanian #language-Serbian #language-Sranan Tongo #language-Saterfriesisch #language-Sundanese #language-Swedish #language-Swahili (macrolanguage) #language-Silesian #language-Tamil #language-Tulu #language-Telugu #language-Tetum #language-Tajik #language-Thai #language-Turkmen #language-Tagalog #language-Tswana #language-Tonga (Tonga Islands) #language-Turkish #language-Tatar #language-Tuvinian #language-Udmurt #language-Uighur #language-Ukrainian #language-Urdu #language-Uzbek #language-Venetian #language-Veps #language-Vietnamese #language-Vlaams #language-Volapük #language-Võro #language-Walloon #language-Waray (Philippines) #language-Wolof #language-Wu Chinese #language-Xhosa #language-Mingrelian #language-Yiddish #language-Yoruba #language-Zeeuws #language-Chinese #license-odbl #language-identification #arxiv-1801.07779 #region-us
# Dataset Card for wili_2018 ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: URL - Leaderboard: - Point of Contact: Thoma, Martin (Email: info@URL) ### Dataset Summary WiLI-2018, the Wikipedia language identification benchmark dataset, contains 235000 paragraphs of 235 languages. The dataset is balanced and a train-test split is provided. ### Supported Tasks and Leaderboards ### Languages 235 Different Languages ## Dataset Structure ### Data Instances ### Data Fields ### Data Splits 175000 lines of text each for train and test data. ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators The dataset was initially created by Thomas Martin ### Licensing Information ODC Open Database License v1.0 ### Contributions Thanks to @Shubhambindal2017 for adding this dataset.
[ "# Dataset Card for wili_2018", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: URL\n- Leaderboard: \n- Point of Contact: Thoma, Martin (Email: info@URL)", "### Dataset Summary\n\nWiLI-2018, the Wikipedia language identification benchmark dataset, contains 235000 paragraphs of 235 languages. The dataset is balanced and a train-test split is provided.", "### Supported Tasks and Leaderboards", "### Languages\n\n235 Different Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits\n\n175000 lines of text each for train and test data.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThe dataset was initially created by Thomas Martin", "### Licensing Information\n\nODC Open Database License v1.0", "### Contributions\n\nThanks to @Shubhambindal2017 for adding this dataset." ]
[ "TAGS\n#task_categories-text-classification #annotations_creators-no-annotation #language_creators-found #multilinguality-multilingual #size_categories-100K<n<1M #source_datasets-original #language-Achinese #language-Afrikaans #language-Tosk Albanian #language-Amharic #language-Aragonese #language-Old English (ca. 450-1100) #language-Arabic #language-Egyptian Arabic #language-Assamese #language-Asturian #language-Avaric #language-Aymara #language-Azerbaijani #language-South Azerbaijani #language-Bashkir #language-Bavarian #language-Central Bikol #language-Belarusian #language-Bulgarian #language-Bhojpuri #language-Banjar #language-Bengali #language-Tibetan #language-Bishnupriya #language-Breton #language-Bosnian #language-Russia Buriat #language-Catalan #language-Chavacano #language-Min Dong Chinese #language-Chechen #language-Cebuano #language-Cherokee #language-Central Kurdish #language-Corsican #language-Crimean Tatar #language-Czech #language-Kashubian #language-Chuvash #language-Welsh #language-Danish #language-German #language-Dimli (individual language) #language-Lower Sorbian #language-Dotyali #language-Dhivehi #language-Emilian #language-Modern Greek (1453-) #language-English #language-Esperanto #language-Spanish #language-Estonian #language-Basque #language-Extremaduran #language-Persian #language-Finnish #language-Faroese #language-French #language-Arpitan #language-Friulian #language-Western Frisian #language-Irish #language-Gagauz #language-Scottish Gaelic #language-Galician #language-Gilaki #language-Guarani #language-Gujarati #language-Manx #language-Hausa #language-Hakka Chinese #language-Hebrew #language-Hindi #language-Fiji Hindi #language-Croatian #language-Upper Sorbian #language-Haitian #language-Hungarian #language-Armenian #language-Interlingua (International Auxiliary Language Association) #language-Indonesian #language-Interlingue #language-Igbo #language-Iloko #language-Ido #language-Icelandic #language-Italian #language-Japanese #language-Jamaican Creole English #language-Lojban #language-Javanese #language-Georgian #language-Kara-Kalpak #language-Kabyle #language-Kabardian #language-Kazakh #language-Khmer #language-Kannada #language-Korean #language-Komi-Permyak #language-Konkani (macrolanguage) #language-Karachay-Balkar #language-Kölsch #language-Kurdish #language-Komi #language-Cornish #language-Kirghiz #language-Latin #language-Ladino #language-Luxembourgish #language-Lezghian #language-Ganda #language-Limburgan #language-Ligurian #language-Lombard #language-Lingala #language-Lao #language-Northern Luri #language-Lithuanian #language-Latgalian #language-Latvian #language-Literary Chinese #language-Maithili #language-map #language-Moksha #language-Malagasy #language-Eastern Mari #language-Maori #language-Minangkabau #language-Macedonian #language-Malayalam #language-Mongolian #language-Marathi #language-Western Mari #language-Malay (macrolanguage) #language-Maltese #language-Mirandese #language-Burmese #language-Erzya #language-Mazanderani #language-Min Nan Chinese #language-Neapolitan #language-Norwegian Bokmål #language-Classical Nahuatl #language-Low German #language-Nepali (macrolanguage) #language-Newari #language-Dutch #language-Norwegian Nynorsk #language-Narom #language-Pedi #language-Navajo #language-Occitan (post 1500) #language-Livvi #language-Oromo #language-Oriya (macrolanguage) #language-Ossetian #language-Panjabi #language-Pangasinan #language-Pampanga #language-Papiamento #language-Picard #language-Pennsylvania German #language-Pfaelzisch #language-Polish #language-Western Panjabi #language-Pushto #language-Portuguese #language-Quechua #language-Romansh #language-Romanian #language-roa #language-Russian #language-Rusyn #language-Macedo-Romanian #language-Kinyarwanda #language-Sanskrit #language-Yakut #language-Sardinian #language-Sicilian #language-Scots #language-Sindhi #language-Samogitian #language-Serbo-Croatian #language-Sinhala #language-Slovak #language-Slovenian #language-Northern Sami #language-Shona #language-Somali #language-Albanian #language-Serbian #language-Sranan Tongo #language-Saterfriesisch #language-Sundanese #language-Swedish #language-Swahili (macrolanguage) #language-Silesian #language-Tamil #language-Tulu #language-Telugu #language-Tetum #language-Tajik #language-Thai #language-Turkmen #language-Tagalog #language-Tswana #language-Tonga (Tonga Islands) #language-Turkish #language-Tatar #language-Tuvinian #language-Udmurt #language-Uighur #language-Ukrainian #language-Urdu #language-Uzbek #language-Venetian #language-Veps #language-Vietnamese #language-Vlaams #language-Volapük #language-Võro #language-Walloon #language-Waray (Philippines) #language-Wolof #language-Wu Chinese #language-Xhosa #language-Mingrelian #language-Yiddish #language-Yoruba #language-Zeeuws #language-Chinese #license-odbl #language-identification #arxiv-1801.07779 #region-us \n", "# Dataset Card for wili_2018", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: URL\n- Leaderboard: \n- Point of Contact: Thoma, Martin (Email: info@URL)", "### Dataset Summary\n\nWiLI-2018, the Wikipedia language identification benchmark dataset, contains 235000 paragraphs of 235 languages. The dataset is balanced and a train-test split is provided.", "### Supported Tasks and Leaderboards", "### Languages\n\n235 Different Languages", "## Dataset Structure", "### Data Instances", "### Data Fields", "### Data Splits\n\n175000 lines of text each for train and test data.", "## Dataset Creation", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThe dataset was initially created by Thomas Martin", "### Licensing Information\n\nODC Open Database License v1.0", "### Contributions\n\nThanks to @Shubhambindal2017 for adding this dataset." ]
[ 1521, 9, 120, 37, 45, 10, 8, 6, 6, 5, 17, 5, 7, 4, 10, 10, 5, 5, 9, 8, 8, 7, 8, 7, 5, 16, 13, 20 ]
[ "passage: " ]
3f31267586e4408e3b3f77ec22198fd24ea8dc1d
# Dataset Card for Wino_Bias dataset ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [WinoBias](https://uclanlp.github.io/corefBias/overview) - **Repository:** - **Paper:** [Arxiv](https://arxiv.org/abs/1804.06876) - **Leaderboard:** - **Point of Contact:** ### Dataset Summary WinoBias, a Winograd-schema dataset for coreference resolution focused on gender bias. The corpus contains Winograd-schema style sentences with entities corresponding to people referred by their occupation (e.g. the nurse, the doctor, the carpenter). ### Supported Tasks and Leaderboards The underlying task is coreference resolution. ### Languages English ## Dataset Structure ### Data Instances The dataset has 4 subsets: `type1_pro`, `type1_anti`, `type2_pro` and `type2_anti`. The `*_pro` subsets contain sentences that reinforce gender stereotypes (e.g. mechanics are male, nurses are female), whereas the `*_anti` datasets contain "anti-stereotypical" sentences (e.g. mechanics are female, nurses are male). The `type1` (*WB-Knowledge*) subsets contain sentences for which world knowledge is necessary to resolve the co-references, and `type2` (*WB-Syntax*) subsets require only the syntactic information present in the sentence to resolve them. ### Data Fields - document_id = This is a variation on the document filename - part_number = Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. - word_num = This is the word index of the word in that sentence. - tokens = This is the token as segmented/tokenized in the Treebank. - pos_tags = This is the Penn Treebank style part of speech. When parse information is missing, all part of speeches except the one for which there is some sense or proposition annotation are marked with a XX tag. The verb is marked with just a VERB tag. - parse_bit = This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. When the parse information is missing, the first word of a sentence is tagged as "(TOP*" and the last word is tagged as "*)" and all intermediate words are tagged with a "*". - predicate_lemma = The predicate lemma is mentioned for the rows for which we have semantic role information or word sense information. All other rows are marked with a "-". - predicate_framenet_id = This is the PropBank frameset ID of the predicate in predicate_lemma. - word_sense = This is the word sense of the word in Column tokens. - speaker = This is the speaker or author name where available. - ner_tags = These columns identifies the spans representing various named entities. For documents which do not have named entity annotation, each line is represented with an "*". - verbal_predicates = There is one column each of predicate argument structure information for the predicate mentioned in predicate_lemma. If there are no predicates tagged in a sentence this is a single column with all rows marked with an "*". ### Data Splits Dev and Test Split available ## Dataset Creation ### Curation Rationale The WinoBias dataset was introduced in 2018 (see [paper](https://arxiv.org/abs/1804.06876)), with its original task being *coreference resolution*, which is a task that aims to identify mentions that refer to the same entity or person. ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? The dataset was created by researchers familiar with the WinoBias project, based on two prototypical templates provided by the authors, in which entities interact in plausible ways. ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? "Researchers familiar with the [WinoBias] project" ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [Recent work](https://www.microsoft.com/en-us/research/uploads/prod/2021/06/The_Salmon_paper.pdf) has shown that this dataset contains grammatical issues, incorrect or ambiguous labels, and stereotype conflation, among other limitations. ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez and Kai-Wei Chan ### Licensing Information MIT Licence ### Citation Information @article{DBLP:journals/corr/abs-1804-06876, author = {Jieyu Zhao and Tianlu Wang and Mark Yatskar and Vicente Ordonez and Kai{-}Wei Chang}, title = {Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods}, journal = {CoRR}, volume = {abs/1804.06876}, year = {2018}, url = {http://arxiv.org/abs/1804.06876}, archivePrefix = {arXiv}, eprint = {1804.06876}, timestamp = {Mon, 13 Aug 2018 16:47:01 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-1804-06876.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ### Contributions Thanks to [@akshayb7](https://github.com/akshayb7) for adding this dataset. Updated by [@JieyuZhao](https://github.com/JieyuZhao).
wino_bias
[ "task_categories:token-classification", "task_ids:coreference-resolution", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:mit", "arxiv:1804.06876", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["expert-generated"], "language": ["en"], "license": ["mit"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["token-classification"], "task_ids": ["coreference-resolution"], "paperswithcode_id": "winobias", "pretty_name": "WinoBias", "dataset_info": [{"config_name": "type1_anti", "features": [{"name": "document_id", "dtype": "string"}, {"name": "part_number", "dtype": "string"}, {"name": "word_number", "sequence": "int32"}, {"name": "tokens", "sequence": "string"}, {"name": "pos_tags", "sequence": {"class_label": {"names": {"0": "\"", "1": "''", "2": "#", "3": "$", "4": "(", "5": ")", "6": ",", "7": ".", "8": ":", "9": "``", "10": "CC", "11": "CD", "12": "DT", "13": "EX", "14": "FW", "15": "IN", "16": "JJ", "17": "JJR", "18": "JJS", "19": "LS", "20": "MD", "21": "NN", "22": "NNP", "23": "NNPS", "24": "NNS", "25": "NN|SYM", "26": "PDT", "27": "POS", "28": "PRP", "29": "PRP$", "30": "RB", "31": "RBR", "32": "RBS", "33": "RP", "34": "SYM", "35": "TO", "36": "UH", "37": "VB", "38": "VBD", "39": "VBG", "40": "VBN", "41": "VBP", "42": "VBZ", "43": "WDT", "44": "WP", "45": "WP$", "46": "WRB", "47": "HYPH", "48": "XX", "49": "NFP", "50": "AFX", "51": "ADD", "52": "-LRB-", "53": "-RRB-", "54": "-"}}}}, {"name": "parse_bit", "sequence": "string"}, {"name": "predicate_lemma", "sequence": "string"}, {"name": "predicate_framenet_id", "sequence": "string"}, {"name": "word_sense", "sequence": "string"}, {"name": "speaker", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "B-PERSON", "1": "I-PERSON", "2": "B-NORP", "3": "I-NORP", "4": "B-FAC", "5": "I-FAC", "6": "B-ORG", "7": "I-ORG", "8": "B-GPE", "9": "I-GPE", "10": "B-LOC", "11": "I-LOC", "12": "B-PRODUCT", "13": "I-PRODUCT", "14": "B-EVENT", "15": "I-EVENT", "16": "B-WORK_OF_ART", "17": "I-WORK_OF_ART", "18": "B-LAW", "19": "I-LAW", "20": "B-LANGUAGE", "21": "I-LANGUAGE", "22": "B-DATE", "23": "I-DATE", "24": "B-TIME", "25": "I-TIME", "26": "B-PERCENT", "27": "I-PERCENT", "28": "B-MONEY", "29": "I-MONEY", "30": "B-QUANTITY", "31": "I-QUANTITY", "32": "B-ORDINAL", "33": "I-ORDINAL", "34": "B-CARDINAL", "35": "I-CARDINAL", "36": "*", "37": "0", "38": "-"}}}}, {"name": "verbal_predicates", "sequence": "string"}, {"name": "coreference_clusters", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 380510, "num_examples": 396}, {"name": "test", "num_bytes": 402893, "num_examples": 396}], "download_size": 65383, "dataset_size": 783403}, {"config_name": "type1_pro", "features": [{"name": "document_id", "dtype": "string"}, {"name": "part_number", "dtype": "string"}, {"name": "word_number", "sequence": "int32"}, {"name": "tokens", "sequence": "string"}, {"name": "pos_tags", "sequence": {"class_label": {"names": {"0": "\"", "1": "''", "2": "#", "3": "$", "4": "(", "5": ")", "6": ",", "7": ".", "8": ":", "9": "``", "10": "CC", "11": "CD", "12": "DT", "13": "EX", "14": "FW", "15": "IN", "16": "JJ", "17": "JJR", "18": "JJS", "19": "LS", "20": "MD", "21": "NN", "22": "NNP", "23": "NNPS", "24": "NNS", "25": "NN|SYM", "26": "PDT", "27": "POS", "28": "PRP", "29": "PRP$", "30": "RB", "31": "RBR", "32": "RBS", "33": "RP", "34": "SYM", "35": "TO", "36": "UH", "37": "VB", "38": "VBD", "39": "VBG", "40": "VBN", "41": "VBP", "42": "VBZ", "43": "WDT", "44": "WP", "45": "WP$", "46": "WRB", "47": "HYPH", "48": "XX", "49": "NFP", "50": "AFX", "51": "ADD", "52": "-LRB-", "53": "-RRB-", "54": "-"}}}}, {"name": "parse_bit", "sequence": "string"}, {"name": "predicate_lemma", "sequence": "string"}, {"name": "predicate_framenet_id", "sequence": "string"}, {"name": "word_sense", "sequence": "string"}, {"name": "speaker", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "B-PERSON", "1": "I-PERSON", "2": "B-NORP", "3": "I-NORP", "4": "B-FAC", "5": "I-FAC", "6": "B-ORG", "7": "I-ORG", "8": "B-GPE", "9": "I-GPE", "10": "B-LOC", "11": "I-LOC", "12": "B-PRODUCT", "13": "I-PRODUCT", "14": "B-EVENT", "15": "I-EVENT", "16": "B-WORK_OF_ART", "17": "I-WORK_OF_ART", "18": "B-LAW", "19": "I-LAW", "20": "B-LANGUAGE", "21": "I-LANGUAGE", "22": "B-DATE", "23": "I-DATE", "24": "B-TIME", "25": "I-TIME", "26": "B-PERCENT", "27": "I-PERCENT", "28": "B-MONEY", "29": "I-MONEY", "30": "B-QUANTITY", "31": "I-QUANTITY", "32": "B-ORDINAL", "33": "I-ORDINAL", "34": "B-CARDINAL", "35": "I-CARDINAL", "36": "*", "37": "0", "38": "-"}}}}, {"name": "verbal_predicates", "sequence": "string"}, {"name": "coreference_clusters", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 379044, "num_examples": 396}, {"name": "test", "num_bytes": 401705, "num_examples": 396}], "download_size": 65516, "dataset_size": 780749}, {"config_name": "type2_anti", "features": [{"name": "document_id", "dtype": "string"}, {"name": "part_number", "dtype": "string"}, {"name": "word_number", "sequence": "int32"}, {"name": "tokens", "sequence": "string"}, {"name": "pos_tags", "sequence": {"class_label": {"names": {"0": "\"", "1": "''", "2": "#", "3": "$", "4": "(", "5": ")", "6": ",", "7": ".", "8": ":", "9": "``", "10": "CC", "11": "CD", "12": "DT", "13": "EX", "14": "FW", "15": "IN", "16": "JJ", "17": "JJR", "18": "JJS", "19": "LS", "20": "MD", "21": "NN", "22": "NNP", "23": "NNPS", "24": "NNS", "25": "NN|SYM", "26": "PDT", "27": "POS", "28": "PRP", "29": "PRP$", "30": "RB", "31": "RBR", "32": "RBS", "33": "RP", "34": "SYM", "35": "TO", "36": "UH", "37": "VB", "38": "VBD", "39": "VBG", "40": "VBN", "41": "VBP", "42": "VBZ", "43": "WDT", "44": "WP", "45": "WP$", "46": "WRB", "47": "HYPH", "48": "XX", "49": "NFP", "50": "AFX", "51": "ADD", "52": "-LRB-", "53": "-RRB-", "54": "-"}}}}, {"name": "parse_bit", "sequence": "string"}, {"name": "predicate_lemma", "sequence": "string"}, {"name": "predicate_framenet_id", "sequence": "string"}, {"name": "word_sense", "sequence": "string"}, {"name": "speaker", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "B-PERSON", "1": "I-PERSON", "2": "B-NORP", "3": "I-NORP", "4": "B-FAC", "5": "I-FAC", "6": "B-ORG", "7": "I-ORG", "8": "B-GPE", "9": "I-GPE", "10": "B-LOC", "11": "I-LOC", "12": "B-PRODUCT", "13": "I-PRODUCT", "14": "B-EVENT", "15": "I-EVENT", "16": "B-WORK_OF_ART", "17": "I-WORK_OF_ART", "18": "B-LAW", "19": "I-LAW", "20": "B-LANGUAGE", "21": "I-LANGUAGE", "22": "B-DATE", "23": "I-DATE", "24": "B-TIME", "25": "I-TIME", "26": "B-PERCENT", "27": "I-PERCENT", "28": "B-MONEY", "29": "I-MONEY", "30": "B-QUANTITY", "31": "I-QUANTITY", "32": "B-ORDINAL", "33": "I-ORDINAL", "34": "B-CARDINAL", "35": "I-CARDINAL", "36": "*", "37": "0", "38": "-"}}}}, {"name": "verbal_predicates", "sequence": "string"}, {"name": "coreference_clusters", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 368421, "num_examples": 396}, {"name": "test", "num_bytes": 376926, "num_examples": 396}], "download_size": 62555, "dataset_size": 745347}, {"config_name": "type2_pro", "features": [{"name": "document_id", "dtype": "string"}, {"name": "part_number", "dtype": "string"}, {"name": "word_number", "sequence": "int32"}, {"name": "tokens", "sequence": "string"}, {"name": "pos_tags", "sequence": {"class_label": {"names": {"0": "\"", "1": "''", "2": "#", "3": "$", "4": "(", "5": ")", "6": ",", "7": ".", "8": ":", "9": "``", "10": "CC", "11": "CD", "12": "DT", "13": "EX", "14": "FW", "15": "IN", "16": "JJ", "17": "JJR", "18": "JJS", "19": "LS", "20": "MD", "21": "NN", "22": "NNP", "23": "NNPS", "24": "NNS", "25": "NN|SYM", "26": "PDT", "27": "POS", "28": "PRP", "29": "PRP$", "30": "RB", "31": "RBR", "32": "RBS", "33": "RP", "34": "SYM", "35": "TO", "36": "UH", "37": "VB", "38": "VBD", "39": "VBG", "40": "VBN", "41": "VBP", "42": "VBZ", "43": "WDT", "44": "WP", "45": "WP$", "46": "WRB", "47": "HYPH", "48": "XX", "49": "NFP", "50": "AFX", "51": "ADD", "52": "-LRB-", "53": "-RRB-", "54": "-"}}}}, {"name": "parse_bit", "sequence": "string"}, {"name": "predicate_lemma", "sequence": "string"}, {"name": "predicate_framenet_id", "sequence": "string"}, {"name": "word_sense", "sequence": "string"}, {"name": "speaker", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "B-PERSON", "1": "I-PERSON", "2": "B-NORP", "3": "I-NORP", "4": "B-FAC", "5": "I-FAC", "6": "B-ORG", "7": "I-ORG", "8": "B-GPE", "9": "I-GPE", "10": "B-LOC", "11": "I-LOC", "12": "B-PRODUCT", "13": "I-PRODUCT", "14": "B-EVENT", "15": "I-EVENT", "16": "B-WORK_OF_ART", "17": "I-WORK_OF_ART", "18": "B-LAW", "19": "I-LAW", "20": "B-LANGUAGE", "21": "I-LANGUAGE", "22": "B-DATE", "23": "I-DATE", "24": "B-TIME", "25": "I-TIME", "26": "B-PERCENT", "27": "I-PERCENT", "28": "B-MONEY", "29": "I-MONEY", "30": "B-QUANTITY", "31": "I-QUANTITY", "32": "B-ORDINAL", "33": "I-ORDINAL", "34": "B-CARDINAL", "35": "I-CARDINAL", "36": "*", "37": "0", "38": "-"}}}}, {"name": "verbal_predicates", "sequence": "string"}, {"name": "coreference_clusters", "sequence": "string"}], "splits": [{"name": "validation", "num_bytes": 366957, "num_examples": 396}, {"name": "test", "num_bytes": 375144, "num_examples": 396}], "download_size": 62483, "dataset_size": 742101}, {"config_name": "wino_bias", "features": [{"name": "document_id", "dtype": "string"}, {"name": "part_number", "dtype": "string"}, {"name": "word_number", "sequence": "int32"}, {"name": "tokens", "sequence": "string"}, {"name": "pos_tags", "sequence": {"class_label": {"names": {"0": "\"", "1": "''", "2": "#", "3": "$", "4": "(", "5": ")", "6": ",", "7": ".", "8": ":", "9": "``", "10": "CC", "11": "CD", "12": "DT", "13": "EX", "14": "FW", "15": "IN", "16": "JJ", "17": "JJR", "18": "JJS", "19": "LS", "20": "MD", "21": "NN", "22": "NNP", "23": "NNPS", "24": "NNS", "25": "NN|SYM", "26": "PDT", "27": "POS", "28": "PRP", "29": "PRP$", "30": "RB", "31": "RBR", "32": "RBS", "33": "RP", "34": "SYM", "35": "TO", "36": "UH", "37": "VB", "38": "VBD", "39": "VBG", "40": "VBN", "41": "VBP", "42": "VBZ", "43": "WDT", "44": "WP", "45": "WP$", "46": "WRB", "47": "HYPH", "48": "XX", "49": "NFP", "50": "AFX", "51": "ADD", "52": "-LRB-", "53": "-RRB-"}}}}, {"name": "parse_bit", "sequence": "string"}, {"name": "predicate_lemma", "sequence": "string"}, {"name": "predicate_framenet_id", "sequence": "string"}, {"name": "word_sense", "sequence": "string"}, {"name": "speaker", "sequence": "string"}, {"name": "ner_tags", "sequence": {"class_label": {"names": {"0": "B-PERSON", "1": "I-PERSON", "2": "B-NORP", "3": "I-NORP", "4": "B-FAC", "5": "I-FAC", "6": "B-ORG", "7": "I-ORG", "8": "B-GPE", "9": "I-GPE", "10": "B-LOC", "11": "I-LOC", "12": "B-PRODUCT", "13": "I-PRODUCT", "14": "B-EVENT", "15": "I-EVENT", "16": "B-WORK_OF_ART", "17": "I-WORK_OF_ART", "18": "B-LAW", "19": "I-LAW", "20": "B-LANGUAGE", "21": "I-LANGUAGE", "22": "B-DATE", "23": "I-DATE", "24": "B-TIME", "25": "I-TIME", "26": "B-PERCENT", "27": "I-PERCENT", "28": "B-MONEY", "29": "I-MONEY", "30": "B-QUANTITY", "31": "I-QUANTITY", "32": "B-ORDINAL", "33": "I-ORDINAL", "34": "B-CARDINAL", "35": "I-CARDINAL", "36": "*", "37": "0"}}}}, {"name": "verbal_predicates", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 173899234, "num_examples": 150335}], "download_size": 268725744, "dataset_size": 173899234}], "configs": [{"config_name": "type1_anti", "data_files": [{"split": "validation", "path": "type1_anti/validation-*"}, {"split": "test", "path": "type1_anti/test-*"}]}, {"config_name": "type1_pro", "data_files": [{"split": "validation", "path": "type1_pro/validation-*"}, {"split": "test", "path": "type1_pro/test-*"}]}, {"config_name": "type2_anti", "data_files": [{"split": "validation", "path": "type2_anti/validation-*"}, {"split": "test", "path": "type2_anti/test-*"}]}, {"config_name": "type2_pro", "data_files": [{"split": "validation", "path": "type2_pro/validation-*"}, {"split": "test", "path": "type2_pro/test-*"}]}]}
2024-01-04T16:50:33+00:00
[ "1804.06876" ]
[ "en" ]
TAGS #task_categories-token-classification #task_ids-coreference-resolution #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-mit #arxiv-1804.06876 #region-us
# Dataset Card for Wino_Bias dataset ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: WinoBias - Repository: - Paper: Arxiv - Leaderboard: - Point of Contact: ### Dataset Summary WinoBias, a Winograd-schema dataset for coreference resolution focused on gender bias. The corpus contains Winograd-schema style sentences with entities corresponding to people referred by their occupation (e.g. the nurse, the doctor, the carpenter). ### Supported Tasks and Leaderboards The underlying task is coreference resolution. ### Languages English ## Dataset Structure ### Data Instances The dataset has 4 subsets: 'type1_pro', 'type1_anti', 'type2_pro' and 'type2_anti'. The '*_pro' subsets contain sentences that reinforce gender stereotypes (e.g. mechanics are male, nurses are female), whereas the '*_anti' datasets contain "anti-stereotypical" sentences (e.g. mechanics are female, nurses are male). The 'type1' (*WB-Knowledge*) subsets contain sentences for which world knowledge is necessary to resolve the co-references, and 'type2' (*WB-Syntax*) subsets require only the syntactic information present in the sentence to resolve them. ### Data Fields - document_id = This is a variation on the document filename - part_number = Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. - word_num = This is the word index of the word in that sentence. - tokens = This is the token as segmented/tokenized in the Treebank. - pos_tags = This is the Penn Treebank style part of speech. When parse information is missing, all part of speeches except the one for which there is some sense or proposition annotation are marked with a XX tag. The verb is marked with just a VERB tag. - parse_bit = This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. When the parse information is missing, the first word of a sentence is tagged as "(TOP*" and the last word is tagged as "*)" and all intermediate words are tagged with a "*". - predicate_lemma = The predicate lemma is mentioned for the rows for which we have semantic role information or word sense information. All other rows are marked with a "-". - predicate_framenet_id = This is the PropBank frameset ID of the predicate in predicate_lemma. - word_sense = This is the word sense of the word in Column tokens. - speaker = This is the speaker or author name where available. - ner_tags = These columns identifies the spans representing various named entities. For documents which do not have named entity annotation, each line is represented with an "*". - verbal_predicates = There is one column each of predicate argument structure information for the predicate mentioned in predicate_lemma. If there are no predicates tagged in a sentence this is a single column with all rows marked with an "*". ### Data Splits Dev and Test Split available ## Dataset Creation ### Curation Rationale The WinoBias dataset was introduced in 2018 (see paper), with its original task being *coreference resolution*, which is a task that aims to identify mentions that refer to the same entity or person. ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? The dataset was created by researchers familiar with the WinoBias project, based on two prototypical templates provided by the authors, in which entities interact in plausible ways. ### Annotations #### Annotation process #### Who are the annotators? "Researchers familiar with the [WinoBias] project" ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases Recent work has shown that this dataset contains grammatical issues, incorrect or ambiguous labels, and stereotype conflation, among other limitations. ### Other Known Limitations ## Additional Information ### Dataset Curators Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez and Kai-Wei Chan ### Licensing Information MIT Licence @article{DBLP:journals/corr/abs-1804-06876, author = {Jieyu Zhao and Tianlu Wang and Mark Yatskar and Vicente Ordonez and Kai{-}Wei Chang}, title = {Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods}, journal = {CoRR}, volume = {abs/1804.06876}, year = {2018}, url = {URL archivePrefix = {arXiv}, eprint = {1804.06876}, timestamp = {Mon, 13 Aug 2018 16:47:01 +0200}, biburl = {URL bibsource = {dblp computer science bibliography, URL} } ### Contributions Thanks to @akshayb7 for adding this dataset. Updated by @JieyuZhao.
[ "# Dataset Card for Wino_Bias dataset", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: WinoBias\n- Repository:\n- Paper: Arxiv\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nWinoBias, a Winograd-schema dataset for coreference resolution focused on gender bias.\nThe corpus contains Winograd-schema style sentences with entities corresponding to people referred by their occupation (e.g. the nurse, the doctor, the carpenter).", "### Supported Tasks and Leaderboards\n\nThe underlying task is coreference resolution.", "### Languages\n\nEnglish", "## Dataset Structure", "### Data Instances\n\nThe dataset has 4 subsets: 'type1_pro', 'type1_anti', 'type2_pro' and 'type2_anti'.\n\nThe '*_pro' subsets contain sentences that reinforce gender stereotypes (e.g. mechanics are male, nurses are female), whereas the '*_anti' datasets contain \"anti-stereotypical\" sentences (e.g. mechanics are female, nurses are male).\n\nThe 'type1' (*WB-Knowledge*) subsets contain sentences for which world knowledge is necessary to resolve the co-references, and 'type2' (*WB-Syntax*) subsets require only the syntactic information present in the sentence to resolve them.", "### Data Fields\n\n - document_id = This is a variation on the document filename\n - part_number = Some files are divided into multiple parts numbered as 000, 001, 002, ... etc.\n - word_num = This is the word index of the word in that sentence.\n - tokens = This is the token as segmented/tokenized in the Treebank.\n - pos_tags = This is the Penn Treebank style part of speech. When parse information is missing, all part of speeches except the one for which there is some sense or proposition annotation are marked with a XX tag. The verb is marked with just a VERB tag.\n - parse_bit = This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the \"([pos] [word])\" string (or leaf) and concatenating the items in the rows of that column. When the parse information is missing, the first word of a sentence is tagged as \"(TOP*\" and the last word is tagged as \"*)\" and all intermediate words are tagged with a \"*\".\n - predicate_lemma = The predicate lemma is mentioned for the rows for which we have semantic role information or word sense information. All other rows are marked with a \"-\".\n - predicate_framenet_id = This is the PropBank frameset ID of the predicate in predicate_lemma.\n - word_sense = This is the word sense of the word in Column tokens.\n - speaker = This is the speaker or author name where available.\n - ner_tags = These columns identifies the spans representing various named entities. For documents which do not have named entity annotation, each line is represented with an \"*\".\n - verbal_predicates = There is one column each of predicate argument structure information for the predicate mentioned in predicate_lemma. If there are no predicates tagged in a sentence this is a single column with all rows marked with an \"*\".", "### Data Splits\n\nDev and Test Split available", "## Dataset Creation", "### Curation Rationale\n\nThe WinoBias dataset was introduced in 2018 (see paper), with its original task being *coreference resolution*, which is a task that aims to identify mentions that refer to the same entity or person.", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?\n\n The dataset was created by researchers familiar with the WinoBias project, based on two prototypical templates provided by the authors, in which entities interact in plausible ways.", "### Annotations", "#### Annotation process", "#### Who are the annotators?\n\n\"Researchers familiar with the [WinoBias] project\"", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases\n\nRecent work has shown that this dataset contains grammatical issues, incorrect or ambiguous labels, and stereotype conflation, among other limitations.", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nJieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez and Kai-Wei Chan", "### Licensing Information\n\nMIT Licence\n\n\n\n@article{DBLP:journals/corr/abs-1804-06876,\n author = {Jieyu Zhao and\n Tianlu Wang and\n Mark Yatskar and\n Vicente Ordonez and\n Kai{-}Wei Chang},\n title = {Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods},\n journal = {CoRR},\n volume = {abs/1804.06876},\n year = {2018},\n url = {URL\n archivePrefix = {arXiv},\n eprint = {1804.06876},\n timestamp = {Mon, 13 Aug 2018 16:47:01 +0200},\n biburl = {URL\n bibsource = {dblp computer science bibliography, URL}\n}", "### Contributions\n\nThanks to @akshayb7 for adding this dataset. Updated by @JieyuZhao." ]
[ "TAGS\n#task_categories-token-classification #task_ids-coreference-resolution #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-mit #arxiv-1804.06876 #region-us \n", "# Dataset Card for Wino_Bias dataset", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: WinoBias\n- Repository:\n- Paper: Arxiv\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nWinoBias, a Winograd-schema dataset for coreference resolution focused on gender bias.\nThe corpus contains Winograd-schema style sentences with entities corresponding to people referred by their occupation (e.g. the nurse, the doctor, the carpenter).", "### Supported Tasks and Leaderboards\n\nThe underlying task is coreference resolution.", "### Languages\n\nEnglish", "## Dataset Structure", "### Data Instances\n\nThe dataset has 4 subsets: 'type1_pro', 'type1_anti', 'type2_pro' and 'type2_anti'.\n\nThe '*_pro' subsets contain sentences that reinforce gender stereotypes (e.g. mechanics are male, nurses are female), whereas the '*_anti' datasets contain \"anti-stereotypical\" sentences (e.g. mechanics are female, nurses are male).\n\nThe 'type1' (*WB-Knowledge*) subsets contain sentences for which world knowledge is necessary to resolve the co-references, and 'type2' (*WB-Syntax*) subsets require only the syntactic information present in the sentence to resolve them.", "### Data Fields\n\n - document_id = This is a variation on the document filename\n - part_number = Some files are divided into multiple parts numbered as 000, 001, 002, ... etc.\n - word_num = This is the word index of the word in that sentence.\n - tokens = This is the token as segmented/tokenized in the Treebank.\n - pos_tags = This is the Penn Treebank style part of speech. When parse information is missing, all part of speeches except the one for which there is some sense or proposition annotation are marked with a XX tag. The verb is marked with just a VERB tag.\n - parse_bit = This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the \"([pos] [word])\" string (or leaf) and concatenating the items in the rows of that column. When the parse information is missing, the first word of a sentence is tagged as \"(TOP*\" and the last word is tagged as \"*)\" and all intermediate words are tagged with a \"*\".\n - predicate_lemma = The predicate lemma is mentioned for the rows for which we have semantic role information or word sense information. All other rows are marked with a \"-\".\n - predicate_framenet_id = This is the PropBank frameset ID of the predicate in predicate_lemma.\n - word_sense = This is the word sense of the word in Column tokens.\n - speaker = This is the speaker or author name where available.\n - ner_tags = These columns identifies the spans representing various named entities. For documents which do not have named entity annotation, each line is represented with an \"*\".\n - verbal_predicates = There is one column each of predicate argument structure information for the predicate mentioned in predicate_lemma. If there are no predicates tagged in a sentence this is a single column with all rows marked with an \"*\".", "### Data Splits\n\nDev and Test Split available", "## Dataset Creation", "### Curation Rationale\n\nThe WinoBias dataset was introduced in 2018 (see paper), with its original task being *coreference resolution*, which is a task that aims to identify mentions that refer to the same entity or person.", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?\n\n The dataset was created by researchers familiar with the WinoBias project, based on two prototypical templates provided by the authors, in which entities interact in plausible ways.", "### Annotations", "#### Annotation process", "#### Who are the annotators?\n\n\"Researchers familiar with the [WinoBias] project\"", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases\n\nRecent work has shown that this dataset contains grammatical issues, incorrect or ambiguous labels, and stereotype conflation, among other limitations.", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nJieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez and Kai-Wei Chan", "### Licensing Information\n\nMIT Licence\n\n\n\n@article{DBLP:journals/corr/abs-1804-06876,\n author = {Jieyu Zhao and\n Tianlu Wang and\n Mark Yatskar and\n Vicente Ordonez and\n Kai{-}Wei Chang},\n title = {Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods},\n journal = {CoRR},\n volume = {abs/1804.06876},\n year = {2018},\n url = {URL\n archivePrefix = {arXiv},\n eprint = {1804.06876},\n timestamp = {Mon, 13 Aug 2018 16:47:01 +0200},\n biburl = {URL\n bibsource = {dblp computer science bibliography, URL}\n}", "### Contributions\n\nThanks to @akshayb7 for adding this dataset. Updated by @JieyuZhao." ]
[ 99, 12, 120, 31, 72, 20, 5, 6, 175, 500, 10, 5, 54, 4, 10, 50, 5, 5, 24, 8, 8, 7, 43, 7, 5, 31, 174, 29 ]
[ "passage: TAGS\n#task_categories-token-classification #task_ids-coreference-resolution #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-1K<n<10K #source_datasets-original #language-English #license-mit #arxiv-1804.06876 #region-us \n# Dataset Card for Wino_Bias dataset## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: WinoBias\n- Repository:\n- Paper: Arxiv\n- Leaderboard:\n- Point of Contact:### Dataset Summary\n\nWinoBias, a Winograd-schema dataset for coreference resolution focused on gender bias.\nThe corpus contains Winograd-schema style sentences with entities corresponding to people referred by their occupation (e.g. the nurse, the doctor, the carpenter).### Supported Tasks and Leaderboards\n\nThe underlying task is coreference resolution.### Languages\n\nEnglish## Dataset Structure", "passage: ### Data Instances\n\nThe dataset has 4 subsets: 'type1_pro', 'type1_anti', 'type2_pro' and 'type2_anti'.\n\nThe '*_pro' subsets contain sentences that reinforce gender stereotypes (e.g. mechanics are male, nurses are female), whereas the '*_anti' datasets contain \"anti-stereotypical\" sentences (e.g. mechanics are female, nurses are male).\n\nThe 'type1' (*WB-Knowledge*) subsets contain sentences for which world knowledge is necessary to resolve the co-references, and 'type2' (*WB-Syntax*) subsets require only the syntactic information present in the sentence to resolve them.### Data Fields\n\n - document_id = This is a variation on the document filename\n - part_number = Some files are divided into multiple parts numbered as 000, 001, 002, ... etc.\n - word_num = This is the word index of the word in that sentence.\n - tokens = This is the token as segmented/tokenized in the Treebank.\n - pos_tags = This is the Penn Treebank style part of speech. When parse information is missing, all part of speeches except the one for which there is some sense or proposition annotation are marked with a XX tag. The verb is marked with just a VERB tag.\n - parse_bit = This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the \"([pos] [word])\" string (or leaf) and concatenating the items in the rows of that column. When the parse information is missing, the first word of a sentence is tagged as \"(TOP*\" and the last word is tagged as \"*)\" and all intermediate words are tagged with a \"*\".\n - predicate_lemma = The predicate lemma is mentioned for the rows for which we have semantic role information or word sense information. All other rows are marked with a \"-\".\n - predicate_framenet_id = This is the PropBank frameset ID of the predicate in predicate_lemma.\n - word_sense = This is the word sense of the word in Column tokens.\n - speaker = This is the speaker or author name where available.\n - ner_tags = These columns identifies the spans representing various named entities. For documents which do not have named entity annotation, each line is represented with an \"*\".\n - verbal_predicates = There is one column each of predicate argument structure information for the predicate mentioned in predicate_lemma. If there are no predicates tagged in a sentence this is a single column with all rows marked with an \"*\"." ]
8052c534a4cecd8328c0bc879f6a5aa07404c4da
# Dataset Card for The Winograd Schema Challenge ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://cs.nyu.edu/faculty/davise/papers/WinogradSchemas/WS.html - **Repository:** - **Paper:** https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.729.9814&rep=rep1&type=pdf - **Leaderboard:** - **Point of Contact:** ### Dataset Summary A Winograd schema is a pair of sentences that differ in only one or two words and that contain an ambiguity that is resolved in opposite ways in the two sentences and requires the use of world knowledge and reasoning for its resolution. The schema takes its name from a well-known example by Terry Winograd: > The city councilmen refused the demonstrators a permit because they [feared/advocated] violence. If the word is ``feared'', then ``they'' presumably refers to the city council; if it is ``advocated'' then ``they'' presumably refers to the demonstrators. ### Supported Tasks and Leaderboards From the official webpage: > A contest, entitled the Winograd Schema Challenge was run once, in 2016. At that time, there was a cash prize offered for achieving human-level performance in the contest. Since then, the sponsor has withdrawn; therefore NO CASH PRIZES CAN BE OFFERED OR WILL BE AWARDED FOR ANY KIND OF PERFORMANCE OR ACHIEVEMENT ON THIS CHALLENGE. ### Languages The dataset is in English. [Translation of 12 WSs into Chinese ](https://cs.nyu.edu/faculty/davise/papers/WinogradSchemas/WSChinese.html)(translated by Wei Xu). Translations into Japanese, by Soichiro Tanaka, Rafal Rzepka, and Shiho Katajima\ **Translation changing English names to Japanese **[PDF ](https://cs.nyu.edu/faculty/davise/papers/WinogradSchemas/collection_ja.pdf)    [HTML](http://arakilab.media.eng.hokudai.ac.jp/~kabura/collection_ja.html)\ **Translation preserving English names** [PDF ](https://cs.nyu.edu/faculty/davise/papers/WinogradSchemas/collection_katakana.pdf)    [HTML](http://arakilab.media.eng.hokudai.ac.jp/~kabura/collection_katakana.html) [Translation into French, ](http://www.llf.cnrs.fr/winograd-fr)by Pascal Amsili and Olga Seminck [Winograd Schemas in Portuguese](https://sol.sbc.org.br/index.php/eniac/article/view/9334) by Gabriela Melo, Vinicius Imaizumi, and Fábio Cozman. [Mandarinograd: A Chinese Collection of Winograd Schemas](https://www.aclweb.org/anthology/2020.lrec-1.3) by Timothée Bernard and Ting Han, LREC-2020. ## Dataset Structure ### Data Instances Each instance contains a text passage with a designated pronoun and two possible answers indicating which entity in the passage the pronoun represents. An example instance looks like the following: ```python { 'label': 0, 'options': ['The city councilmen', 'The demonstrators'], 'pronoun': 'they', 'pronoun_loc': 63, 'quote': 'they feared violence', 'quote_loc': 63, 'source': '(Winograd 1972)', 'text': 'The city councilmen refused the demonstrators a permit because they feared violence.' } ``` ### Data Fields - `text` (str): The text sequence - `options` (list[str]): The two entity options that the pronoun may be referring to - `label` (int): The index of the correct option in the `options` field - `pronoun` (str): The pronoun in the sequence to be resolved - `pronoun_loc` (int): The starting position of the pronoun in the sequence - `quote` (str): The substr with the key action or context surrounding the pronoun - `quote_loc` (int): The starting position of the quote in the sequence - `source` (str): A description of the source who contributed the example ### Data Splits Only a test split is included. ## Dataset Creation ### Curation Rationale The Winograd Schema Challenge was proposed as an automated evaluation of an AI system's commonsense linguistic understanding. From the webpage: > The strengths of the challenge are that it is clear-cut, in that the answer to each schema is a binary choice; vivid, in that it is obvious to non-experts that a program that fails to get the right answers clearly has serious gaps in its understanding; and difficult, in that it is far beyond the current state of the art. ### Source Data #### Initial Data Collection and Normalization This data was manually written by experts such that the schemas are: - easily disambiguated by the human reader (ideally, so easily that the reader does not even notice that there is an ambiguity); - not solvable by simple techniques such as selectional restrictions; - Google-proof; that is, there is no obvious statistical test over text corpora that will reliably disambiguate these correctly. #### Who are the source language producers? This dataset has grown over time, and so was produced by a variety of lingustic and AI researchers. See the `source` field for the source of each instance. ### Annotations #### Annotation process Annotations are produced by the experts who construct the examples. #### Who are the annotators? See above. ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators This dataset has grown over time, and so was produced by a variety of lingustic and AI researchers. See the `source` field for the source of each instance. ### Licensing Information This work is licensed under a [Creative Commons Attribution 4.0 International License](https://creativecommons.org/licenses/by/4.0/). ### Citation Information The Winograd Schema Challenge including many of the examples here was proposed by [Levesque et al 2012](https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.729.9814&rep=rep1&type=pdf): ``` @inproceedings{levesque2012winograd, title={The winograd schema challenge}, author={Levesque, Hector and Davis, Ernest and Morgenstern, Leora}, booktitle={Thirteenth International Conference on the Principles of Knowledge Representation and Reasoning}, year={2012}, organization={Citeseer} } ``` ### Contributions Thanks to [@joeddav](https://github.com/joeddav) for adding this dataset.
winograd_wsc
[ "task_categories:multiple-choice", "task_ids:multiple-choice-coreference-resolution", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:n<1K", "source_datasets:original", "language:en", "license:cc-by-4.0", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["expert-generated"], "language": ["en"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["n<1K"], "source_datasets": ["original"], "task_categories": ["multiple-choice"], "task_ids": ["multiple-choice-coreference-resolution"], "paperswithcode_id": "wsc", "pretty_name": "Winograd Schema Challenge", "dataset_info": [{"config_name": "wsc285", "features": [{"name": "text", "dtype": "string"}, {"name": "pronoun", "dtype": "string"}, {"name": "pronoun_loc", "dtype": "int32"}, {"name": "quote", "dtype": "string"}, {"name": "quote_loc", "dtype": "int32"}, {"name": "options", "sequence": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "0", "1": "1"}}}}, {"name": "source", "dtype": "string"}], "splits": [{"name": "test", "num_bytes": 52281, "num_examples": 285}], "download_size": 113235, "dataset_size": 52281}, {"config_name": "wsc273", "features": [{"name": "text", "dtype": "string"}, {"name": "pronoun", "dtype": "string"}, {"name": "pronoun_loc", "dtype": "int32"}, {"name": "quote", "dtype": "string"}, {"name": "quote_loc", "dtype": "int32"}, {"name": "options", "sequence": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "0", "1": "1"}}}}, {"name": "source", "dtype": "string"}], "splits": [{"name": "test", "num_bytes": 49674, "num_examples": 273}], "download_size": 113235, "dataset_size": 49674}]}
2024-01-18T11:18:21+00:00
[]
[ "en" ]
TAGS #task_categories-multiple-choice #task_ids-multiple-choice-coreference-resolution #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-n<1K #source_datasets-original #language-English #license-cc-by-4.0 #region-us
# Dataset Card for The Winograd Schema Challenge ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: - Paper: URL - Leaderboard: - Point of Contact: ### Dataset Summary A Winograd schema is a pair of sentences that differ in only one or two words and that contain an ambiguity that is resolved in opposite ways in the two sentences and requires the use of world knowledge and reasoning for its resolution. The schema takes its name from a well-known example by Terry Winograd: > The city councilmen refused the demonstrators a permit because they [feared/advocated] violence. If the word is ''feared'', then ''they'' presumably refers to the city council; if it is ''advocated'' then ''they'' presumably refers to the demonstrators. ### Supported Tasks and Leaderboards From the official webpage: > A contest, entitled the Winograd Schema Challenge was run once, in 2016. At that time, there was a cash prize offered for achieving human-level performance in the contest. Since then, the sponsor has withdrawn; therefore NO CASH PRIZES CAN BE OFFERED OR WILL BE AWARDED FOR ANY KIND OF PERFORMANCE OR ACHIEVEMENT ON THIS CHALLENGE. ### Languages The dataset is in English. Translation of 12 WSs into Chinese (translated by Wei Xu). Translations into Japanese, by Soichiro Tanaka, Rafal Rzepka, and Shiho Katajima\ Translation changing English names to Japanese PDF     HTML\ Translation preserving English names PDF     HTML Translation into French, by Pascal Amsili and Olga Seminck Winograd Schemas in Portuguese by Gabriela Melo, Vinicius Imaizumi, and Fábio Cozman. Mandarinograd: A Chinese Collection of Winograd Schemas by Timothée Bernard and Ting Han, LREC-2020. ## Dataset Structure ### Data Instances Each instance contains a text passage with a designated pronoun and two possible answers indicating which entity in the passage the pronoun represents. An example instance looks like the following: ### Data Fields - 'text' (str): The text sequence - 'options' (list[str]): The two entity options that the pronoun may be referring to - 'label' (int): The index of the correct option in the 'options' field - 'pronoun' (str): The pronoun in the sequence to be resolved - 'pronoun_loc' (int): The starting position of the pronoun in the sequence - 'quote' (str): The substr with the key action or context surrounding the pronoun - 'quote_loc' (int): The starting position of the quote in the sequence - 'source' (str): A description of the source who contributed the example ### Data Splits Only a test split is included. ## Dataset Creation ### Curation Rationale The Winograd Schema Challenge was proposed as an automated evaluation of an AI system's commonsense linguistic understanding. From the webpage: > The strengths of the challenge are that it is clear-cut, in that the answer to each schema is a binary choice; vivid, in that it is obvious to non-experts that a program that fails to get the right answers clearly has serious gaps in its understanding; and difficult, in that it is far beyond the current state of the art. ### Source Data #### Initial Data Collection and Normalization This data was manually written by experts such that the schemas are: - easily disambiguated by the human reader (ideally, so easily that the reader does not even notice that there is an ambiguity); - not solvable by simple techniques such as selectional restrictions; - Google-proof; that is, there is no obvious statistical test over text corpora that will reliably disambiguate these correctly. #### Who are the source language producers? This dataset has grown over time, and so was produced by a variety of lingustic and AI researchers. See the 'source' field for the source of each instance. ### Annotations #### Annotation process Annotations are produced by the experts who construct the examples. #### Who are the annotators? See above. ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators This dataset has grown over time, and so was produced by a variety of lingustic and AI researchers. See the 'source' field for the source of each instance. ### Licensing Information This work is licensed under a Creative Commons Attribution 4.0 International License. The Winograd Schema Challenge including many of the examples here was proposed by Levesque et al 2012: ### Contributions Thanks to @joeddav for adding this dataset.
[ "# Dataset Card for The Winograd Schema Challenge", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: URL\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nA Winograd schema is a pair of sentences that differ in only one or two words and that contain an ambiguity that is\nresolved in opposite ways in the two sentences and requires the use of world knowledge and reasoning for its\nresolution. The schema takes its name from a well-known example by Terry Winograd:\n\n> The city councilmen refused the demonstrators a permit because they [feared/advocated] violence.\n\nIf the word is ''feared'', then ''they'' presumably refers to the city council; if it is ''advocated'' then ''they''\npresumably refers to the demonstrators.", "### Supported Tasks and Leaderboards\n\nFrom the official webpage:\n\n> A contest, entitled the Winograd Schema Challenge was run once, in 2016. At that time, there was a cash prize\noffered for achieving human-level performance in the contest. Since then, the sponsor has withdrawn; therefore NO\nCASH PRIZES CAN BE OFFERED OR WILL BE AWARDED FOR ANY KIND OF PERFORMANCE OR ACHIEVEMENT ON THIS CHALLENGE.", "### Languages\n\nThe dataset is in English.\n\nTranslation of 12 WSs into Chinese (translated by Wei Xu).\n\nTranslations into Japanese, by Soichiro Tanaka, Rafal Rzepka, and Shiho Katajima\\\nTranslation changing English names to Japanese PDF     HTML\\\nTranslation preserving English names PDF     HTML\n\nTranslation into French, by Pascal Amsili and Olga Seminck\n\nWinograd Schemas in Portuguese by Gabriela Melo, Vinicius Imaizumi, and Fábio Cozman.\n\nMandarinograd: A Chinese Collection of Winograd Schemas by Timothée Bernard and Ting Han, LREC-2020.", "## Dataset Structure", "### Data Instances\n\nEach instance contains a text passage with a designated pronoun and two possible answers indicating which entity in\nthe passage the pronoun represents. An example instance looks like the following:", "### Data Fields\n\n- 'text' (str): The text sequence\n- 'options' (list[str]): The two entity options that the pronoun may be referring to\n- 'label' (int): The index of the correct option in the 'options' field\n- 'pronoun' (str): The pronoun in the sequence to be resolved\n- 'pronoun_loc' (int): The starting position of the pronoun in the sequence\n- 'quote' (str): The substr with the key action or context surrounding the pronoun\n- 'quote_loc' (int): The starting position of the quote in the sequence\n- 'source' (str): A description of the source who contributed the example", "### Data Splits\n\nOnly a test split is included.", "## Dataset Creation", "### Curation Rationale\n\nThe Winograd Schema Challenge was proposed as an automated evaluation of an AI system's commonsense linguistic\nunderstanding. From the webpage:\n\n> The strengths of the challenge are that it is clear-cut, in that the answer to each schema is a binary choice;\nvivid, in that it is obvious to non-experts that a program that fails to get the right answers clearly has serious\ngaps in its understanding; and difficult, in that it is far beyond the current state of the art.", "### Source Data", "#### Initial Data Collection and Normalization\n\nThis data was manually written by experts such that the schemas are:\n\n- easily disambiguated by the human reader (ideally, so easily that the reader does not even notice that there is an ambiguity);\n\n- not solvable by simple techniques such as selectional restrictions;\n\n- Google-proof; that is, there is no obvious statistical test over text corpora that will reliably disambiguate these correctly.", "#### Who are the source language producers?\n\nThis dataset has grown over time, and so was produced by a variety of lingustic and AI researchers. See the 'source'\nfield for the source of each instance.", "### Annotations", "#### Annotation process\n\nAnnotations are produced by the experts who construct the examples.", "#### Who are the annotators?\n\nSee above.", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThis dataset has grown over time, and so was produced by a variety of lingustic and AI researchers. See the 'source'\nfield for the source of each instance.", "### Licensing Information\n\nThis work is licensed under a Creative Commons Attribution 4.0 International\nLicense.\n\n\n\nThe Winograd Schema Challenge including many of the examples here was proposed by\nLevesque et al 2012:", "### Contributions\n\nThanks to @joeddav for adding this dataset." ]
[ "TAGS\n#task_categories-multiple-choice #task_ids-multiple-choice-coreference-resolution #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-n<1K #source_datasets-original #language-English #license-cc-by-4.0 #region-us \n", "# Dataset Card for The Winograd Schema Challenge", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: URL\n- Leaderboard:\n- Point of Contact:", "### Dataset Summary\n\nA Winograd schema is a pair of sentences that differ in only one or two words and that contain an ambiguity that is\nresolved in opposite ways in the two sentences and requires the use of world knowledge and reasoning for its\nresolution. The schema takes its name from a well-known example by Terry Winograd:\n\n> The city councilmen refused the demonstrators a permit because they [feared/advocated] violence.\n\nIf the word is ''feared'', then ''they'' presumably refers to the city council; if it is ''advocated'' then ''they''\npresumably refers to the demonstrators.", "### Supported Tasks and Leaderboards\n\nFrom the official webpage:\n\n> A contest, entitled the Winograd Schema Challenge was run once, in 2016. At that time, there was a cash prize\noffered for achieving human-level performance in the contest. Since then, the sponsor has withdrawn; therefore NO\nCASH PRIZES CAN BE OFFERED OR WILL BE AWARDED FOR ANY KIND OF PERFORMANCE OR ACHIEVEMENT ON THIS CHALLENGE.", "### Languages\n\nThe dataset is in English.\n\nTranslation of 12 WSs into Chinese (translated by Wei Xu).\n\nTranslations into Japanese, by Soichiro Tanaka, Rafal Rzepka, and Shiho Katajima\\\nTranslation changing English names to Japanese PDF     HTML\\\nTranslation preserving English names PDF     HTML\n\nTranslation into French, by Pascal Amsili and Olga Seminck\n\nWinograd Schemas in Portuguese by Gabriela Melo, Vinicius Imaizumi, and Fábio Cozman.\n\nMandarinograd: A Chinese Collection of Winograd Schemas by Timothée Bernard and Ting Han, LREC-2020.", "## Dataset Structure", "### Data Instances\n\nEach instance contains a text passage with a designated pronoun and two possible answers indicating which entity in\nthe passage the pronoun represents. An example instance looks like the following:", "### Data Fields\n\n- 'text' (str): The text sequence\n- 'options' (list[str]): The two entity options that the pronoun may be referring to\n- 'label' (int): The index of the correct option in the 'options' field\n- 'pronoun' (str): The pronoun in the sequence to be resolved\n- 'pronoun_loc' (int): The starting position of the pronoun in the sequence\n- 'quote' (str): The substr with the key action or context surrounding the pronoun\n- 'quote_loc' (int): The starting position of the quote in the sequence\n- 'source' (str): A description of the source who contributed the example", "### Data Splits\n\nOnly a test split is included.", "## Dataset Creation", "### Curation Rationale\n\nThe Winograd Schema Challenge was proposed as an automated evaluation of an AI system's commonsense linguistic\nunderstanding. From the webpage:\n\n> The strengths of the challenge are that it is clear-cut, in that the answer to each schema is a binary choice;\nvivid, in that it is obvious to non-experts that a program that fails to get the right answers clearly has serious\ngaps in its understanding; and difficult, in that it is far beyond the current state of the art.", "### Source Data", "#### Initial Data Collection and Normalization\n\nThis data was manually written by experts such that the schemas are:\n\n- easily disambiguated by the human reader (ideally, so easily that the reader does not even notice that there is an ambiguity);\n\n- not solvable by simple techniques such as selectional restrictions;\n\n- Google-proof; that is, there is no obvious statistical test over text corpora that will reliably disambiguate these correctly.", "#### Who are the source language producers?\n\nThis dataset has grown over time, and so was produced by a variety of lingustic and AI researchers. See the 'source'\nfield for the source of each instance.", "### Annotations", "#### Annotation process\n\nAnnotations are produced by the experts who construct the examples.", "#### Who are the annotators?\n\nSee above.", "### Personal and Sensitive Information", "## Considerations for Using the Data", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThis dataset has grown over time, and so was produced by a variety of lingustic and AI researchers. See the 'source'\nfield for the source of each instance.", "### Licensing Information\n\nThis work is licensed under a Creative Commons Attribution 4.0 International\nLicense.\n\n\n\nThe Winograd Schema Challenge including many of the examples here was proposed by\nLevesque et al 2012:", "### Contributions\n\nThanks to @joeddav for adding this dataset." ]
[ 98, 12, 120, 26, 147, 107, 136, 6, 46, 163, 12, 5, 117, 4, 101, 48, 5, 19, 12, 8, 8, 7, 8, 7, 5, 44, 44, 17 ]
[ "passage: TAGS\n#task_categories-multiple-choice #task_ids-multiple-choice-coreference-resolution #annotations_creators-expert-generated #language_creators-expert-generated #multilinguality-monolingual #size_categories-n<1K #source_datasets-original #language-English #license-cc-by-4.0 #region-us \n# Dataset Card for The Winograd Schema Challenge## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: \n- Paper: URL\n- Leaderboard:\n- Point of Contact:### Dataset Summary\n\nA Winograd schema is a pair of sentences that differ in only one or two words and that contain an ambiguity that is\nresolved in opposite ways in the two sentences and requires the use of world knowledge and reasoning for its\nresolution. The schema takes its name from a well-known example by Terry Winograd:\n\n> The city councilmen refused the demonstrators a permit because they [feared/advocated] violence.\n\nIf the word is ''feared'', then ''they'' presumably refers to the city council; if it is ''advocated'' then ''they''\npresumably refers to the demonstrators.", "passage: ### Supported Tasks and Leaderboards\n\nFrom the official webpage:\n\n> A contest, entitled the Winograd Schema Challenge was run once, in 2016. At that time, there was a cash prize\noffered for achieving human-level performance in the contest. Since then, the sponsor has withdrawn; therefore NO\nCASH PRIZES CAN BE OFFERED OR WILL BE AWARDED FOR ANY KIND OF PERFORMANCE OR ACHIEVEMENT ON THIS CHALLENGE.### Languages\n\nThe dataset is in English.\n\nTranslation of 12 WSs into Chinese (translated by Wei Xu).\n\nTranslations into Japanese, by Soichiro Tanaka, Rafal Rzepka, and Shiho Katajima\\\nTranslation changing English names to Japanese PDF     HTML\\\nTranslation preserving English names PDF     HTML\n\nTranslation into French, by Pascal Amsili and Olga Seminck\n\nWinograd Schemas in Portuguese by Gabriela Melo, Vinicius Imaizumi, and Fábio Cozman.\n\nMandarinograd: A Chinese Collection of Winograd Schemas by Timothée Bernard and Ting Han, LREC-2020.## Dataset Structure### Data Instances\n\nEach instance contains a text passage with a designated pronoun and two possible answers indicating which entity in\nthe passage the pronoun represents. An example instance looks like the following:### Data Fields\n\n- 'text' (str): The text sequence\n- 'options' (list[str]): The two entity options that the pronoun may be referring to\n- 'label' (int): The index of the correct option in the 'options' field\n- 'pronoun' (str): The pronoun in the sequence to be resolved\n- 'pronoun_loc' (int): The starting position of the pronoun in the sequence\n- 'quote' (str): The substr with the key action or context surrounding the pronoun\n- 'quote_loc' (int): The starting position of the quote in the sequence\n- 'source' (str): A description of the source who contributed the example### Data Splits\n\nOnly a test split is included.## Dataset Creation### Curation Rationale\n\nThe Winograd Schema Challenge was proposed as an automated evaluation of an AI system's commonsense linguistic\nunderstanding. From the webpage:\n\n> The strengths of the challenge are that it is clear-cut, in that the answer to each schema is a binary choice;\nvivid, in that it is obvious to non-experts that a program that fails to get the right answers clearly has serious\ngaps in its understanding; and difficult, in that it is far beyond the current state of the art.### Source Data" ]
85ac5b5a3b7a930e22d590176e39460400d19e41
# Dataset Card for "winogrande" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://leaderboard.allenai.org/winogrande/submissions/get-started](https://leaderboard.allenai.org/winogrande/submissions/get-started) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 20.37 MB - **Size of the generated dataset:** 10.50 MB - **Total amount of disk used:** 30.87 MB ### Dataset Summary WinoGrande is a new collection of 44k problems, inspired by Winograd Schema Challenge (Levesque, Davis, and Morgenstern 2011), but adjusted to improve the scale and robustness against the dataset-specific bias. Formulated as a fill-in-a-blank task with binary options, the goal is to choose the right option for a given sentence which requires commonsense reasoning. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### winogrande_debiased - **Size of downloaded dataset files:** 3.40 MB - **Size of the generated dataset:** 1.59 MB - **Total amount of disk used:** 4.99 MB An example of 'train' looks as follows. ``` ``` #### winogrande_l - **Size of downloaded dataset files:** 3.40 MB - **Size of the generated dataset:** 1.71 MB - **Total amount of disk used:** 5.11 MB An example of 'validation' looks as follows. ``` ``` #### winogrande_m - **Size of downloaded dataset files:** 3.40 MB - **Size of the generated dataset:** 0.72 MB - **Total amount of disk used:** 4.12 MB An example of 'validation' looks as follows. ``` ``` #### winogrande_s - **Size of downloaded dataset files:** 3.40 MB - **Size of the generated dataset:** 0.47 MB - **Total amount of disk used:** 3.87 MB An example of 'validation' looks as follows. ``` ``` #### winogrande_xl - **Size of downloaded dataset files:** 3.40 MB - **Size of the generated dataset:** 5.58 MB - **Total amount of disk used:** 8.98 MB An example of 'train' looks as follows. ``` ``` ### Data Fields The data fields are the same among all splits. #### winogrande_debiased - `sentence`: a `string` feature. - `option1`: a `string` feature. - `option2`: a `string` feature. - `answer`: a `string` feature. #### winogrande_l - `sentence`: a `string` feature. - `option1`: a `string` feature. - `option2`: a `string` feature. - `answer`: a `string` feature. #### winogrande_m - `sentence`: a `string` feature. - `option1`: a `string` feature. - `option2`: a `string` feature. - `answer`: a `string` feature. #### winogrande_s - `sentence`: a `string` feature. - `option1`: a `string` feature. - `option2`: a `string` feature. - `answer`: a `string` feature. #### winogrande_xl - `sentence`: a `string` feature. - `option1`: a `string` feature. - `option2`: a `string` feature. - `answer`: a `string` feature. ### Data Splits | name |train|validation|test| |-------------------|----:|---------:|---:| |winogrande_debiased| 9248| 1267|1767| |winogrande_l |10234| 1267|1767| |winogrande_m | 2558| 1267|1767| |winogrande_s | 640| 1267|1767| |winogrande_xl |40398| 1267|1767| |winogrande_xs | 160| 1267|1767| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @InProceedings{ai2:winogrande, title = {WinoGrande: An Adversarial Winograd Schema Challenge at Scale}, authors={Keisuke, Sakaguchi and Ronan, Le Bras and Chandra, Bhagavatula and Yejin, Choi }, year={2019} } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@TevenLeScao](https://github.com/TevenLeScao), [@patrickvonplaten](https://github.com/patrickvonplaten), [@lewtun](https://github.com/lewtun) for adding this dataset.
winogrande
[ "language:en", "region:us" ]
2022-03-02T23:29:22+00:00
{"language": ["en"], "paperswithcode_id": "winogrande", "pretty_name": "WinoGrande", "dataset_info": [{"config_name": "winogrande_xs", "features": [{"name": "sentence", "dtype": "string"}, {"name": "option1", "dtype": "string"}, {"name": "option2", "dtype": "string"}, {"name": "answer", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 20704, "num_examples": 160}, {"name": "test", "num_bytes": 227649, "num_examples": 1767}, {"name": "validation", "num_bytes": 164199, "num_examples": 1267}], "download_size": 3395492, "dataset_size": 412552}, {"config_name": "winogrande_s", "features": [{"name": "sentence", "dtype": "string"}, {"name": "option1", "dtype": "string"}, {"name": "option2", "dtype": "string"}, {"name": "answer", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 82308, "num_examples": 640}, {"name": "test", "num_bytes": 227649, "num_examples": 1767}, {"name": "validation", "num_bytes": 164199, "num_examples": 1267}], "download_size": 3395492, "dataset_size": 474156}, {"config_name": "winogrande_m", "features": [{"name": "sentence", "dtype": "string"}, {"name": "option1", "dtype": "string"}, {"name": "option2", "dtype": "string"}, {"name": "answer", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 329001, "num_examples": 2558}, {"name": "test", "num_bytes": 227649, "num_examples": 1767}, {"name": "validation", "num_bytes": 164199, "num_examples": 1267}], "download_size": 3395492, "dataset_size": 720849}, {"config_name": "winogrande_l", "features": [{"name": "sentence", "dtype": "string"}, {"name": "option1", "dtype": "string"}, {"name": "option2", "dtype": "string"}, {"name": "answer", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1319576, "num_examples": 10234}, {"name": "test", "num_bytes": 227649, "num_examples": 1767}, {"name": "validation", "num_bytes": 164199, "num_examples": 1267}], "download_size": 3395492, "dataset_size": 1711424}, {"config_name": "winogrande_xl", "features": [{"name": "sentence", "dtype": "string"}, {"name": "option1", "dtype": "string"}, {"name": "option2", "dtype": "string"}, {"name": "answer", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 5185832, "num_examples": 40398}, {"name": "test", "num_bytes": 227649, "num_examples": 1767}, {"name": "validation", "num_bytes": 164199, "num_examples": 1267}], "download_size": 3395492, "dataset_size": 5577680}, {"config_name": "winogrande_debiased", "features": [{"name": "sentence", "dtype": "string"}, {"name": "option1", "dtype": "string"}, {"name": "option2", "dtype": "string"}, {"name": "answer", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1203420, "num_examples": 9248}, {"name": "test", "num_bytes": 227649, "num_examples": 1767}, {"name": "validation", "num_bytes": 164199, "num_examples": 1267}], "download_size": 3395492, "dataset_size": 1595268}]}
2024-01-18T11:18:22+00:00
[]
[ "en" ]
TAGS #language-English #region-us
Dataset Card for "winogrande" ============================= Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: * Point of Contact: * Size of downloaded dataset files: 20.37 MB * Size of the generated dataset: 10.50 MB * Total amount of disk used: 30.87 MB ### Dataset Summary WinoGrande is a new collection of 44k problems, inspired by Winograd Schema Challenge (Levesque, Davis, and Morgenstern 2011), but adjusted to improve the scale and robustness against the dataset-specific bias. Formulated as a fill-in-a-blank task with binary options, the goal is to choose the right option for a given sentence which requires commonsense reasoning. ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- ### Data Instances #### winogrande\_debiased * Size of downloaded dataset files: 3.40 MB * Size of the generated dataset: 1.59 MB * Total amount of disk used: 4.99 MB An example of 'train' looks as follows. #### winogrande\_l * Size of downloaded dataset files: 3.40 MB * Size of the generated dataset: 1.71 MB * Total amount of disk used: 5.11 MB An example of 'validation' looks as follows. #### winogrande\_m * Size of downloaded dataset files: 3.40 MB * Size of the generated dataset: 0.72 MB * Total amount of disk used: 4.12 MB An example of 'validation' looks as follows. #### winogrande\_s * Size of downloaded dataset files: 3.40 MB * Size of the generated dataset: 0.47 MB * Total amount of disk used: 3.87 MB An example of 'validation' looks as follows. #### winogrande\_xl * Size of downloaded dataset files: 3.40 MB * Size of the generated dataset: 5.58 MB * Total amount of disk used: 8.98 MB An example of 'train' looks as follows. ### Data Fields The data fields are the same among all splits. #### winogrande\_debiased * 'sentence': a 'string' feature. * 'option1': a 'string' feature. * 'option2': a 'string' feature. * 'answer': a 'string' feature. #### winogrande\_l * 'sentence': a 'string' feature. * 'option1': a 'string' feature. * 'option2': a 'string' feature. * 'answer': a 'string' feature. #### winogrande\_m * 'sentence': a 'string' feature. * 'option1': a 'string' feature. * 'option2': a 'string' feature. * 'answer': a 'string' feature. #### winogrande\_s * 'sentence': a 'string' feature. * 'option1': a 'string' feature. * 'option2': a 'string' feature. * 'answer': a 'string' feature. #### winogrande\_xl * 'sentence': a 'string' feature. * 'option1': a 'string' feature. * 'option2': a 'string' feature. * 'answer': a 'string' feature. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @thomwolf, @TevenLeScao, @patrickvonplaten, @lewtun for adding this dataset.
[ "### Dataset Summary\n\n\nWinoGrande is a new collection of 44k problems, inspired by Winograd Schema Challenge (Levesque, Davis, and Morgenstern\n2011), but adjusted to improve the scale and robustness against the dataset-specific bias. Formulated as a\nfill-in-a-blank task with binary options, the goal is to choose the right option for a given sentence which requires\ncommonsense reasoning.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### winogrande\\_debiased\n\n\n* Size of downloaded dataset files: 3.40 MB\n* Size of the generated dataset: 1.59 MB\n* Total amount of disk used: 4.99 MB\n\n\nAn example of 'train' looks as follows.", "#### winogrande\\_l\n\n\n* Size of downloaded dataset files: 3.40 MB\n* Size of the generated dataset: 1.71 MB\n* Total amount of disk used: 5.11 MB\n\n\nAn example of 'validation' looks as follows.", "#### winogrande\\_m\n\n\n* Size of downloaded dataset files: 3.40 MB\n* Size of the generated dataset: 0.72 MB\n* Total amount of disk used: 4.12 MB\n\n\nAn example of 'validation' looks as follows.", "#### winogrande\\_s\n\n\n* Size of downloaded dataset files: 3.40 MB\n* Size of the generated dataset: 0.47 MB\n* Total amount of disk used: 3.87 MB\n\n\nAn example of 'validation' looks as follows.", "#### winogrande\\_xl\n\n\n* Size of downloaded dataset files: 3.40 MB\n* Size of the generated dataset: 5.58 MB\n* Total amount of disk used: 8.98 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### winogrande\\_debiased\n\n\n* 'sentence': a 'string' feature.\n* 'option1': a 'string' feature.\n* 'option2': a 'string' feature.\n* 'answer': a 'string' feature.", "#### winogrande\\_l\n\n\n* 'sentence': a 'string' feature.\n* 'option1': a 'string' feature.\n* 'option2': a 'string' feature.\n* 'answer': a 'string' feature.", "#### winogrande\\_m\n\n\n* 'sentence': a 'string' feature.\n* 'option1': a 'string' feature.\n* 'option2': a 'string' feature.\n* 'answer': a 'string' feature.", "#### winogrande\\_s\n\n\n* 'sentence': a 'string' feature.\n* 'option1': a 'string' feature.\n* 'option2': a 'string' feature.\n* 'answer': a 'string' feature.", "#### winogrande\\_xl\n\n\n* 'sentence': a 'string' feature.\n* 'option1': a 'string' feature.\n* 'option2': a 'string' feature.\n* 'answer': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @thomwolf, @TevenLeScao, @patrickvonplaten, @lewtun for adding this dataset." ]
[ "TAGS\n#language-English #region-us \n", "### Dataset Summary\n\n\nWinoGrande is a new collection of 44k problems, inspired by Winograd Schema Challenge (Levesque, Davis, and Morgenstern\n2011), but adjusted to improve the scale and robustness against the dataset-specific bias. Formulated as a\nfill-in-a-blank task with binary options, the goal is to choose the right option for a given sentence which requires\ncommonsense reasoning.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### winogrande\\_debiased\n\n\n* Size of downloaded dataset files: 3.40 MB\n* Size of the generated dataset: 1.59 MB\n* Total amount of disk used: 4.99 MB\n\n\nAn example of 'train' looks as follows.", "#### winogrande\\_l\n\n\n* Size of downloaded dataset files: 3.40 MB\n* Size of the generated dataset: 1.71 MB\n* Total amount of disk used: 5.11 MB\n\n\nAn example of 'validation' looks as follows.", "#### winogrande\\_m\n\n\n* Size of downloaded dataset files: 3.40 MB\n* Size of the generated dataset: 0.72 MB\n* Total amount of disk used: 4.12 MB\n\n\nAn example of 'validation' looks as follows.", "#### winogrande\\_s\n\n\n* Size of downloaded dataset files: 3.40 MB\n* Size of the generated dataset: 0.47 MB\n* Total amount of disk used: 3.87 MB\n\n\nAn example of 'validation' looks as follows.", "#### winogrande\\_xl\n\n\n* Size of downloaded dataset files: 3.40 MB\n* Size of the generated dataset: 5.58 MB\n* Total amount of disk used: 8.98 MB\n\n\nAn example of 'train' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### winogrande\\_debiased\n\n\n* 'sentence': a 'string' feature.\n* 'option1': a 'string' feature.\n* 'option2': a 'string' feature.\n* 'answer': a 'string' feature.", "#### winogrande\\_l\n\n\n* 'sentence': a 'string' feature.\n* 'option1': a 'string' feature.\n* 'option2': a 'string' feature.\n* 'answer': a 'string' feature.", "#### winogrande\\_m\n\n\n* 'sentence': a 'string' feature.\n* 'option1': a 'string' feature.\n* 'option2': a 'string' feature.\n* 'answer': a 'string' feature.", "#### winogrande\\_s\n\n\n* 'sentence': a 'string' feature.\n* 'option1': a 'string' feature.\n* 'option2': a 'string' feature.\n* 'answer': a 'string' feature.", "#### winogrande\\_xl\n\n\n* 'sentence': a 'string' feature.\n* 'option1': a 'string' feature.\n* 'option2': a 'string' feature.\n* 'answer': a 'string' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @thomwolf, @TevenLeScao, @patrickvonplaten, @lewtun for adding this dataset." ]
[ 10, 94, 10, 11, 6, 56, 55, 56, 56, 54, 17, 58, 56, 56, 56, 56, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 6, 35 ]
[ "passage: TAGS\n#language-English #region-us \n### Dataset Summary\n\n\nWinoGrande is a new collection of 44k problems, inspired by Winograd Schema Challenge (Levesque, Davis, and Morgenstern\n2011), but adjusted to improve the scale and robustness against the dataset-specific bias. Formulated as a\nfill-in-a-blank task with binary options, the goal is to choose the right option for a given sentence which requires\ncommonsense reasoning.### Supported Tasks and Leaderboards### Languages\n\n\nDataset Structure\n-----------------### Data Instances#### winogrande\\_debiased\n\n\n* Size of downloaded dataset files: 3.40 MB\n* Size of the generated dataset: 1.59 MB\n* Total amount of disk used: 4.99 MB\n\n\nAn example of 'train' looks as follows.#### winogrande\\_l\n\n\n* Size of downloaded dataset files: 3.40 MB\n* Size of the generated dataset: 1.71 MB\n* Total amount of disk used: 5.11 MB\n\n\nAn example of 'validation' looks as follows.#### winogrande\\_m\n\n\n* Size of downloaded dataset files: 3.40 MB\n* Size of the generated dataset: 0.72 MB\n* Total amount of disk used: 4.12 MB\n\n\nAn example of 'validation' looks as follows.#### winogrande\\_s\n\n\n* Size of downloaded dataset files: 3.40 MB\n* Size of the generated dataset: 0.47 MB\n* Total amount of disk used: 3.87 MB\n\n\nAn example of 'validation' looks as follows.#### winogrande\\_xl\n\n\n* Size of downloaded dataset files: 3.40 MB\n* Size of the generated dataset: 5.58 MB\n* Total amount of disk used: 8.98 MB\n\n\nAn example of 'train' looks as follows.### Data Fields\n\n\nThe data fields are the same among all splits.#### winogrande\\_debiased\n\n\n* 'sentence': a 'string' feature.\n* 'option1': a 'string' feature.\n* 'option2': a 'string' feature.\n* 'answer': a 'string' feature." ]
8dda4b5e237452fb939b39326f68ad6607e75ab6
# Dataset Card for "wiqa" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://allenai.org/data/wiqa](https://allenai.org/data/wiqa) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 5.24 MB - **Size of the generated dataset:** 22.40 MB - **Total amount of disk used:** 27.65 MB ### Dataset Summary The WIQA dataset V1 has 39705 questions containing a perturbation and a possible effect in the context of a paragraph. The dataset is split into 29808 train questions, 6894 dev questions and 3003 test questions. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### default - **Size of downloaded dataset files:** 5.24 MB - **Size of the generated dataset:** 22.40 MB - **Total amount of disk used:** 27.65 MB An example of 'validation' looks as follows. ``` { "answer_label": "more", "answer_label_as_choice": "A", "choices": { "label": ["A", "B", "C"], "text": ["more", "less", "no effect"] }, "metadata_graph_id": "481", "metadata_para_id": "528", "metadata_path_len": 3, "metadata_question_id": "influence_graph:528:481:77#0", "metadata_question_type": "INPARA_EFFECT", "question_para_step": ["A male and female rabbit mate", "The female rabbit becomes pregnant", "Baby rabbits form inside of the mother rabbit", "The female rabbit gives birth to a litter", "The newborn rabbits grow up to become adults", "The adult rabbits find mates."], "question_stem": "suppose the female is sterile happens, how will it affect LESS rabbits." } ``` ### Data Fields The data fields are the same among all splits. #### default - `question_stem`: a `string` feature. - `question_para_step`: a `list` of `string` features. - `answer_label`: a `string` feature. - `answer_label_as_choice`: a `string` feature. - `choices`: a dictionary feature containing: - `text`: a `string` feature. - `label`: a `string` feature. - `metadata_question_id`: a `string` feature. - `metadata_graph_id`: a `string` feature. - `metadata_para_id`: a `string` feature. - `metadata_question_type`: a `string` feature. - `metadata_path_len`: a `int32` feature. ### Data Splits | name |train|validation|test| |-------|----:|---------:|---:| |default|29808| 6894|3003| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @article{wiqa, author = {Niket Tandon and Bhavana Dalvi Mishra and Keisuke Sakaguchi and Antoine Bosselut and Peter Clark} title = {WIQA: A dataset for "What if..." reasoning over procedural text}, journal = {arXiv:1909.04739v1}, year = {2019}, } ``` ### Contributions Thanks to [@patrickvonplaten](https://github.com/patrickvonplaten), [@lewtun](https://github.com/lewtun), [@thomwolf](https://github.com/thomwolf), [@mariamabarham](https://github.com/mariamabarham), [@lhoestq](https://github.com/lhoestq) for adding this dataset.
wiqa
[ "language:en", "region:us" ]
2022-03-02T23:29:22+00:00
{"language": ["en"], "paperswithcode_id": "wiqa", "pretty_name": "What-If Question Answering", "dataset_info": {"features": [{"name": "question_stem", "dtype": "string"}, {"name": "question_para_step", "sequence": "string"}, {"name": "answer_label", "dtype": "string"}, {"name": "answer_label_as_choice", "dtype": "string"}, {"name": "choices", "sequence": [{"name": "text", "dtype": "string"}, {"name": "label", "dtype": "string"}]}, {"name": "metadata_question_id", "dtype": "string"}, {"name": "metadata_graph_id", "dtype": "string"}, {"name": "metadata_para_id", "dtype": "string"}, {"name": "metadata_question_type", "dtype": "string"}, {"name": "metadata_path_len", "dtype": "int32"}], "splits": [{"name": "train", "num_bytes": 17089298, "num_examples": 29808}, {"name": "test", "num_bytes": 1532223, "num_examples": 3003}, {"name": "validation", "num_bytes": 3779584, "num_examples": 6894}], "download_size": 5247733, "dataset_size": 22401105}}
2024-01-18T11:18:23+00:00
[]
[ "en" ]
TAGS #language-English #region-us
Dataset Card for "wiqa" ======================= Table of Contents ----------------- * Dataset Description + Dataset Summary + Supported Tasks and Leaderboards + Languages * Dataset Structure + Data Instances + Data Fields + Data Splits * Dataset Creation + Curation Rationale + Source Data + Annotations + Personal and Sensitive Information * Considerations for Using the Data + Social Impact of Dataset + Discussion of Biases + Other Known Limitations * Additional Information + Dataset Curators + Licensing Information + Citation Information + Contributions Dataset Description ------------------- * Homepage: URL * Repository: * Paper: * Point of Contact: * Size of downloaded dataset files: 5.24 MB * Size of the generated dataset: 22.40 MB * Total amount of disk used: 27.65 MB ### Dataset Summary The WIQA dataset V1 has 39705 questions containing a perturbation and a possible effect in the context of a paragraph. The dataset is split into 29808 train questions, 6894 dev questions and 3003 test questions. ### Supported Tasks and Leaderboards ### Languages Dataset Structure ----------------- ### Data Instances #### default * Size of downloaded dataset files: 5.24 MB * Size of the generated dataset: 22.40 MB * Total amount of disk used: 27.65 MB An example of 'validation' looks as follows. ### Data Fields The data fields are the same among all splits. #### default * 'question\_stem': a 'string' feature. * 'question\_para\_step': a 'list' of 'string' features. * 'answer\_label': a 'string' feature. * 'answer\_label\_as\_choice': a 'string' feature. * 'choices': a dictionary feature containing: + 'text': a 'string' feature. + 'label': a 'string' feature. * 'metadata\_question\_id': a 'string' feature. * 'metadata\_graph\_id': a 'string' feature. * 'metadata\_para\_id': a 'string' feature. * 'metadata\_question\_type': a 'string' feature. * 'metadata\_path\_len': a 'int32' feature. ### Data Splits Dataset Creation ---------------- ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information Considerations for Using the Data --------------------------------- ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations Additional Information ---------------------- ### Dataset Curators ### Licensing Information ### Contributions Thanks to @patrickvonplaten, @lewtun, @thomwolf, @mariamabarham, @lhoestq for adding this dataset.
[ "### Dataset Summary\n\n\nThe WIQA dataset V1 has 39705 questions containing a perturbation and a possible effect in the context of a paragraph.\nThe dataset is split into 29808 train questions, 6894 dev questions and 3003 test questions.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### default\n\n\n* Size of downloaded dataset files: 5.24 MB\n* Size of the generated dataset: 22.40 MB\n* Total amount of disk used: 27.65 MB\n\n\nAn example of 'validation' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### default\n\n\n* 'question\\_stem': a 'string' feature.\n* 'question\\_para\\_step': a 'list' of 'string' features.\n* 'answer\\_label': a 'string' feature.\n* 'answer\\_label\\_as\\_choice': a 'string' feature.\n* 'choices': a dictionary feature containing:\n\t+ 'text': a 'string' feature.\n\t+ 'label': a 'string' feature.\n* 'metadata\\_question\\_id': a 'string' feature.\n* 'metadata\\_graph\\_id': a 'string' feature.\n* 'metadata\\_para\\_id': a 'string' feature.\n* 'metadata\\_question\\_type': a 'string' feature.\n* 'metadata\\_path\\_len': a 'int32' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @patrickvonplaten, @lewtun, @thomwolf, @mariamabarham, @lhoestq for adding this dataset." ]
[ "TAGS\n#language-English #region-us \n", "### Dataset Summary\n\n\nThe WIQA dataset V1 has 39705 questions containing a perturbation and a possible effect in the context of a paragraph.\nThe dataset is split into 29808 train questions, 6894 dev questions and 3003 test questions.", "### Supported Tasks and Leaderboards", "### Languages\n\n\nDataset Structure\n-----------------", "### Data Instances", "#### default\n\n\n* Size of downloaded dataset files: 5.24 MB\n* Size of the generated dataset: 22.40 MB\n* Total amount of disk used: 27.65 MB\n\n\nAn example of 'validation' looks as follows.", "### Data Fields\n\n\nThe data fields are the same among all splits.", "#### default\n\n\n* 'question\\_stem': a 'string' feature.\n* 'question\\_para\\_step': a 'list' of 'string' features.\n* 'answer\\_label': a 'string' feature.\n* 'answer\\_label\\_as\\_choice': a 'string' feature.\n* 'choices': a dictionary feature containing:\n\t+ 'text': a 'string' feature.\n\t+ 'label': a 'string' feature.\n* 'metadata\\_question\\_id': a 'string' feature.\n* 'metadata\\_graph\\_id': a 'string' feature.\n* 'metadata\\_para\\_id': a 'string' feature.\n* 'metadata\\_question\\_type': a 'string' feature.\n* 'metadata\\_path\\_len': a 'int32' feature.", "### Data Splits\n\n\n\nDataset Creation\n----------------", "### Curation Rationale", "### Source Data", "#### Initial Data Collection and Normalization", "#### Who are the source language producers?", "### Annotations", "#### Annotation process", "#### Who are the annotators?", "### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------", "### Social Impact of Dataset", "### Discussion of Biases", "### Other Known Limitations\n\n\nAdditional Information\n----------------------", "### Dataset Curators", "### Licensing Information", "### Contributions\n\n\nThanks to @patrickvonplaten, @lewtun, @thomwolf, @mariamabarham, @lhoestq for adding this dataset." ]
[ 10, 54, 10, 11, 6, 50, 17, 207, 11, 7, 4, 10, 10, 5, 5, 9, 18, 7, 8, 14, 6, 6, 39 ]
[ "passage: TAGS\n#language-English #region-us \n### Dataset Summary\n\n\nThe WIQA dataset V1 has 39705 questions containing a perturbation and a possible effect in the context of a paragraph.\nThe dataset is split into 29808 train questions, 6894 dev questions and 3003 test questions.### Supported Tasks and Leaderboards### Languages\n\n\nDataset Structure\n-----------------### Data Instances#### default\n\n\n* Size of downloaded dataset files: 5.24 MB\n* Size of the generated dataset: 22.40 MB\n* Total amount of disk used: 27.65 MB\n\n\nAn example of 'validation' looks as follows.### Data Fields\n\n\nThe data fields are the same among all splits.#### default\n\n\n* 'question\\_stem': a 'string' feature.\n* 'question\\_para\\_step': a 'list' of 'string' features.\n* 'answer\\_label': a 'string' feature.\n* 'answer\\_label\\_as\\_choice': a 'string' feature.\n* 'choices': a dictionary feature containing:\n\t+ 'text': a 'string' feature.\n\t+ 'label': a 'string' feature.\n* 'metadata\\_question\\_id': a 'string' feature.\n* 'metadata\\_graph\\_id': a 'string' feature.\n* 'metadata\\_para\\_id': a 'string' feature.\n* 'metadata\\_question\\_type': a 'string' feature.\n* 'metadata\\_path\\_len': a 'int32' feature.### Data Splits\n\n\n\nDataset Creation\n----------------### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information\n\n\nConsiderations for Using the Data\n---------------------------------### Social Impact of Dataset### Discussion of Biases### Other Known Limitations\n\n\nAdditional Information\n----------------------### Dataset Curators### Licensing Information" ]
c6f339a893fddc7fcb2bda6b9dfea04843d12e3b
# Dataset Card for `wisesight1000` ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://github.com/PyThaiNLP/wisesight-sentiment - **Repository:** https://github.com/PyThaiNLP/wisesight-sentiment/blob/master/word-tokenization/ - **Paper:** - **Leaderboard:** - **Point of Contact:** https://github.com/PyThaiNLP/ ### Dataset Summary `wisesight1000` contains Thai social media texts randomly drawn from the full `wisesight-sentiment`, tokenized by human annotators. Out of the labels `neg` (negative), `neu` (neutral), `pos` (positive), `q` (question), 250 samples each. Some texts are removed because they look like spam. Because these samples are representative of real world content, we believe having these annotaed samples will allow the community to robustly evaluate tokenization algorithms. ### Supported Tasks and Leaderboards word tokenization ### Languages Thai ## Dataset Structure ### Data Instances ``` {'char': ['E', 'u', 'c', 'e', 'r', 'i', 'n', ' ', 'p', 'r', 'o', ' ', 'a', 'c', 'n', 'e', ' ', 'ค', '่', 'ะ', ' ', 'ใ', 'ช', '้', 'แ', 'ล', '้', 'ว', 'ส', 'ิ', 'ว', 'ข', 'ึ', '้', 'น', 'เ', 'พ', 'ิ', '่', 'ม', 'ท', 'ุ', 'ก', 'ว', 'ั', 'น', ' ', 'ม', 'า', 'ด', 'ู', 'ก', 'ั', 'น', 'น', 'ะ', 'ค', 'ะ', ' ', 'ว', '่', 'า', 'จ', 'ั', 'ด', 'ก', 'า', 'ร', 'ป', 'ั', 'ญ', 'ห', 'า', 'ส', 'ิ', 'ว', 'ใ', 'น', '7', 'ว', 'ั', 'น', 'ไ', 'ด', '้', 'ร', 'ึ', 'ม', 'ั', '่', 'ย', 'ย', 'ย', 'ย', 'ย', 'ย', 'ย', 'ย', ' ', 'ล', '่', 'า', 'ส', 'ุ', 'ด', 'ไ', 'ป', 'ล', '้', 'า', 'ง', 'ห', 'น', '้', '…', '\n'], 'char_type': [0, 8, 8, 8, 8, 8, 8, 5, 8, 8, 8, 5, 8, 8, 8, 8, 5, 1, 9, 10, 5, 11, 1, 9, 11, 1, 9, 1, 1, 10, 1, 1, 10, 9, 1, 11, 1, 10, 9, 1, 1, 10, 1, 1, 4, 1, 5, 1, 10, 1, 10, 1, 4, 1, 1, 10, 1, 10, 5, 1, 9, 10, 1, 4, 1, 1, 10, 1, 1, 4, 1, 3, 10, 1, 10, 1, 11, 1, 2, 1, 4, 1, 11, 1, 9, 1, 10, 1, 4, 9, 1, 1, 1, 1, 1, 1, 1, 1, 5, 1, 9, 10, 1, 10, 1, 11, 1, 1, 9, 10, 1, 3, 1, 9, 4, 4], 'is_beginning': [1, 0, 0, 0, 0, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 0, 1, 0, 0, 1, 0, 0, 0, 1, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 0, 1, 0, 0, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 0, 1, 0, 0, 0, 0]} {'char': ['แ', 'พ', 'ง', 'เ', 'ว', '่', 'อ', 'ร', '์', ' ', 'เ', 'บ', 'ี', 'ย', 'ร', '์', 'ช', '้', 'า', 'ง', 'ต', '้', 'น', 'ท', 'ุ', 'น', 'ข', 'ว', 'ด', 'ล', 'ะ', 'ไ', 'ม', '่', 'ถ', 'ึ', 'ง', ' ', '5', '0', ' ', 'ข', 'า', 'ย', ' ', '1', '2', '0', ' ', '😰', '😰', '😰', '์', '\n'], 'char_type': [11, 1, 1, 11, 1, 9, 1, 1, 7, 5, 11, 1, 10, 1, 1, 7, 1, 9, 10, 1, 1, 9, 1, 1, 10, 1, 1, 1, 1, 1, 10, 11, 1, 9, 1, 10, 1, 5, 2, 2, 5, 1, 10, 1, 5, 2, 2, 2, 5, 4, 4, 4, 7, 4], 'is_beginning': [1, 0, 0, 1, 0, 0, 0, 0, 0, 1, 1, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1, 0, 0, 0, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 1, 1, 1, 0, 0]} ``` ### Data Fields - `char`: characters - `char_type`: character types as adopted from []() by [deepcut](https://github.com/rkcosmos/deepcut) - `is_beginning`: 1 if beginning of word else 0 ### Data Splits No explicit split is given. ## Dataset Creation ### Curation Rationale The dataset was created from `wisesight-sentiment` to be a word tokenization benchmark that is closer to texts in the wild, since other Thai word tokenization datasets such as [BEST](https://aiforthai.in.th/corpus.php) are mostly texts from news articles, which do not have some real-world features like misspellings. ### Source Data #### Initial Data Collection and Normalization The data are sampled from `wisesight-sentiment` which has the following data collection and normalization: - Style: Informal and conversational. With some news headlines and advertisement. - Time period: Around 2016 to early 2019. With small amount from other period. - Domains: Mixed. Majority are consumer products and services (restaurants, cosmetics, drinks, car, hotels), with some current affairs. - Privacy: - Only messages that made available to the public on the internet (websites, blogs, social network sites). - For Facebook, this means the public comments (everyone can see) that made on a public page. - Private/protected messages and messages in groups, chat, and inbox are not included. - Usernames and non-public figure names are removed - Phone numbers are masked (e.g. 088-888-8888, 09-9999-9999, 0-2222-2222) - If you see any personal data still remain in the set, please tell us - so we can remove them. - Alternations and modifications: - Keep in mind that this corpus does not statistically represent anything in the language register. - Large amount of messages are not in their original form. Personal data are removed or masked. - Duplicated, leading, and trailing whitespaces are removed. Other punctuations, symbols, and emojis are kept intact. - (Mis)spellings are kept intact. - Messages longer than 2,000 characters are removed. - Long non-Thai messages are removed. Duplicated message (exact match) are removed. #### Who are the source language producers? Social media users in Thailand ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? The annotation was done by several people, including Nitchakarn Chantarapratin, [Pattarawat Chormai](https://github.com/heytitle), [Ponrawee Prasertsom](https://github.com/ponrawee), [Jitkapat Sawatphol](https://github.com/jitkapat), [Nozomi Yamada](https://github.com/nozomiyamada), and [Attapol Rutherford](https://attapol.github.io/). ### Personal and Sensitive Information - The authors tried to exclude any known personally identifiable information from this data set. - Usernames and non-public figure names are removed - Phone numbers are masked (e.g. 088-888-8888, 09-9999-9999, 0-2222-2222) - If you see any personal data still remain in the set, please tell us - so we can remove them. ## Considerations for Using the Data ### Social Impact of Dataset - word tokenization dataset from texts in the wild ### Discussion of Biases - no guideline is given by the authors on word tokenization ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators Thanks [PyThaiNLP](https://github.com/PyThaiNLP/pythainlp) community, [Kitsuchart Pasupa](http://www.it.kmitl.ac.th/~kitsuchart/) (Faculty of Information Technology, King Mongkut's Institute of Technology Ladkrabang), and [Ekapol Chuangsuwanich](https://www.cp.eng.chula.ac.th/en/about/faculty/ekapolc/) (Faculty of Engineering, Chulalongkorn University) for advice. The original Kaggle competition, using the first version of this corpus, can be found at https://www.kaggle.com/c/wisesight-sentiment/ ### Licensing Information CC0 ### Citation Information Dataset: ``` @software{bact_2019_3457447, author = {Suriyawongkul, Arthit and Chuangsuwanich, Ekapol and Chormai, Pattarawat and Polpanumas, Charin}, title = {PyThaiNLP/wisesight-sentiment: First release}, month = sep, year = 2019, publisher = {Zenodo}, version = {v1.0}, doi = {10.5281/zenodo.3457447}, url = {https://doi.org/10.5281/zenodo.3457447} } ``` Character type features: ``` @inproceedings{haruechaiyasak2009tlex, title={TLex: Thai lexeme analyser based on the conditional random fields}, author={Haruechaiyasak, Choochart and Kongyoung, Sarawoot}, booktitle={Proceedings of 8th International Symposium on Natural Language Processing}, year={2009} } ``` ### Contributions Thanks to [@cstorm125](https://github.com/cstorm125) for adding this dataset.
wisesight1000
[ "task_categories:token-classification", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:n<1K", "source_datasets:extended|wisesight_sentiment", "language:th", "license:cc0-1.0", "word-tokenization", "region:us" ]
2022-03-02T23:29:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found"], "language": ["th"], "license": ["cc0-1.0"], "multilinguality": ["monolingual"], "size_categories": ["n<1K"], "source_datasets": ["extended|wisesight_sentiment"], "task_categories": ["token-classification"], "task_ids": [], "pretty_name": "wisesight1000", "tags": ["word-tokenization"], "dataset_info": {"features": [{"name": "char", "sequence": "string"}, {"name": "char_type", "sequence": {"class_label": {"names": {"0": "b_e", "1": "c", "2": "d", "3": "n", "4": "o", "5": "p", "6": "q", "7": "s", "8": "s_e", "9": "t", "10": "v", "11": "w"}}}}, {"name": "is_beginning", "sequence": {"class_label": {"names": {"0": "neg", "1": "pos"}}}}], "config_name": "wisesight1000", "splits": [{"name": "train", "num_bytes": 1735438, "num_examples": 993}], "download_size": 222691, "dataset_size": 1735438}}
2023-06-14T07:20:50+00:00
[]
[ "th" ]
TAGS #task_categories-token-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-n<1K #source_datasets-extended|wisesight_sentiment #language-Thai #license-cc0-1.0 #word-tokenization #region-us
# Dataset Card for 'wisesight1000' ## Table of Contents - Dataset Description - Dataset Summary - Supported Tasks and Leaderboards - Languages - Dataset Structure - Data Instances - Data Fields - Data Splits - Dataset Creation - Curation Rationale - Source Data - Annotations - Personal and Sensitive Information - Considerations for Using the Data - Social Impact of Dataset - Discussion of Biases - Other Known Limitations - Additional Information - Dataset Curators - Licensing Information - Citation Information - Contributions ## Dataset Description - Homepage: URL - Repository: URL - Paper: - Leaderboard: - Point of Contact: URL ### Dataset Summary 'wisesight1000' contains Thai social media texts randomly drawn from the full 'wisesight-sentiment', tokenized by human annotators. Out of the labels 'neg' (negative), 'neu' (neutral), 'pos' (positive), 'q' (question), 250 samples each. Some texts are removed because they look like spam. Because these samples are representative of real world content, we believe having these annotaed samples will allow the community to robustly evaluate tokenization algorithms. ### Supported Tasks and Leaderboards word tokenization ### Languages Thai ## Dataset Structure ### Data Instances ### Data Fields - 'char': characters - 'char_type': character types as adopted from []() by deepcut - 'is_beginning': 1 if beginning of word else 0 ### Data Splits No explicit split is given. ## Dataset Creation ### Curation Rationale The dataset was created from 'wisesight-sentiment' to be a word tokenization benchmark that is closer to texts in the wild, since other Thai word tokenization datasets such as BEST are mostly texts from news articles, which do not have some real-world features like misspellings. ### Source Data #### Initial Data Collection and Normalization The data are sampled from 'wisesight-sentiment' which has the following data collection and normalization: - Style: Informal and conversational. With some news headlines and advertisement. - Time period: Around 2016 to early 2019. With small amount from other period. - Domains: Mixed. Majority are consumer products and services (restaurants, cosmetics, drinks, car, hotels), with some current affairs. - Privacy: - Only messages that made available to the public on the internet (websites, blogs, social network sites). - For Facebook, this means the public comments (everyone can see) that made on a public page. - Private/protected messages and messages in groups, chat, and inbox are not included. - Usernames and non-public figure names are removed - Phone numbers are masked (e.g. 088-888-8888, 09-9999-9999, 0-2222-2222) - If you see any personal data still remain in the set, please tell us - so we can remove them. - Alternations and modifications: - Keep in mind that this corpus does not statistically represent anything in the language register. - Large amount of messages are not in their original form. Personal data are removed or masked. - Duplicated, leading, and trailing whitespaces are removed. Other punctuations, symbols, and emojis are kept intact. - (Mis)spellings are kept intact. - Messages longer than 2,000 characters are removed. - Long non-Thai messages are removed. Duplicated message (exact match) are removed. #### Who are the source language producers? Social media users in Thailand ### Annotations #### Annotation process #### Who are the annotators? The annotation was done by several people, including Nitchakarn Chantarapratin, Pattarawat Chormai, Ponrawee Prasertsom, Jitkapat Sawatphol, Nozomi Yamada, and Attapol Rutherford. ### Personal and Sensitive Information - The authors tried to exclude any known personally identifiable information from this data set. - Usernames and non-public figure names are removed - Phone numbers are masked (e.g. 088-888-8888, 09-9999-9999, 0-2222-2222) - If you see any personal data still remain in the set, please tell us - so we can remove them. ## Considerations for Using the Data ### Social Impact of Dataset - word tokenization dataset from texts in the wild ### Discussion of Biases - no guideline is given by the authors on word tokenization ### Other Known Limitations ## Additional Information ### Dataset Curators Thanks PyThaiNLP community, Kitsuchart Pasupa (Faculty of Information Technology, King Mongkut's Institute of Technology Ladkrabang), and Ekapol Chuangsuwanich (Faculty of Engineering, Chulalongkorn University) for advice. The original Kaggle competition, using the first version of this corpus, can be found at URL ### Licensing Information CC0 Dataset: Character type features: ### Contributions Thanks to @cstorm125 for adding this dataset.
[ "# Dataset Card for 'wisesight1000'", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper:\n- Leaderboard:\n- Point of Contact: URL", "### Dataset Summary\n\n'wisesight1000' contains Thai social media texts randomly drawn from the full 'wisesight-sentiment', tokenized by human annotators.\nOut of the labels 'neg' (negative), 'neu' (neutral), 'pos' (positive), 'q' (question), 250 samples each. Some texts are removed because they look like spam. Because these samples are representative of real world content, we believe having these annotaed samples will allow the community to robustly evaluate tokenization algorithms.", "### Supported Tasks and Leaderboards\n\nword tokenization", "### Languages\n\nThai", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- 'char': characters\n- 'char_type': character types as adopted from []() by deepcut\n- 'is_beginning': 1 if beginning of word else 0", "### Data Splits\n\nNo explicit split is given.", "## Dataset Creation", "### Curation Rationale\n\nThe dataset was created from 'wisesight-sentiment' to be a word tokenization benchmark that is closer to texts in the wild, since other Thai word tokenization datasets such as BEST are mostly texts from news articles, which do not have some real-world features like misspellings.", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe data are sampled from 'wisesight-sentiment' which has the following data collection and normalization:\n- Style: Informal and conversational. With some news headlines and advertisement.\n- Time period: Around 2016 to early 2019. With small amount from other period.\n- Domains: Mixed. Majority are consumer products and services (restaurants, cosmetics, drinks, car, hotels), with some current affairs.\n- Privacy:\n - Only messages that made available to the public on the internet (websites, blogs, social network sites).\n - For Facebook, this means the public comments (everyone can see) that made on a public page.\n - Private/protected messages and messages in groups, chat, and inbox are not included.\n - Usernames and non-public figure names are removed\n - Phone numbers are masked (e.g. 088-888-8888, 09-9999-9999, 0-2222-2222)\n - If you see any personal data still remain in the set, please tell us - so we can remove them.\n- Alternations and modifications:\n - Keep in mind that this corpus does not statistically represent anything in the language register.\n - Large amount of messages are not in their original form. Personal data are removed or masked.\n - Duplicated, leading, and trailing whitespaces are removed. Other punctuations, symbols, and emojis are kept intact.\n - (Mis)spellings are kept intact.\n - Messages longer than 2,000 characters are removed.\n - Long non-Thai messages are removed. Duplicated message (exact match) are removed.", "#### Who are the source language producers?\n\nSocial media users in Thailand", "### Annotations", "#### Annotation process", "#### Who are the annotators?\n\nThe annotation was done by several people, including Nitchakarn Chantarapratin, Pattarawat Chormai, Ponrawee Prasertsom, Jitkapat Sawatphol, Nozomi Yamada, and Attapol Rutherford.", "### Personal and Sensitive Information\n\n- The authors tried to exclude any known personally identifiable information from this data set.\n- Usernames and non-public figure names are removed\n- Phone numbers are masked (e.g. 088-888-8888, 09-9999-9999, 0-2222-2222)\n- If you see any personal data still remain in the set, please tell us - so we can remove them.", "## Considerations for Using the Data", "### Social Impact of Dataset\n\n- word tokenization dataset from texts in the wild", "### Discussion of Biases\n\n- no guideline is given by the authors on word tokenization", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThanks PyThaiNLP community, Kitsuchart Pasupa (Faculty of Information Technology, King Mongkut's Institute of Technology Ladkrabang), and Ekapol Chuangsuwanich (Faculty of Engineering, Chulalongkorn University) for advice. The original Kaggle competition, using the first version of this corpus, can be found at URL", "### Licensing Information\n\nCC0\n\n\n\nDataset:\n\n\nCharacter type features:", "### Contributions\n\nThanks to @cstorm125 for adding this dataset." ]
[ "TAGS\n#task_categories-token-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-n<1K #source_datasets-extended|wisesight_sentiment #language-Thai #license-cc0-1.0 #word-tokenization #region-us \n", "# Dataset Card for 'wisesight1000'", "## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions", "## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper:\n- Leaderboard:\n- Point of Contact: URL", "### Dataset Summary\n\n'wisesight1000' contains Thai social media texts randomly drawn from the full 'wisesight-sentiment', tokenized by human annotators.\nOut of the labels 'neg' (negative), 'neu' (neutral), 'pos' (positive), 'q' (question), 250 samples each. Some texts are removed because they look like spam. Because these samples are representative of real world content, we believe having these annotaed samples will allow the community to robustly evaluate tokenization algorithms.", "### Supported Tasks and Leaderboards\n\nword tokenization", "### Languages\n\nThai", "## Dataset Structure", "### Data Instances", "### Data Fields\n\n- 'char': characters\n- 'char_type': character types as adopted from []() by deepcut\n- 'is_beginning': 1 if beginning of word else 0", "### Data Splits\n\nNo explicit split is given.", "## Dataset Creation", "### Curation Rationale\n\nThe dataset was created from 'wisesight-sentiment' to be a word tokenization benchmark that is closer to texts in the wild, since other Thai word tokenization datasets such as BEST are mostly texts from news articles, which do not have some real-world features like misspellings.", "### Source Data", "#### Initial Data Collection and Normalization\n\nThe data are sampled from 'wisesight-sentiment' which has the following data collection and normalization:\n- Style: Informal and conversational. With some news headlines and advertisement.\n- Time period: Around 2016 to early 2019. With small amount from other period.\n- Domains: Mixed. Majority are consumer products and services (restaurants, cosmetics, drinks, car, hotels), with some current affairs.\n- Privacy:\n - Only messages that made available to the public on the internet (websites, blogs, social network sites).\n - For Facebook, this means the public comments (everyone can see) that made on a public page.\n - Private/protected messages and messages in groups, chat, and inbox are not included.\n - Usernames and non-public figure names are removed\n - Phone numbers are masked (e.g. 088-888-8888, 09-9999-9999, 0-2222-2222)\n - If you see any personal data still remain in the set, please tell us - so we can remove them.\n- Alternations and modifications:\n - Keep in mind that this corpus does not statistically represent anything in the language register.\n - Large amount of messages are not in their original form. Personal data are removed or masked.\n - Duplicated, leading, and trailing whitespaces are removed. Other punctuations, symbols, and emojis are kept intact.\n - (Mis)spellings are kept intact.\n - Messages longer than 2,000 characters are removed.\n - Long non-Thai messages are removed. Duplicated message (exact match) are removed.", "#### Who are the source language producers?\n\nSocial media users in Thailand", "### Annotations", "#### Annotation process", "#### Who are the annotators?\n\nThe annotation was done by several people, including Nitchakarn Chantarapratin, Pattarawat Chormai, Ponrawee Prasertsom, Jitkapat Sawatphol, Nozomi Yamada, and Attapol Rutherford.", "### Personal and Sensitive Information\n\n- The authors tried to exclude any known personally identifiable information from this data set.\n- Usernames and non-public figure names are removed\n- Phone numbers are masked (e.g. 088-888-8888, 09-9999-9999, 0-2222-2222)\n- If you see any personal data still remain in the set, please tell us - so we can remove them.", "## Considerations for Using the Data", "### Social Impact of Dataset\n\n- word tokenization dataset from texts in the wild", "### Discussion of Biases\n\n- no guideline is given by the authors on word tokenization", "### Other Known Limitations", "## Additional Information", "### Dataset Curators\n\nThanks PyThaiNLP community, Kitsuchart Pasupa (Faculty of Information Technology, King Mongkut's Institute of Technology Ladkrabang), and Ekapol Chuangsuwanich (Faculty of Engineering, Chulalongkorn University) for advice. The original Kaggle competition, using the first version of this corpus, can be found at URL", "### Licensing Information\n\nCC0\n\n\n\nDataset:\n\n\nCharacter type features:", "### Contributions\n\nThanks to @cstorm125 for adding this dataset." ]
[ 93, 11, 120, 27, 129, 14, 5, 6, 6, 47, 11, 5, 75, 4, 352, 15, 5, 5, 64, 91, 8, 20, 23, 7, 5, 85, 17, 17 ]
[ "passage: TAGS\n#task_categories-token-classification #annotations_creators-expert-generated #language_creators-found #multilinguality-monolingual #size_categories-n<1K #source_datasets-extended|wisesight_sentiment #language-Thai #license-cc0-1.0 #word-tokenization #region-us \n# Dataset Card for 'wisesight1000'## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n - Personal and Sensitive Information\n- Considerations for Using the Data\n - Social Impact of Dataset\n - Discussion of Biases\n - Other Known Limitations\n- Additional Information\n - Dataset Curators\n - Licensing Information\n - Citation Information\n - Contributions## Dataset Description\n\n- Homepage: URL\n- Repository: URL\n- Paper:\n- Leaderboard:\n- Point of Contact: URL### Dataset Summary\n\n'wisesight1000' contains Thai social media texts randomly drawn from the full 'wisesight-sentiment', tokenized by human annotators.\nOut of the labels 'neg' (negative), 'neu' (neutral), 'pos' (positive), 'q' (question), 250 samples each. Some texts are removed because they look like spam. Because these samples are representative of real world content, we believe having these annotaed samples will allow the community to robustly evaluate tokenization algorithms.### Supported Tasks and Leaderboards\n\nword tokenization### Languages\n\nThai## Dataset Structure### Data Instances### Data Fields\n\n- 'char': characters\n- 'char_type': character types as adopted from []() by deepcut\n- 'is_beginning': 1 if beginning of word else 0### Data Splits\n\nNo explicit split is given.## Dataset Creation", "passage: ### Curation Rationale\n\nThe dataset was created from 'wisesight-sentiment' to be a word tokenization benchmark that is closer to texts in the wild, since other Thai word tokenization datasets such as BEST are mostly texts from news articles, which do not have some real-world features like misspellings.### Source Data#### Initial Data Collection and Normalization\n\nThe data are sampled from 'wisesight-sentiment' which has the following data collection and normalization:\n- Style: Informal and conversational. With some news headlines and advertisement.\n- Time period: Around 2016 to early 2019. With small amount from other period.\n- Domains: Mixed. Majority are consumer products and services (restaurants, cosmetics, drinks, car, hotels), with some current affairs.\n- Privacy:\n - Only messages that made available to the public on the internet (websites, blogs, social network sites).\n - For Facebook, this means the public comments (everyone can see) that made on a public page.\n - Private/protected messages and messages in groups, chat, and inbox are not included.\n - Usernames and non-public figure names are removed\n - Phone numbers are masked (e.g. 088-888-8888, 09-9999-9999, 0-2222-2222)\n - If you see any personal data still remain in the set, please tell us - so we can remove them.\n- Alternations and modifications:\n - Keep in mind that this corpus does not statistically represent anything in the language register.\n - Large amount of messages are not in their original form. Personal data are removed or masked.\n - Duplicated, leading, and trailing whitespaces are removed. Other punctuations, symbols, and emojis are kept intact.\n - (Mis)spellings are kept intact.\n - Messages longer than 2,000 characters are removed.\n - Long non-Thai messages are removed. Duplicated message (exact match) are removed.#### Who are the source language producers?\n\nSocial media users in Thailand### Annotations#### Annotation process#### Who are the annotators?\n\nThe annotation was done by several people, including Nitchakarn Chantarapratin, Pattarawat Chormai, Ponrawee Prasertsom, Jitkapat Sawatphol, Nozomi Yamada, and Attapol Rutherford." ]