sha
stringlengths
40
40
text
stringlengths
0
13.4M
id
stringlengths
2
117
tags
sequence
created_at
stringlengths
25
25
metadata
stringlengths
2
31.7M
last_modified
stringlengths
25
25
29b145026c2ec661f3ad581e42f73c68be4f4a13
# Dataset Description Out of **20,577** human proteins (from [UniProt human proteome](https://www.uniprot.org/proteomes/UP000005640)), sequences shorter than 20 amino acids or longer than 512 amino acids were removed, resulting in a set of **12,703** proteins. The uShuffle algorithm ([python pacakge](https://github.com/guma44/ushuffle)) was then used to shuffle these protein sequences while maintaining their singlet distribution. Afterwards, h-CD-HIT algorithm ([web server](http://weizhong-lab.ucsd.edu/cdhit-web-server/cgi-bin/index.cgi)) was used with three subsequent filter stages at pairwise identity cutoffs of 0.9, 0.5 and 0.1, resulting in a total of **11,698** sequences. # **Citation** If you use this dataset, please cite our paper: ``` @article { author = {Geffen, Yaron and Ofran, Yanay and Unger, Ron}, title = {DistilProtBert: A distilled protein language model used to distinguish between real proteins and their randomly shuffled counterparts}, year = {2022}, doi = {10.1093/bioinformatics/btac474}, URL = {https://doi.org/10.1093/bioinformatics/btac474}, journal = {Bioinformatics} } ```
yarongef/human_proteome_singlets
[ "license:mit", "region:us" ]
2022-04-12T07:20:31+00:00
{"license": "mit"}
2022-09-21T07:45:02+00:00
68b2211a56c1f2a2276d10c2d0f31a416ed9a2c9
# Dataset Card for Snacks ## Dataset Summary This is a dataset of 20 different types of snack foods that accompanies the book [Machine Learning by Tutorials](https://www.raywenderlich.com/books/machine-learning-by-tutorials/v2.0). The images were taken from the [Google Open Images dataset](https://storage.googleapis.com/openimages/web/index.html), release 2017_11. ## Dataset Structure Number of images in the train/validation/test splits: ```nohighlight train 4838 val 955 test 952 total 6745 ``` Total images in each category: ```nohighlight apple 350 banana 350 cake 349 candy 349 carrot 349 cookie 349 doughnut 350 grape 350 hot dog 350 ice cream 350 juice 350 muffin 348 orange 349 pineapple 340 popcorn 260 pretzel 204 salad 350 strawberry 348 waffle 350 watermelon 350 ``` To save space in the download, the images were resized so that their smallest side is 256 pixels. All EXIF information was removed. ### Data Splits Train, Test, Validation ## Licensing Information Just like the images from Google Open Images, the snacks dataset is licensed under the terms of the Creative Commons license. The images are listed as having a [CC BY 2.0](https://creativecommons.org/licenses/by/2.0/) license. The annotations are licensed by Google Inc. under a [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) license. The **credits.csv** file contains the original URL, author information and license for each image.
Matthijs/snacks
[ "task_categories:image-classification", "license:cc-by-4.0", "region:us" ]
2022-04-12T07:30:24+00:00
{"license": "cc-by-4.0", "task_categories": ["image-classification", "computer-vision"], "pretty_name": "Snacks"}
2022-04-12T13:26:59+00:00
14aba009b5fcd97b1a9ee6f3e3b0da0e308cf7cb
### Dataset Summary This dataset is extracted from Fever dataset (https://fever.ai), pre-processed and ready to train and evaluate. The training objective is a text classification task - given a claim and evidence, predict if evidence is related to claim.
mwong/fever-evidence-related
[ "task_categories:text-classification", "task_ids:fact-checking", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:extended|fever", "language:en", "license:cc-by-sa-3.0", "license:gpl-3.0", "region:us" ]
2022-04-12T07:39:59+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["cc-by-sa-3.0", "gpl-3.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["extended|fever"], "task_categories": ["text-classification"], "task_ids": ["fact-checking"], "paperswithcode_id": "fever", "pretty_name": "fever"}
2022-10-25T09:06:51+00:00
0f4beb45dfe3526ba6e132375982ca27cee6eb47
yuanjie/demo
[ "license:apache-2.0", "region:us" ]
2022-04-12T09:00:12+00:00
{"license": "apache-2.0"}
2022-04-12T09:00:12+00:00
e53f048856ff4f594e959d75785d2c2d37b678ee
# Dataset Card for GSM8K ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://openai.com/blog/grade-school-math/ - **Repository:** https://github.com/openai/grade-school-math - **Paper:** https://arxiv.org/abs/2110.14168 - **Leaderboard:** [Needs More Information] - **Point of Contact:** [Needs More Information] ### Dataset Summary GSM8K (Grade School Math 8K) is a dataset of 8.5K high quality linguistically diverse grade school math word problems. The dataset was created to support the task of question answering on basic mathematical problems that require multi-step reasoning. - These problems take between 2 and 8 steps to solve. - Solutions primarily involve performing a sequence of elementary calculations using basic arithmetic operations (+ − ×÷) to reach the final answer. - A bright middle school student should be able to solve every problem: from the paper, "Problems require no concepts beyond the level of early Algebra, and the vast majority of problems can be solved without explicitly defining a variable." - Solutions are provided in natural language, as opposed to pure math expressions. From the paper: "We believe this is the most generally useful data format, and we expect it to shed light on the properties of large language models’ internal monologues"" ### Supported Tasks and Leaderboards This dataset is generally used to test logic and math in language modelling. It has been used for many benchmarks, including the [LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). ### Languages The text in the dataset is in English. The associated BCP-47 code is `en`. ## Dataset Structure ### Data Instances For the `main` configuration, each instance contains a string for the grade-school level math question and a string for the corresponding answer with multiple steps of reasoning and calculator annotations (explained [here](https://github.com/openai/grade-school-math#calculation-annotations)). ```python { 'question': 'Natalia sold clips to 48 of her friends in April, and then she sold half as many clips in May. How many clips did Natalia sell altogether in April and May?', 'answer': 'Natalia sold 48/2 = <<48/2=24>>24 clips in May.\nNatalia sold 48+24 = <<48+24=72>>72 clips altogether in April and May.\n#### 72', } ``` For the `socratic` configuration, each instance contains a string for a grade-school level math question, a string for the corresponding answer with multiple steps of reasoning, calculator annotations (explained [here](https://github.com/openai/grade-school-math#calculation-annotations)), and *Socratic sub-questions*. ```python { 'question': 'Natalia sold clips to 48 of her friends in April, and then she sold half as many clips in May. How many clips did Natalia sell altogether in April and May?', 'answer': 'How many clips did Natalia sell in May? ** Natalia sold 48/2 = <<48/2=24>>24 clips in May.\nHow many clips did Natalia sell altogether in April and May? ** Natalia sold 48+24 = <<48+24=72>>72 clips altogether in April and May.\n#### 72', } ``` ### Data Fields The data fields are the same among `main` and `socratic` configurations and their individual splits. - question: The question string to a grade school math problem. - answer: The full solution string to the `question`. It contains multiple steps of reasoning with calculator annotations and the final numeric solution. ### Data Splits | name |train|validation| |--------|----:|---------:| |main | 7473| 1319| |socratic| 7473| 1319| ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization From the paper, appendix A: > We initially collected a starting set of a thousand problems and natural language solutions by hiring freelance contractors on Upwork (upwork.com). We then worked with Surge AI (surgehq.ai), an NLP data labeling platform, to scale up our data collection. After collecting the full dataset, we asked workers to re-solve all problems, with no workers re-solving problems they originally wrote. We checked whether their final answers agreed with the original solutions, and any problems that produced disagreements were either repaired or discarded. We then performed another round of agreement checks on a smaller subset of problems, finding that 1.7% of problems still produce disagreements among contractors. We estimate this to be the fraction of problems that contain breaking errors or ambiguities. It is possible that a larger percentage of problems contain subtle errors. #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? Surge AI (surgehq.ai) ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information The GSM8K dataset is licensed under the [MIT License](https://opensource.org/licenses/MIT). ### Citation Information ```bibtex @article{cobbe2021gsm8k, title={Training Verifiers to Solve Math Word Problems}, author={Cobbe, Karl and Kosaraju, Vineet and Bavarian, Mohammad and Chen, Mark and Jun, Heewoo and Kaiser, Lukasz and Plappert, Matthias and Tworek, Jerry and Hilton, Jacob and Nakano, Reiichiro and Hesse, Christopher and Schulman, John}, journal={arXiv preprint arXiv:2110.14168}, year={2021} } ``` ### Contributions Thanks to [@jon-tow](https://github.com/jon-tow) for adding this dataset.
gsm8k
[ "task_categories:text2text-generation", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:mit", "math-word-problems", "arxiv:2110.14168", "region:us" ]
2022-04-12T09:22:10+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["mit"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text2text-generation"], "task_ids": [], "paperswithcode_id": "gsm8k", "pretty_name": "Grade School Math 8K", "tags": ["math-word-problems"], "dataset_info": [{"config_name": "main", "features": [{"name": "question", "dtype": "string"}, {"name": "answer", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 3963202, "num_examples": 7473}, {"name": "test", "num_bytes": 713732, "num_examples": 1319}], "download_size": 2725633, "dataset_size": 4676934}, {"config_name": "socratic", "features": [{"name": "question", "dtype": "string"}, {"name": "answer", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 5198108, "num_examples": 7473}, {"name": "test", "num_bytes": 936859, "num_examples": 1319}], "download_size": 3164254, "dataset_size": 6134967}], "configs": [{"config_name": "main", "data_files": [{"split": "train", "path": "main/train-*"}, {"split": "test", "path": "main/test-*"}]}, {"config_name": "socratic", "data_files": [{"split": "train", "path": "socratic/train-*"}, {"split": "test", "path": "socratic/test-*"}]}]}
2024-01-04T12:05:15+00:00
d7ec87c8656a85d795e41f727f9ef2286310eee9
# Dataset Card for SBU Captioned Photo Dataset ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Dataset Preprocessing](#dataset-preprocessing) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.cs.rice.edu/~vo9/sbucaptions/ - **Repository:** - **Paper:** [Im2Text: Describing Images Using 1 Million Captioned Photographs](https://papers.nips.cc/paper/2011/hash/5dd9db5e033da9c6fb5ba83c7a7ebea9-Abstract.html) - **Leaderboard:** - **Point of Contact:** [Vicente Ordóñez Román](mailto:vicenteor@rice.edu) ### Dataset Summary SBU Captioned Photo Dataset is a collection of associated captions and images from Flickr. ### Dataset Preprocessing This dataset doesn't download the images locally by default. Instead, it exposes URLs to the images. To fetch the images, use the following code: ```python from concurrent.futures import ThreadPoolExecutor from functools import partial import io import urllib import PIL.Image from datasets import load_dataset from datasets.utils.file_utils import get_datasets_user_agent USER_AGENT = get_datasets_user_agent() def fetch_single_image(image_url, timeout=None, retries=0): for _ in range(retries + 1): try: request = urllib.request.Request( image_url, data=None, headers={"user-agent": USER_AGENT}, ) with urllib.request.urlopen(request, timeout=timeout) as req: image = PIL.Image.open(io.BytesIO(req.read())) break except Exception: image = None return image def fetch_images(batch, num_threads, timeout=None, retries=0): fetch_single_image_with_args = partial(fetch_single_image, timeout=timeout, retries=retries) with ThreadPoolExecutor(max_workers=num_threads) as executor: batch["image"] = list(executor.map(fetch_single_image_with_args, batch["image_url"])) return batch num_threads = 20 dset = load_dataset("sbu_captions") dset = dset.map(fetch_images, batched=True, batch_size=100, fn_kwargs={"num_threads": num_threads}) ``` ### Supported Tasks and Leaderboards - `image-to-text`: This dataset can be used to train a model for Image Captioning where the goal is to predict a caption given the image. ### Languages All captions are in English. ## Dataset Structure ### Data Instances Each instance in SBU Captioned Photo Dataset represents a single image with a caption and a user_id: ``` { 'img_url': 'http://static.flickr.com/2723/4385058960_b0f291553e.jpg', 'user_id': '47889917@N08', 'caption': 'A wooden chair in the living room' } ``` ### Data Fields - `image_url`: Static URL for downloading the image associated with the post. - `caption`: Textual description of the image. - `user_id`: Author of caption. ### Data Splits All the data is contained in training split. The training set has 1M instances. ## Dataset Creation ### Curation Rationale From the paper: > One contribution is our technique for the automatic collection of this new dataset – performing a huge number of Flickr queries and then filtering the noisy results down to 1 million images with associated visually relevant captions. Such a collection allows us to approach the extremely challenging problem of description generation using relatively simple non-parametric methods and produces surprisingly effective results. ### Source Data The source images come from Flickr. #### Initial Data Collection and Normalization One key contribution of our paper is a novel web-scale database of photographs with associated descriptive text. To enable effective captioning of novel images, this database must be good in two ways: 1) It must be large so that image based matches to a query are reasonably similar, 2) The captions associated with the data base photographs must be visually relevant so that transferring captions between pictures is useful. To achieve the first requirement we query Flickr using a huge number of pairs of query terms (objects, attributes, actions, stuff, and scenes). This produces a very large, but noisy initial set of photographs with associated text. #### Who are the source language producers? The Flickr users. ### Annotations #### Annotation process Text descriptions associated with the images are inherited as annotations/captions. #### Who are the annotators? The Flickr users. ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators Vicente Ordonez, Girish Kulkarni and Tamara L. Berg. ### Licensing Information Not specified. ### Citation Information ```bibtex @inproceedings{NIPS2011_5dd9db5e, author = {Ordonez, Vicente and Kulkarni, Girish and Berg, Tamara}, booktitle = {Advances in Neural Information Processing Systems}, editor = {J. Shawe-Taylor and R. Zemel and P. Bartlett and F. Pereira and K.Q. Weinberger}, pages = {}, publisher = {Curran Associates, Inc.}, title = {Im2Text: Describing Images Using 1 Million Captioned Photographs}, url = {https://proceedings.neurips.cc/paper/2011/file/5dd9db5e033da9c6fb5ba83c7a7ebea9-Paper.pdf}, volume = {24}, year = {2011} } ``` ### Contributions Thanks to [@thomasw21](https://github.com/thomasw21) for adding this dataset
sbu_captions
[ "task_categories:image-to-text", "task_ids:image-captioning", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:1M<n<10M", "source_datasets:original", "language:en", "license:unknown", "region:us" ]
2022-04-12T09:41:52+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["1M<n<10M"], "source_datasets": ["original"], "task_categories": ["image-to-text"], "task_ids": ["image-captioning"], "paperswithcode_id": "sbu-captions-dataset", "pretty_name": "SBU Captioned Photo Dataset", "dataset_info": {"features": [{"name": "image_url", "dtype": "string"}, {"name": "user_id", "dtype": "string"}, {"name": "caption", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 143795586, "num_examples": 1000000}], "download_size": 49787719, "dataset_size": 143795586}}
2024-01-18T11:19:05+00:00
4e2d0dd3b956ed9e3c623e087a9e800afee79572
# Dataset Description Out of **20,577** human proteins (from [UniProt human proteome](https://www.uniprot.org/proteomes/UP000005640)), sequences shorter than 20 amino acids or longer than 512 amino acids were removed, resulting in a set of **12,703** proteins. The uShuffle algorithm ([python pacakge](https://github.com/guma44/ushuffle)) was then used to shuffle these protein sequences while maintaining their doublet distribution. The very few sequences for which uShuffle failed to create a shuffled version were eliminated. Afterwards, h-CD-HIT algorithm ([web server](http://weizhong-lab.ucsd.edu/cdhit-web-server/cgi-bin/index.cgi)) was used with three subsequent filter stages at pairwise identity cutoffs of 0.9, 0.5 and 0.1, resulting in a total of **11,658** sequences. # Citation If you use this dataset, please cite our paper: ``` @article { author = {Geffen, Yaron and Ofran, Yanay and Unger, Ron}, title = {DistilProtBert: A distilled protein language model used to distinguish between real proteins and their randomly shuffled counterparts}, year = {2022}, doi = {10.1093/bioinformatics/btac474}, URL = {https://doi.org/10.1093/bioinformatics/btac474}, journal = {Bioinformatics} } ```
yarongef/human_proteome_doublets
[ "license:mit", "region:us" ]
2022-04-12T09:42:22+00:00
{"license": "mit"}
2022-09-21T07:43:43+00:00
7a048b602b00c84360b440c8f13b198b67c14ae4
# Dataset Description Out of **20,577** human proteins (from [UniProt human proteome](https://www.uniprot.org/proteomes/UP000005640)), sequences shorter than 20 amino acids or longer than 512 amino acids were removed, resulting in a set of **12,703** proteins. The uShuffle algorithm ([python pacakge](https://github.com/guma44/ushuffle)) was then used to shuffle these protein sequences while maintaining their triplet distribution. The sequences for which uShuffle failed to create a shuffled version were eliminated. Afterwards, h-CD-HIT algorithm ([web server](http://weizhong-lab.ucsd.edu/cdhit-web-server/cgi-bin/index.cgi)) was used with three subsequent filter stages at pairwise identity cutoffs of 0.9, 0.5 and 0.1, resulting in a total of **3,688** sequences. # Citation If you use this dataset, please cite our paper: ``` @article { author = {Geffen, Yaron and Ofran, Yanay and Unger, Ron}, title = {DistilProtBert: A distilled protein language model used to distinguish between real proteins and their randomly shuffled counterparts}, year = {2022}, doi = {10.1093/bioinformatics/btac474}, URL = {https://doi.org/10.1093/bioinformatics/btac474}, journal = {Bioinformatics} } ```
yarongef/human_proteome_triplets
[ "license:mit", "region:us" ]
2022-04-12T09:44:30+00:00
{"license": "mit"}
2022-09-21T07:44:27+00:00
4a4b251e2258a5d44e0b258c1ea7b026d4e7147e
### Dataset Summary This dataset is extracted from Climate Fever dataset (https://www.sustainablefinance.uzh.ch/en/research/climate-fever.html), pre-processed and ready to train and evaluate. The training objective is a text classification task - given a claim and evidence, predict if evidence is related to claim.
mwong/climate-evidence-related
[ "task_categories:text-classification", "task_ids:fact-checking", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:extended|climate_fever", "language:en", "license:cc-by-sa-3.0", "license:gpl-3.0", "region:us" ]
2022-04-12T09:58:49+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["cc-by-sa-3.0", "gpl-3.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["extended|climate_fever"], "task_categories": ["text-classification"], "task_ids": ["fact-checking"], "paperswithcode_id": "climate-fever", "pretty_name": "climate-fever"}
2022-10-25T09:06:54+00:00
15e4d1933a321088880215257aa25b621a091335
This dataset is the subset of original eli5 dataset available on hugging face
Pavithree/eli5
[ "region:us" ]
2022-04-12T11:31:47+00:00
{}
2022-04-23T07:38:44+00:00
785573d6c226c7c07a7dc67f3e8739c25dad3927
enimai/MuST-C-and-WMT16-de-en
[ "license:afl-3.0", "region:us" ]
2022-04-12T12:48:41+00:00
{"license": "afl-3.0"}
2022-04-12T19:03:25+00:00
ca047290468c7f565f248c16139d2a096230112b
# Dataset Card for Snacks (Detection) ## Dataset Summary This is a dataset of 20 different types of snack foods that accompanies the book [Machine Learning by Tutorials](https://www.raywenderlich.com/books/machine-learning-by-tutorials/v2.0). The images were taken from the [Google Open Images dataset](https://storage.googleapis.com/openimages/web/index.html), release 2017_11. ## Dataset Structure Included in the **data** folder are three CSV files with bounding box annotations for the images in the dataset, although not all images have annotations and some images have multiple annotations. The columns in the CSV files are: - `image_id`: the filename of the image without the .jpg extension - `x_min, x_max, y_min, y_max`: normalized bounding box coordinates, i.e. in the range [0, 1] - `class_name`: the class that belongs to the bounding box - `folder`: the class that belongs to the image as a whole, which is also the name of the folder that contains the image The class names are: ```nohighlight apple banana cake candy carrot cookie doughnut grape hot dog ice cream juice muffin orange pineapple popcorn pretzel salad strawberry waffle watermelon ``` **Note:** The image files are not part of this repo but [can be found here](https://huggingface.co/datasets/Matthijs/snacks). ### Data Splits Train, Test, Validation ## Licensing Information Just like the images from Google Open Images, the snacks dataset is licensed under the terms of the Creative Commons license. The images are listed as having a [CC BY 2.0](https://creativecommons.org/licenses/by/2.0/) license. The annotations are licensed by Google Inc. under a [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) license.
Matthijs/snacks-detection
[ "task_categories:object-detection", "license:cc-by-4.0", "region:us" ]
2022-04-12T13:17:41+00:00
{"license": "cc-by-4.0", "task_categories": ["object-detection", "computer-vision"], "pretty_name": "Snacks (Detection)"}
2022-04-12T13:26:04+00:00
d3d4c6fc780dc8d0de62aeef28a38ba2fbed3606
| images | semantic maps | instance ids | | --- | --- | --- | | available | available | available | ``` dataset-size: 107Mb resolution: 1024x1024 license: ... sample-size: ./pix2pixHD_person_synthesis ├── test_img [10 entries] ├── test_inst [10 entries] ├── test_label [10 entries] ├── train_img [160 entries] ├── train_inst [160 entries] └── train_label [160 entries] ```
arakesh/PennFudanPedestrian-1024x512
[ "region:us" ]
2022-04-12T14:58:09+00:00
{}
2022-04-12T15:14:33+00:00
b46b232f33c35877e75081f948a89357ea4f1016
Data source: http://deepglobe.org/ | images | semantic maps | instance ids | | --- | --- | --- | | available | available | n/a | ``` dataset-size: 2.0G resolution: 2448x2448 license: ... sample-size: ./pix2pixHD-deepglobe-synthesis ├── test_img [30 entries] ├── test_label [30 entries] ├── train_img [773 entries] └── train_label [773 entries] ```
arakesh/deepglobe-2448x2448
[ "region:us" ]
2022-04-12T15:20:33+00:00
{}
2022-04-12T16:20:26+00:00
22c6b6002b2100e073b4146f20c22fa534d6e536
csteinmetz1/test-dataset
[ "license:cc", "region:us" ]
2022-04-12T15:47:54+00:00
{"license": "cc"}
2022-04-12T15:47:54+00:00
bdf451283a48e53f34fea37a8ad1c475175308cf
Data source: https://uavid.nl/ | images | semantic maps | instance ids | | --- | --- | --- | | available | available | n/a | ``` dataset-size: 6.1G resolution: mixed (3840x2160, 4096x2060) - because drone cameras are different for different faces. license: ... sample-size: + train: 200 + test: 70 ```
arakesh/uavid-15-hq-mixedres
[ "region:us" ]
2022-04-12T16:04:13+00:00
{}
2022-04-12T16:19:47+00:00
e5981976381334c87f5917b4743d749726a9e21b
## Problem and Opportunity In the United States, voting is largely a private matter. A registered voter is given a randomized ballot form or machine to prevent linkage between their voting choices and their identity. This disconnect supports confidence in the election process, but it provides obstacles to an election's analysis. A common solution is to field exit polls, interviewing voters immediately after leaving their polling location. This method is rife with bias, however, and functionally limited in direct demographics data collected. For the 2020 general election, though, most states published their election results for each voting location. These publications were additionally supported by the geographical areas assigned to each location, the voting precincts. As a result, geographic processing can now be applied to project precinct election results onto Census block groups. While precinct have few demographic traits directly, their geographies have characteristics that make them projectable onto U.S. Census geographies. Both state voting precincts and U.S. Census block groups: * are exclusive, and do not overlap * are adjacent, fully covering their corresponding state and potentially county * have roughly the same size in area, population and voter presence Analytically, a projection of local demographics does not allow conclusions about voters themselves. However, the dataset does allow statements related to the geographies that yield voting behavior. One could say, for example, that an area dominated by a particular voting pattern would have mean traits of age, race, income or household structure. The dataset that results from this programming provides voting results allocated by Census block groups. The block group identifier can be joined to Census Decennial and American Community Survey demographic estimates.
openenvironments/blockgroupvoting
[ "license:mit", "region:us" ]
2022-04-12T20:15:18+00:00
{"license": "mit"}
2022-04-12T20:19:08+00:00
58bafe5544a7b2eb3303c7d2ff3ea282a6446ba4
# Dataset Card for HatemojiCheck ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Content Warning This datasets contains examples of hateful language. ## Dataset Description and Details - **Repository:** https://github.com/HannahKirk/Hatemoji - **Paper:** https://arxiv.org/abs/2108.05921 - **Point of Contact:** hannah.kirk@oii.ox.ac.uk ### Dataset Summary HatemojiCheck is a test suite of 3,930 test cases covering seven functionalities of emoji-based hate and six identities. HatemojiCheck contains the text for each test case and its gold-standard label from majority agreement of three annotators. We provide labels by target of hate. HatemojiCheck can be used to evaluate the robustness of hate speech classifiers to constructions of emoji-based hate. ### Supported Tasks Hate Speech Detection ### Languages English ## Dataset Structure ### Data Instances 3,930 test cases ### Data Fields case_id: The unique ID of the test case (assigned to each of the 3,930 cases generated) templ_id: The unique ID of the template (original=.0, identity perturbation=.1, polarity perturbation=.2, emoji perturbation = .3) from which the test case was generated test_grp_id: The ID of the set of templates (original, identity perturbation, polarity perturbation, no emoji perturbation) from which the test case was generated. text: The text of the test case. target: Where applicable, the protected group targeted or referenced by the test case. We cover six protected groups in the test suite: women, trans people, gay people, black people, disabled people and Muslims. functionality: The shorthand for the functionality tested by the test case. set: Whether the test case is an original statement, a identity perturbation, a polarity perturbation or a no emoji perturbation. label_gold: The gold standard label (hateful/non-hateful) of the test case. All test cases within a given functionality have the same gold standard label. unrealistic_flags: The number of annotators (/3) who flagged the test case as unrealistic. included_in_test_suite: Indicator for whether test case is included in final HatemojiCheck test suite. All 3,930 test cases are included. ### Data Splits All of HatemojiCheck is designated for testing models so only test is provided. ## Dataset Creation ### Curation Rationale The purpose of HatemojiCheck is to evaluate the performance of black-box models against varied constructions of emoji-based hate. To construct HatemojiCheck, we hand-crafted 3,930 short form English-language texts using a template-based method for group identities and slurs. Each test case exemplifies one functionality and is associated with a binary gold standard label _hateful_ versus _not hateful_. All 3,930 cases were labeled by a trained team of three annotators, who could also flag examples that were unrealistic. Any test cases with multiple disagreements or flags were replaced with alternative templates and re-issued for annotation to improve the quality of examples in the final set of test cases. ### Source Data #### Initial Data Collection and Normalization Based on the literature, we define a list of potentially hateful emoji and words, and use Twitter's Streaming API to search for the Cartesian products of emoji--emoji and emoji--word pairs over a two week period. To identify different forms of emoji-based hate, we apply a grounded theory approach on a sample of 3,295 tweets, splitting out distinctive categories, and recursively selecting sub-categories until all key parts of the data are captured and the framework is `saturated'. #### Who are the source language producers? All test cases were hand-crafted by the lead author, who is a native English-speaking researcher at a UK university with extensive subject matter expertise in online harms. The test cases are in English. This choice was motivated by the researchers' and annotators' expertise, and to maximize HatemojiCheck's applicability to previous hate speech detection studies, which are predominantly conducted on English-language data. We discuss the limitations of restricting HatemojiCheck to one language and suggest that future work should prioritize expanding the test suite to other languages. ### Annotations #### Annotation process To validate the gold-standard labels assigned to each test case, we recruited three annotators with prior experience on hate speech projects. Annotators were given extensive guidelines, test tasks and training sessions, which included examining real-world examples of emoji-based hate from Twitter. We followed guidance for protecting annotator well-being. There were two iterative rounds of annotation. In the first round, each annotator labeled all 3,930 test cases as hateful or non-hateful, and had the option to flag unrealistic entries. Test cases with any disagreement or unrealistic flags were reviewed by the study authors (n=289). One-on-one interviews were conducted with annotators to identify dataset issues versus annotator error. From 289 test cases, 119 were identified as ambiguous or unrealistic, replaced with alternatives and re-issued to annotators for labeling. No further issues were raised. We measured inter-annotator agreement using Randolph's Kappa, obtaining a value of 0.85 for the final set of test cases, which indicates "almost perfect agreement". #### Who are the annotators? We recruited a team of three annotators who worked for two weeks in May 2021 and were paid £16. All annotators were female and between 30--39 years old. One had an undergraduate degree, one a taught graduate degree and one a post-graduate research degree. There were three nationalities: Argentinian, British and Iraqi, two ethnicities: White and Arab, and three religious affiliations: Catholic, Muslim and None. One annotator was a native English speaker and the others were non-native but fluent. All annotators used emoji and social media more than once per day. All annotators had seen others targeted by abuse online, and one had been targeted personally. ### Personal and Sensitive Information HatemojiCheck contains synthetic statements so has no personal information. It does however contains harmful examples of emoji-based hate which could be disturbing or damaging to view. ## Considerations for Using the Data ### Social Impact of Dataset HatemojiCheck contains challenging emoji examples on which commercial solutions and state-of-the-art transformer models have been proven to fail. Malicious actors could take inspiration for bypassing current detection systems on internet platforms, or in principal train a generative hate speech model. However, it also helps to evaluate model's weaknesses to emoji-based hate, so can be used to mitigate the harm to victims before a model is deployed. ### Discussion of Biases HatemojiCheck only contains test cases against 6 identities: woman, trans people, gay people, disabled people, Black people and Muslims. It thus is biased towards evaluating hate directed at these targets. Additionally, HatemojiCheck was motivated by an empirical study of English-language tweets. The usage of emoji varies significantly across culture, country and demographic so there may be biases towards Western, English-language use of emoji. ### Other Known Limitations While inspired by real-world instances of emoji-based hate, HatemojiCheck contains synthetic, hand-crafted test cases. These test cases are designed to be a "minimum performance standard" against which to hold models accountable. However, because the test cases are designed to have one "clear, gold-standard label" they may be easier to predict than more nuanced, complex and real-world instances of emoji-based hate. ## Additional Information ### Dataset Curators The dataset was created by the lead author (Hannah Rose Kirk), then validated by the other authors and three annotators. ### Licensing Information Creative Commons Attribution 4.0 International Public License. For full detail see: https://github.com/HannahKirk/Hatemoji/blob/main/LICENSE ### Citation Information If you use this dataset, please cite our paper: Kirk, H. R., Vidgen, B., Röttger, P., Thrush, T., & Hale, S. A. (2021). Hatemoji: A test suite and adversarially-generated dataset for benchmarking and detecting emoji-based hate. arXiv preprint arXiv:2108.05921. ``` @article{kirk2021hatemoji, title={Hatemoji: A test suite and adversarially-generated dataset for benchmarking and detecting emoji-based hate}, author={Kirk, Hannah Rose and Vidgen, Bertram and R{\"o}ttger, Paul and Thrush, Tristan and Hale, Scott A}, journal={arXiv preprint arXiv:2108.05921}, year={2021} } ``` ### Contributions Thanks to [@HannahKirk](https://github.com/HannahKirk) for adding this dataset.
HannahRoseKirk/HatemojiCheck
[ "task_categories:text-classification", "task_ids:hate-speech-detection", "annotations_creators:expert", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "license:cc-by-4.0", "arxiv:2108.05921", "region:us" ]
2022-04-13T07:35:38+00:00
{"annotations_creators": ["expert"], "language_creators": ["expert-generated"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["hate-speech-detection"], "pretty_name": "HatemojiCheck", "languages": ["en"], "extra_gated_prompt": "We have deactivated the automatic preview for this dataset because it contains hate speech. If you want to see the preview, you can continue."}
2022-05-15T07:56:10+00:00
17509b6678c3206ce7bfa43107fdadf53f3d06f7
# Dataset Card for HatemojiBuild ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Content Warning This datasets contains examples of hateful language. ## Dataset Description and Details - **Repository:** https://github.com/HannahKirk/Hatemoji - **Paper:** https://arxiv.org/abs/2108.05921 - **Point of Contact:** hannah.kirk@oii.ox.ac.uk ### Dataset Summary HatemojiBuild can be used to train, develop and test models on emoji-based hate with challenging adversarial examples and perturbations. HatemojiBuild is a dataset of 5,912 adversarially-generated examples created on Dynabench using a human-and-model-in-the-loop approach. We collect data in three consecutive rounds. Our work follows on from Vidgen et al (2021) _Learning from the Worst: Dynamically Generated Datasets to Improve Online Hate Detection_ (http://arxiv.org/abs/2012.15761) who collect four rounds of textual adversarial examples. The R1-R4 data is available at https://github.com/bvidgen/Dynamically-Generated-Hate-Speech-Dataset. The entries in HatemojiBuild are labeled by round (R5-7). The text of each entry is given with its gold-standard label from majority agreement of three annotators. Each original entry is associated with a perturbation so each row of the dataset. matches these two cases. We also provide granular labels of type and target for hateful entries. ### Supported Tasks Hate Speech Detection ### Languages English ## Dataset Structure ### Data Instances 5,912 adversarially-generated instances ### Data Fields entry_id: The unique ID of the entry (assigned to each of the 5,912 cases generated). text: The text of the entry. type: The type of hate assigned to hateful entries. target: The target of hate assigned to hateful entries. round.base: The round where the entry was generated. round.set: The round and whether the entry came from an original statement (a) or a perturbation (b). set: Whether the entry is an original or perturbation. split: The randomly-assigned train/dev/test split using in our work (80:10:10). label_gold: The gold standard label (hateful/non-hateful) of the test case. matched_text: The text of the paired perturbation. Each original entry has one perturbation. matched_id: The unique entry ID of the paired perturbation. ### Data Splits Train, Validation and Test. ## Dataset Creation ### Curation Rationale The genre of texts is hateful and non-hateful statements using emoji constructions. The purpose of HatemojiBuild is address the model weaknesses to emoji-baaed hate, to "build" better models. 50% of the 5,912 test cases are hateful. 50% of the entries in the dataset are original content and 50% are perturbations. ### Source Data #### Initial Data Collection and Normalization We use an online interface designed for dynamic dataset generation and model benchmarking (Dynabench) to collect synthetic adversarial examples in three successive rounds, running between 24th May--11th June. Each round contains approximately 2,000 entries, where each original entry inputed to the interface is paired with an offline perturbation. Data was synthetically-generated by a team of trained annotators, i.e., not sampled from social media. #### Who are the source language producers? The language producers are also the annotators. ### Annotations #### Annotation process We implemented three successive rounds of data generation and model re-training to create the HatemojiBuild dataset. In each round we tasked a team of 10 trained annotators with entering content the model-in-the-loop would misclassify. We refer to this model as the target model. Annotators were instructed to generate linguistically diverse entries while ensuring each entry was (1) realistic, (2) clearly hateful or non-hateful and (3) contained at least one emoji. Each entry was first given a binary label of hateful or non-hateful, and hateful content was assigned secondary labels for the type and target of hate. Each entry was validated by two additional annotators, and an expert resolved disagreements. After validation, annotators created a perturbation for each entry that flips the label. To maximize similarity between originals and perturbations, annotators could either make an emoji substitution while fixing the text or fix the emoji and minimally change the surrounding text. Each perturbation received two additional annotations, and disagreements were resolved by the expert. This weekly cadence of annotator tasks was repeated in three consecutive weeks. #### Who are the annotators? Ten annotators were recruited to work for three weeks, and paid £16/hour. An expert annotator was recruited for quality control purposes and paid £20/hour. In total, there were 11 annotators. All annotators received a training session prior to data collection and had previous experience working on hate speech projects. A daily `stand-up' meeting was held every morning to communicate feedback and update guidelines as rounds progressed. Annotators were able to contact the research team at any point using a messaging platform. Of 11 annotators, 8 were between 18--29 years old and 3 between 30--39 years old. The completed education level was high school for 3 annotators, undergraduate degree for 1 annotators, taught graduate degree for 4 annotators and post-graduate research degree for 3 annotators. 6 annotators were female, and 5 were male. Annotators came from a variety of nationalities, with 7 British, as well as Jordanian, Irish, Polish and Spanish. 7 annotators identified as ethnically White and the remaining annotators came from various ethnicities including Turkish, Middle Eastern, and Mixed White and South Asian. 4 annotators were Muslim, and others identified as Atheist or as having no religious affiliation. 9 annotators were native English speakers and 2 were non-native but fluent. The majority of annotators (9) used emoji and social media more than once per day. 10 annotators had seen others targeted by abuse online, and 7 had been personally targeted. ### Personal and Sensitive Information HatemojiBuild contains synthetic statements so has no personal information. It does however contains harmful examples of emoji-based hate which could be disturbing or damaging to view. ## Considerations for Using the Data ### Social Impact of Dataset HatemojiBuild contains challenging emoji examples which have "tricked" state-of-the-art transformers models. Malicious actors could take inspiration for bypassing current detection systems on internet platforms, or in principal train a generative hate speech model. However, it also helps to build model robustness to emoji-based hate, so can be used to mitigate the harm to victims before a model is deployed. ### Discussion of Biases Annotators were given substantial freedom in the targets of hate resulting in 54 unique targets, and 126 unique intersections of these. The entries from R5-R7 contain 1,082 unique emoji out of 3,521 defined in the Unicode Standard as of September 2020. This diversity helped to mitigate biases in classification towards certain targets but biases likely remain, especially since HatemojiBuild was designed for English-language use of emoji. ### Other Known Limitations While annotators were trained on real-world examples of emoji-based hate from Twitter, the entries in HatemojiBuild are synthetically-generated so may deviate from real-world instances of emoji-based hate. ## Additional Information ### Dataset Curators The dataset was curated by the lead author (Hannah Rose Kirk), using the Dynabench platform. ### Licensing Information Creative Commons Attribution 4.0 International Public License. For full detail see: https://github.com/HannahKirk/Hatemoji/blob/main/LICENSE ### Citation Information If you use this dataset, please cite our paper: Kirk, H. R., Vidgen, B., Röttger, P., Thrush, T., & Hale, S. A. (2021). Hatemoji: A test suite and adversarially-generated dataset for benchmarking and detecting emoji-based hate. arXiv preprint arXiv:2108.05921. ``` @article{kirk2021hatemoji, title={Hatemoji: A test suite and adversarially-generated dataset for benchmarking and detecting emoji-based hate}, author={Kirk, Hannah Rose and Vidgen, Bertram and R{\"o}ttger, Paul and Thrush, Tristan and Hale, Scott A}, journal={arXiv preprint arXiv:2108.05921}, year={2021} } ``` ### Contributions Thanks to [@HannahKirk](https://github.com/HannahKirk) for adding this dataset.
HannahRoseKirk/HatemojiBuild
[ "task_categories:text-classification", "task_ids:hate-speech-detection", "annotations_creators:expert", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "license:cc-by-4.0", "arxiv:2108.05921", "arxiv:2012.15761", "region:us" ]
2022-04-13T08:12:14+00:00
{"annotations_creators": ["expert"], "language_creators": ["expert-generated"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["hate-speech-detection"], "pretty_name": "HatemojiBuild", "languages": ["en"], "extra_gated_prompt": "We have deactivated the automatic preview for this dataset because it contains hate speech. If you want to see the preview, you can continue."}
2022-05-15T07:56:35+00:00
37d117aedb1c469ebf2adc217dae40ff31a97a23
# hotpotQA-Extended (Annotated) A version of [HotpotQA-Extended](https://huggingface.co/datasets/ghomasHudson/hotpotExtended) with extra annotations about what part of the input contains the answer.
ghomasHudson/hotpotExtendedAno
[ "region:us" ]
2022-04-13T09:55:51+00:00
{}
2022-04-13T10:01:17+00:00
3cdedf844922ab40393d46d4c7f81c596e1c6d45
This a subset of "ceyda/smithsonian_butterflies" dataset with additional processing done to train the "ceyda/butterfly_gan" model. The preprocessing includes: - Adding "sim_score" to images with CLIP model using "pretty butterfly","one butterfly","butterfly with open wings","colorful butterfly" - Removing butterflies with the same name(species) - Limiting only to the top 1000 images - Removing the background (doing another sim_scoring after bg removal did visually worse so didn't do it) - Detecting contours - Cropping to the bounding box of the contour with the largest area - Converting back to RGB
huggan/smithsonian_butterflies_subset
[ "region:us" ]
2022-04-13T10:36:00+00:00
{}
2022-04-16T07:02:36+00:00
ec73b665e4be0f567b69d39425355401cfe0d29b
SetFit/amazon_reviews_multi_en
[ "license:apache-2.0", "region:us" ]
2022-04-13T17:39:28+00:00
{"license": "apache-2.0"}
2022-04-13T18:06:11+00:00
77b43264a1186dfd3ffdc2bd4018b41f0a6bb689
enimai/MuST-C-it
[ "license:afl-3.0", "region:us" ]
2022-04-14T03:55:01+00:00
{"license": "afl-3.0"}
2022-04-14T03:57:08+00:00
161fc9fce16d7e942e0ed14046c1e98956437061
[Needs More Information] # Dataset Card for squad_it ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** [Needs More Information] - **Repository:** [Needs More Information] - **Paper:** [Needs More Information] - **Leaderboard:** [Needs More Information] - **Point of Contact:** [Needs More Information] ### Dataset Summary Converted dataset version to be used in Huggingface. Originally created by Croce et al. at 2018, the SQuAD-it The dataset contains more than 60,000 question/answer pairs in Italian derived from the original English SQuAD dataset., in Italian language. Containing 60,000+ in JSON file format. ### Supported Tasks and Leaderboards [Needs More Information] ### Languages [Needs More Information] ## Dataset Structure ### Data Instances [Needs More Information] ### Data Fields [Needs More Information] ### Data Splits [Needs More Information] ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information [Needs More Information] ### Citation Information ``` @InProceedings{10.1007/978-3-030-03840-3_29, author="Croce, Danilo and Zelenanska, Alexandra and Basili, Roberto", editor="Ghidini, Chiara and Magnini, Bernardo and Passerini, Andrea and Traverso, Paolo", title="Neural Learning for Question Answering in Italian", booktitle="AI*IA 2018 -- Advances in Artificial Intelligence", year="2018", publisher="Springer International Publishing", address="Cham", pages="389--402", isbn="978-3-030-03840-3" } ```
bullmount/squad_it
[ "region:us" ]
2022-04-14T05:58:53+00:00
{}
2022-04-14T15:06:54+00:00
d40a6b60af1f6d7f320307c3362e8122e8b72006
Aanisha/NeonGAN_dataset
[ "license:mit", "region:us" ]
2022-04-14T06:57:23+00:00
{"license": "mit"}
2022-04-14T06:57:23+00:00
4b51f4dae6b6ad746445f05059d6793c1f6ea988
# KP20k Benchmark Dataset for Keyphrase Generation ## About KP20k is a dataset for benchmarking keyphrase extraction and generation models. The data is composed of 570 809 abstracts and their associated titles from scientific articles. Details about the dataset can be found in the original paper: - Meng et al 2017. [Deep keyphrase Generation](https://aclanthology.org/P17-1054.pdf) Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, pages 582–592 Reference (indexer-assigned) keyphrases are also categorized under the PRMU (<u>P</u>resent-<u>R</u>eordered-<u>M</u>ixed-<u>U</u>nseen) scheme as proposed in the following paper: - Florian Boudin and Ygor Gallina. 2021. [Redefining Absent Keyphrases and their Effect on Retrieval Effectiveness](https://aclanthology.org/2021.naacl-main.330/). In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 4185–4193, Online. Association for Computational Linguistics. Text pre-processing (tokenization) is carried out using spacy (en_core_web_sm model) with a special rule to avoid splitting words with hyphens (e.g. graph-based is kept as one token). Stemming (Porter's stemmer implementation provided in nltk) is applied before reference keyphrases are matched against the source text. ## Content The dataset is divided into the following three splits: | Split | # documents | # keyphrases by document (average) | % Present | % Reordered | % Mixed | % Unseen | | :--------- | ----------: | -----------: | --------: | ----------: | ------: | -------: | | Train | 530 809 | 5.29 | 58.19 | 10.93 | 17.36 | 13.52 | | Test | 20 000 | 5.28 | 58.40 | 10.84 | 17.20 | 13.56 | | Validation | 20 000 | 5.27 | 58.20 | 10.94 | 17.26 | 13.61 | The following data fields are available: - **id**: unique identifier of the document. **NB** There were no ids in the original dataset. The ids were generated using the python module shortuuid (https://pypi.org/project/shortuuid/) - **title**: title of the document. - **abstract**: abstract of the document. - **keyphrases**: list of the author assigned keyphrases. - **prmu**: list of <u>P</u>resent-<u>R</u>eordered-<u>M</u>ixed-<u>U</u>nseen categories for reference keyphrases. **NB**: The present keyphrases (represented by the "P" label in the PRMU column) are sorted by their apparition order in the text (title + abstract).
taln-ls2n/kp20k
[ "task_categories:text-generation", "annotations_creators:unknown", "language_creators:unknown", "multilinguality:monolingual", "size_categories:100K<n<1M", "language:en", "license:unknown", "keyphrase-generation", "keyphrase-extraction", "text-mining", "region:us" ]
2022-04-14T08:00:02+00:00
{"annotations_creators": ["unknown"], "language_creators": ["unknown"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "task_categories": ["text-generation"], "task_ids": [], "pretty_name": "KP20k", "tags": ["keyphrase-generation", "keyphrase-extraction", "text-mining"]}
2023-09-13T12:15:04+00:00
5eb9bca5c7dc850b2a42df268b78b88190ab2466
# PET: A NEW DATASET FOR PROCESS EXTRACTION FROM TEXT # Dataset Card for PET ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) - [Annotation Guidelines](#annotationguidelines) - [Update](#updates) - [Loading data](#loadingdata) ## Dataset Description - **Homepage:** https://pdi.fbk.eu/pet-dataset/ - **Paper:** https://arxiv.org/abs/2203.04860 - **Point of Contact:** [Patrizio Bellan](pbellan@fbk.eu) ### Dataset Summary Abstract. Although there is a long tradition of work in NLP on extracting entities and relations from text, to date there exists little work on the acquisition of business processes from unstructured data such as textual corpora of process descriptions. With this work we aim at filling this gap and establishing the first steps towards bridging data-driven information extraction methodologies from Natural Language Processing and the model-based formalization that is aimed from Business Process Management. For this, we develop the first corpus of business process descriptions annotated with activities, actors, activity data, gateways and their conditions. We present our new resource to benchmark the difficulty and challenges of business process extraction from text. ### Supported Tasks and Leaderboards - Token Classification - Named Entity Recognition - Relations Extraction ### Languages English ## Dataset Structure Test set to beanchmark *Business Process Extraction from Text* approaches. ### Data Instances #### Token Classification For each instance, there is a document name representing the name of the document of the Friedrich *et al.* dataset, an integer representing the number of the sentence, a list of tokens representing the words of the sentence instance, and a list of *ner tags* (in IOB2 format) representing the annotation of process elements of the sentence. Below, an example of data instance. ``` { "document name":"doc-1.1", "sentence-ID":1, "tokens":["Whenever","the","sales","department","receives","an","order",",","a","new","process","instance","is","created","."], "ner-tags":["O","B-Actor","I-Actor","I-Actor","B-Activity","B-Activity Data","I-Activity Data","O","O","O","O","O","O","O","O"] } ``` #### Relations Extraction For each instance, there is a document name representing the name of the document of the Friedrich *et al.* dataset, a list of tokens representing the words of the document instance, a list of interger representing the words position within each sentence of the document instance, a list of *ner tags* (in IOB2 format) representing the annotation of the token, a list of sentence id representing for each token the number of the sentence, and a list of relations of the document. Below, an example of data instance. ``` { "document name": "doc-1.1", "tokens": ["A", "small", "company",...], "tokens-IDs": [0, 1, 2, ...], "ner_tags": ["O", "O", "O", ...], "sentence-IDs": [0, 0, 0, ...], "relations": { "source-head-sentence-ID": [1, 1, 1, ...], "source-head-word-ID": [4, 4, 4, ...], "relation-type": ["uses", "flow", "actor recipient", ...], "target-head-sentence-ID": [1, 2, 1,...], "target-head-word-ID": [5, 9, 1, ...] } } ``` ### Data Fields #### Token Classification - *document name*: a string used to represent the name of the document. - *sentence-ID*: an integer (starting from 0) representing the number of the sentence within the document. - *tokens*: a list of string representing the words of the sentence - *ner-tags*: a list of string representing the annotation for each word. The allowed **ner-tags** are: - **O**: An O tag indicates that a token belongs to no chunk. - **B-Actor**: This tag indicates the beginning of an *Actor* chunk. - **I-Actor**: This tag indicates that the tag is inside an *Actor* chunk. - **B-Activity**: This tag indicates the beginning of an *Activity* chunk. - **I-Activity**: This tag indicates that the tag is inside an *Activity* chunk. - **B-Activity Data**: This tag indicates the beginning of an *Activity Data* chunk. - **I-Activity Data**: This tag indicates that the tag is inside an *Activity Data* chunk. - **B-Further Specification**: This tag indicates the beginning of a *Further Specification* chunk. - **I-Further Specification**: This tag indicates that the tag is inside a *Further Specification* chunk. - **B-XOR Gateway**: This tag indicates the beginning of a *XOR Gateway* chunk. - **I-XOR Gateway**: This tag indicates that the tag is inside a *XOR Gateway* chunk. - **B-Condition Specification**: This tag indicates the beginning of a *Condition Specification* chunk. - **I-Condition Specification**: This tag indicates that the tag is inside a *Condition Specification* chunk. - **B-AND Gateway**: This tag indicates the beginning of an *AND Gateway* chunk. - **I-AND Gateway**: This tag indicates that the tag is inside an *AND Gateway* chunk. To have a complete explanation of each process element tag please refer to the [research paper](https://arxiv.org/abs/2203.04860) and the [annotation guidelines](https://pdi.fbk.eu/pet/annotation-guidelines-for-process-description.pdf). ### Relations Extraction - *document name*: a string used to represent the name of the document. - *tokens*: a list of string representing the words of the document - *tokens-IDs*: a list of interger representing the word position within a sentence. - *ner_tags*: a list of string representing the annotation for each word. (see ner-tags above) - *sentence-IDs*: a list of interger representing the sentence number for each word of the document. - *relations*:: a list of document relations. - *source-head-sentence-ID*: a list of sentence ID pointing to the sentence number of the head (first token) of the source entity. - *source-head-word-ID*: a list of token ID pointing to the word ID of the head (first token) of the source entity. - *relation-type*: a list of relation tags. - *target-head-sentence-ID*: a list of sentence ID pointing to the sentence number of the head (first token) of the target entity. - *target-head-word-ID*: a list of token ID pointing to the word ID of the head (first token) of the target entity. For instance, a relation is defined by the instances of *source-head-sentence-ID*, *source-head-word-ID*, *relation-type*, *target-head-sentence-ID*, and *target-head-word-ID* at the same index position. In the following example, the first relation of the first document is shown: ```python document_1=modelhub_dataset['test'][0] relation = { 'source-head-sentence-ID': document_1['relations']['source-head-sentence-ID'][0], 'source-head-word-ID': document_1['relations']['source-head-word-ID'][0], 'relation-type': document_1['relations']['relation-type'][0], 'target-head-sentence-ID': document_1['relations']['target-head-sentence-ID'][0], 'target-head-word-ID': document_1['relations']['target-head-sentence-ID'][0], } print(relation) ``` the output is: ```python {'relation-type': 'uses', 'source-head-sentence-ID': 1, 'source-head-word-ID': 4, 'target-head-sentence-ID': 1, 'target-head-word-ID': 1} ``` That means: the entity in sentence number *1*, starting at the token position *4* has a *uses* relation with the entity in sentence number *1* starting at token position *1* ### Data Splits The data was not split. It contains the test set only. ## Dataset Creation ### Curation Rationale Although there is a long tradition of work in NLP on extracting entities and relations from text to date there exists little work on the acquisition of business processes from unstructured data such as textual corpora of process descriptions. With this work we aim at filling this gap and establishing the first steps towards bridging data-driven information extraction methodologies from Natural Language Processing and the model-based formalization that is aimed from Business Process Management. ### Source Data #### Initial Data Collection and Normalization The dataset construction process has been split in five main phases: 1. Text pre-processing. As the first operation, we check the content of each document and we tokenized it. This initial check was necessary since some of the original texts were automatically translated into English by the authors of the dataset. The translations were never validated, indeed, several errors have been found and fixed. 2. Text Annotation. Each text has been annotated by using the [guidelines](https://pdi.fbk.eu/pet/annotation-guidelines-for-process-description.pdf). The team was composed by five annotators with high expertise in BPMN. Each document has been assigned to three experts that were in change of identifying all the elements and flows with each document. In this phase, we used the the Inception tool to support annotators. 3. Automatic annotation fixing. After the second phase, we ran an automatic procedure relying on a rule-based script to automatically fix annotations that were not compliant with the guidelines. For example, if a modal verb was erroneously included in the annotation of an Activity, the procedure removed it from the annotation. Another example is the missing of the article within an annotation related to an Actor. In this case, the script included it in the annotation. This phase allowed to remove possible annotation errors and to obtain annotations compliant with the guidelines. 4. Agreement Computation. Here, we computed, on the annotation provided by the experts, the agreement scores for each process element and for each relation between process elements pair adopting the methodology proposed in [Hripcsak *et al.*](https://academic.oup.com/jamia/article/12/3/296/812057?login=true). We measured the agreement in terms of the F1 measure because, besides being straightforward to calculate, it is directly interpretable. Note that chance-corrected measures like *k* approach the F1-measure as the number of cases that raters agree are negative grows. By following such a methodology, an annotation was considered in agreement among the experts if and only if they capture the same span of words and they assign the same process element tag to the annotation. 5. Reconciliation. The last phase consisted of the mitigation of disagreements within the annotations provided by the experts. The aim of this phase is to obtain a shared and agreed set of gold standard annotations on each text for both entities and relations. Such entities also enable the generation of the related full-connected process model flow that can be rendered by using, but not limited to, a BPMN diagram. During this last phase, among the 47 documents originally included into the dataset, 2 of them were discarded. These texts were not fully annotated by the annotators since they were not be able to completely understand which process elements were actually included in some specific parts of the text. For this reason, the final size of the dataset is 45 textual descriptions of the corresponding process models together with their annotations. #### Who are the source language producers? English ### Annotations #### Annotation process You can read about the annotation process in the original paper https://arxiv.org/abs/2203.04860 #### Who are the annotators? Expert Annotators ### Personal and Sensitive Information No personal or sensitive information issues. ## Considerations for Using the Data ### Social Impact of Dataset The dataset has no social impact ### Discussion of Biases No bias found in the dataset ### Other Known Limitations The *Further specification* and *AND Gateway* elements obtained very poor performance on the baselines proposed in the paper. The *AND Gateway* is the less represented process elements in this dataset. The *Further Specification* process element was the most difficult element to annotate. ## Additional Information ### Dataset Curators - Patrizio Bellan (Fondazione Bruno Kessler, Trento, Italy and Free University of Bozen-Bolzano, Bolzano, Italy) - Mauro Dragoni (Fondazione Bruno Kessler, Trento, Italy) - Chiara Ghidini (Fondazione Bruno Kessler, Trento, Italy) - Han van der Aa (University of Mannheim, Mannheim, Germany) - Simone Ponzetto (University of Mannheim, Mannheim, Germany) ### Licensing Information ### Citation Information ``` @inproceedings{DBLP:conf/aiia/BellanGDPA22, author = {Patrizio Bellan and Chiara Ghidini and Mauro Dragoni and Simone Paolo Ponzetto and Han van der Aa}, editor = {Debora Nozza and Lucia C. Passaro and Marco Polignano}, title = {Process Extraction from Natural Language Text: the {PET} Dataset and Annotation Guidelines}, booktitle = {Proceedings of the Sixth Workshop on Natural Language for Artificial Intelligence {(NL4AI} 2022) co-located with 21th International Conference of the Italian Association for Artificial Intelligence (AI*IA 2022), Udine, November 30th, 2022}, series = {{CEUR} Workshop Proceedings}, volume = {3287}, pages = {177--191}, publisher = {CEUR-WS.org}, year = {2022}, url = {https://ceur-ws.org/Vol-3287/paper18.pdf}, timestamp = {Fri, 10 Mar 2023 16:23:01 +0100}, biburl = {https://dblp.org/rec/conf/aiia/BellanGDPA22.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } @inproceedings{DBLP:conf/bpm/BellanADGP22, author = {Patrizio Bellan and Han van der Aa and Mauro Dragoni and Chiara Ghidini and Simone Paolo Ponzetto}, editor = {Cristina Cabanillas and Niels Frederik Garmann{-}Johnsen and Agnes Koschmider}, title = {{PET:} An Annotated Dataset for Process Extraction from Natural Language Text Tasks}, booktitle = {Business Process Management Workshops - {BPM} 2022 International Workshops, M{\"{u}}nster, Germany, September 11-16, 2022, Revised Selected Papers}, series = {Lecture Notes in Business Information Processing}, volume = {460}, pages = {315--321}, publisher = {Springer}, year = {2022}, url = {https://doi.org/10.1007/978-3-031-25383-6\_23}, doi = {10.1007/978-3-031-25383-6\_23}, timestamp = {Tue, 14 Feb 2023 09:47:10 +0100}, biburl = {https://dblp.org/rec/conf/bpm/BellanADGP22.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ``` ### Contributions Thanks to [Patrizio Bellan](https://pdi.fbk.eu/bellan/) for adding this dataset. #### <a name="updates"></a>Update - v1.0.0: Added token classification task - v1.0.1: Added extraction relation task - v1.1.0: Fixed minor errors, fixed performs relations Version 1.1.0 cab be found [here](https://huggingface.co/datasets/patriziobellan/PETv11) ## <a name="annotationguidelines"></a>Annotation Guidelines ### Inception Schema The inception schema can be found [here](https://pdi.fbk.eu/pet/inception-schema.json) ### Annotation Guidelines The Annotation guidelines and procedures adopted to annotate the PET dataset can be downloaded [here](https://pdi.fbk.eu/pet/annotation-guidelines-for-process-description.pdf) ### Article The article can be downloaded [here]({https://ceur-ws.org/Vol-3287/paper18.pdf}) ### Python Interface A Python interface (beta version) to interact with the dataset can be found [here](https://pypi.org/project/petdatasetreader/) You can find the **BASELINES**, the annotation data, and a graphical interface to visualize predictions [here](https://github.com/patriziobellan86/PETbaselines) ### Benchmarks A Python benchmarking procedure package to test approaches on the PET dataset ca be found [here](https://pypi.org/project/petbenchmarks/) ## <a name="loadingdata"></a>Loading data ### Token-classification task ```python from datasets import load_dataset modelhub_dataset = load_dataset("patriziobellan/PET", name='token-classification') ``` ### Relations-extraction task ```python from datasets import load_dataset modelhub_dataset = load_dataset("patriziobellan/PET", name='relations-extraction') ```
patriziobellan/PET
[ "task_categories:token-classification", "size_categories:n<1K", "language:en", "license:mit", "Business Process Management", "NLP", "ML", "DL", "arxiv:2203.04860", "region:us" ]
2022-04-14T08:35:11+00:00
{"language": ["en"], "license": "mit", "size_categories": ["n<1K"], "task_categories": ["token-classification"], "pretty_name": "PET", "tags": ["Business Process Management", "NLP", "ML", "DL"]}
2023-07-05T13:03:24+00:00
c4a3428883440ffabcba3afe9ed7ee94ffd13abb
# Dataset Card ## Disclaimer All rights belong to their owners. Models and datasets can be removed from the site at the request of the copyright holder. ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [How to use](#how-to-use) - [Dataset Structure](#dataset-structure) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [About](#about) ## Dataset Description - **Homepage:** [https://github.com/AlekseyKorshuk/huggingnft](https://github.com/AlekseyKorshuk/huggingnft) - **Repository:** [https://github.com/AlekseyKorshuk/huggingnft](https://github.com/AlekseyKorshuk/huggingnft) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Dataset Summary NFT images dataset for unconditional generation. NFT collection available [here](https://opensea.io/collection/hapeprime). Model is available [here](https://huggingface.co/huggingnft/hapeprime). Check Space: [link](https://huggingface.co/spaces/AlekseyKorshuk/huggingnft). ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## How to use How to load this dataset directly with the datasets library: ```python from datasets import load_dataset dataset = load_dataset("huggingnft/hapeprime") ``` ## Dataset Structure [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Data Fields The data fields are the same among all splits. - `image`: an `image` feature. - `id`: an `int` feature. - `token_metadata`: a `str` feature. - `image_original_url`: a `str` feature. ### Data Splits [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @InProceedings{huggingnft, author={Aleksey Korshuk} year=2022 } ``` ## About *Built by Aleksey Korshuk* [![Follow](https://img.shields.io/github/followers/AlekseyKorshuk?style=social)](https://github.com/AlekseyKorshuk) [![Follow](https://img.shields.io/twitter/follow/alekseykorshuk?style=social)](https://twitter.com/intent/follow?screen_name=alekseykorshuk) [![Follow](https://img.shields.io/badge/dynamic/json?color=blue&label=Telegram%20Channel&query=%24.result&url=https%3A%2F%2Fapi.telegram.org%2Fbot1929545866%3AAAFGhV-KKnegEcLiyYJxsc4zV6C-bdPEBtQ%2FgetChatMemberCount%3Fchat_id%3D-1001253621662&style=social&logo=telegram)](https://t.me/joinchat/_CQ04KjcJ-4yZTky) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/AlekseyKorshuk/huggingnft?style=social)](https://github.com/AlekseyKorshuk/huggingnft)
huggingnft/hapeprime
[ "license:mit", "huggingnft", "nft", "huggan", "gan", "image", "images", "region:us" ]
2022-04-14T10:40:21+00:00
{"license": "mit", "tags": ["huggingnft", "nft", "huggan", "gan", "image", "images"], "task": ["unconditional-image-generation"], "datasets": ["huggingnft/hapeprime"]}
2022-04-16T16:59:08+00:00
61faf004bd4a26daa27ec3127bc55e5c60829cbe
# Pages of Early Soviet Performance (PESP) This dataset was created as part of the [Pages of Early Soviet Performance](https://cdh.princeton.edu/projects/pages-early-soviet-performance/) project at Princeton and provides text and image research data from a previously scanned [collection of illustrated periodicals](https://dpul.princeton.edu/slavic/catalog?f%5Breadonly_collections_ssim%5D%5B%5D=Russian+Illustrated+Periodicals) held by Princeton University's Slavic Collections. The project was a partnership with ITMO University in Saint Petersburg. Our work focused on document segmentation and the prediction of image, text, title, and mixedtext regions in the document images. The mixedtext category refers to segments where the typeface and text layout are mixed with other visual elements such as graphics, photographs, and illustrations. This category identifies sections that present problems for OCR and also highlights the experimental use of text, images, and other elements in the documents. For each of the ten journals of interest in Princeton's digital collections (DPUL), we started with the IIIF manifest URI. With these manifests, we downloaded each of the 24,000 document images. The URI for each of the images is included in the dataset and a full list is available in `IIIF_URIs.json`. ## Authors Natalia Ermolaev, Thomas Keenan, Katherine Reischl, Andrew Janco, Quinn Dombrowski, Antonina Puchkovskaia, Alexander Jacobson, Anastasiia Mamonova, Michael Galperin and Vladislav Tretyak ## Journal manifests - [Эрмитаж](https://figgy.princeton.edu/concern/scanned_resources/6b561fbb-ba28-4afb-91d2-d77b8728d7d9/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/6b561fbb-ba28-4afb-91d2-d77b8728d7d9/manifest) - [Вестник искусств](https://figgy.princeton.edu/concern/scanned_resources/ad256b35-9ad0-4f75-bf83-3bad1a7c6018/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/ad256b35-9ad0-4f75-bf83-3bad1a7c6018/manifest) - [Советский театр](https://figgy.princeton.edu/concern/scanned_resources/f33993bb-a041-40a1-b11f-f660da825583/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/f33993bb-a041-40a1-b11f-f660da825583/manifest) - [Рабис](https://figgy.princeton.edu/concern/scanned_resources/01f4236f-0a2f-473c-946f-d9bbec12f8ea/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/01f4236f-0a2f-473c-946f-d9bbec12f8ea/manifest) - [Даёшь](https://figgy.princeton.edu/concern/scanned_resources/e036a5da-97a8-4041-ad62-a57af44359e2/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/e036a5da-97a8-4041-ad62-a57af44359e2/manifest) - [Персимфанс](https://figgy.princeton.edu/concern/scanned_resources/af43d19a-3659-4dd0-a0fc-4c74ce521ad6/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/af43d19a-3659-4dd0-a0fc-4c74ce521ad6/manifest) - [Тридцать дней](https://figgy.princeton.edu/concern/scanned_resources/d2d488af-2980-4554-a9ef-aacbaf463ec8/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/d2d488af-2980-4554-a9ef-aacbaf463ec8/manifest) - [За пролетарское искусство](https://figgy.princeton.edu/concern/scanned_resources/38f89d57-8e64-4033-97d6-b925c407584a/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/38f89d57-8e64-4033-97d6-b925c407584a/manifest) - [Бригада художников](https://figgy.princeton.edu/concern/scanned_resources/66d00a87-5ea9-439a-a909-95d697401a2b/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/66d00a87-5ea9-439a-a909-95d697401a2b/manifest) - [Зрелища](https://figgy.princeton.edu/concern/scanned_resources/1af8b322-a0b1-46af-8541-5c3054af8098/manifest?manifest=https://figgy.princeton.edu/concern/scanned_resources/1af8b322-a0b1-46af-8541-5c3054af8098/manifest) ## Model Using [makesense.ai](https://www.makesense.ai/) and a custom active learning application called ["Mayakovsky"](https://github.com/CDH-ITMO-Periodicals-Project/mayakovsky) we generated training data for a [YOLOv5 model](https://docs.ultralytics.com/tutorials/train-custom-datasets/). The model was fine-tuned on the new labels and predictions were generated for all images in the collection. ## OCR Using the model's predictions for image, title, text and mixedtext segments, we cropped the image using the bounding boxes and ran OCR on each document segment using Tesseract, Google Vision, and ABBYY FineReader. Given that the output of these various OCR engines can be difficult to compare, the document segments give a common denominator for comparison of OCR outputs. Having three variations of the extracted text can be useful for experiments with OCR post-correction. ## Dataset The dataset contains an entry for each image with the following fields: - filename: the image name (ex. 'Советский театр_1932 No. 4_16') with journal name, year, issue, page. - dpul: the URL for the image's journal in Digital Princeton University Library - journal: the journal name - year: the year of the journal issue - issue: the issue for the image - URI: the IIIF URI used to fetch the image from Princeton's IIIF server - yolo: the raw model prediction (ex '3 0.1655 0.501396 0.311'), in Yolo's normalized xywh format (object-class x y width height). The labels are 'image'=0, 'mixedtext'=1, 'title'=2, 'textblock'=3. - yolo_predictions: a List with a dictionary for each of the model's predictions with fields for: - label: the predicted label - x: the x-value location of the center point of the prediction - y: the y-value location of the center point of the prediction - w: the total width of the prediction's bounding box - h: the total height of the prediction's bounding box - abbyy_text: the text extracted from the predicted document segment using ABBY FineReader. Note that due to costs, only about 800 images have this data - tesseract_text: the text extracted from the predicted document segment using Tesseract. - vision_text: the text extracted from the predicted document segment using Google Vision. - vision_labels: entities recognized by Google Vision in image blocks and separated by | (ex. Boating|Boat|Printmaking) # Useage ```python from datasets import load_dataset dataset = load_dataset('ajanco/pesp') for item in dataset['train']: for prediction in item['yolo_predictions']: print(prediction) ```
ajanco/pesp
[ "task_categories:other", "annotations_creators:expert-generated", "language_creators:expert-generated", "language_creators:machine-generated", "multilinguality:monolingual", "source_datasets:original", "language:ru", "license:afl-3.0", "region:us" ]
2022-04-14T11:18:44+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["expert-generated", "machine-generated"], "language": ["ru"], "license": ["afl-3.0"], "multilinguality": ["monolingual"], "size_categories": [], "source_datasets": ["original"], "task_categories": ["other"], "task_ids": [], "pretty_name": "The Pages of Early Soviet Performance (PESP) uses machine learning to generate multiple datasets of early-Soviet illustrated periodicals related to the performing arts. By using computer vision techniques and training a YOLO (You Only Look Once) real-time object detection model, we are producing textual and image data that will facilitate new avenues of research about Soviet culture during the first decades after the October Revolution (1917-1932).\n\nOur starting point is Princeton University Library's Digital PUL (DPUL) where ten titles - totaling 526 issues and approximately 26,000 pages - of Soviet performance journals have been digitized and can be freely viewed online. Journals are a diverse and complex genre: taken together, this collection contains hundreds of thousands of articles, poems, editorial commentary, advertisements as well as images, illustrations and graphic art. Today, researchers can browse the journals and view and download high-quality page images on DPUL."}
2022-07-01T15:18:15+00:00
ce994f5c4a86e3fb2db2f045dbc600575e2e6fb8
# Dataset Card for Conceptual Captions ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Dataset Preprocessing](#dataset-preprocessing) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** [Conceptual Captions homepage](https://ai.google.com/research/ConceptualCaptions/) - **Repository:** [Conceptual Captions repository](https://github.com/google-research-datasets/conceptual-captions) - **Paper:** [Conceptual Captions: A Cleaned, Hypernymed, Image Alt-text Dataset For Automatic Image Captioning](https://www.aclweb.org/anthology/P18-1238/) - **Leaderboard:** [Conceptual Captions leaderboard](https://ai.google.com/research/ConceptualCaptions/competition?active_tab=leaderboard)https://ai.google.com/research/ConceptualCaptions/leaderboard?active_tab=leaderboard - **Point of Contact:** [Conceptual Captions e-mail](mailto:conceptual-captions@google.com) ### Dataset Summary Conceptual Captions is a dataset consisting of ~3.3M images annotated with captions. In contrast with the curated style of other image caption annotations, Conceptual Caption images and their raw descriptions are harvested from the web, and therefore represent a wider variety of styles. More precisely, the raw descriptions are harvested from the Alt-text HTML attribute associated with web images. To arrive at the current version of the captions, we have developed an automatic pipeline that extracts, filters, and transforms candidate image/caption pairs, with the goal of achieving a balance of cleanliness, informativeness, fluency, and learnability of the resulting captions. ### Dataset Preprocessing This dataset doesn't download the images locally by default. Instead, it exposes URLs to the images. To fetch the images, use the following code: ```python from concurrent.futures import ThreadPoolExecutor from functools import partial import io import urllib import PIL.Image from datasets import load_dataset from datasets.utils.file_utils import get_datasets_user_agent USER_AGENT = get_datasets_user_agent() def fetch_single_image(image_url, timeout=None, retries=0): for _ in range(retries + 1): try: request = urllib.request.Request( image_url, data=None, headers={"user-agent": USER_AGENT}, ) with urllib.request.urlopen(request, timeout=timeout) as req: image = PIL.Image.open(io.BytesIO(req.read())) break except Exception: image = None return image def fetch_images(batch, num_threads, timeout=None, retries=0): fetch_single_image_with_args = partial(fetch_single_image, timeout=timeout, retries=retries) with ThreadPoolExecutor(max_workers=num_threads) as executor: batch["image"] = list(executor.map(fetch_single_image_with_args, batch["image_url"])) return batch num_threads = 20 dset = load_dataset("conceptual_captions") dset = dset.map(fetch_images, batched=True, batch_size=100, fn_kwargs={"num_threads": num_threads}) ``` ### Supported Tasks and Leaderboards - `image-captioning`: This dataset can be used to train model for the Image Captioning task. The leaderboard for this task is available [here](https://ai.google.com/research/ConceptualCaptions/competition?active_tab=leaderboard). Official submission output captions are scored against the reference captions from the hidden test set using [this](https://github.com/tylin/coco-caption) implementation of the CIDEr (primary), ROUGE-L and SPICE metrics. ### Languages All captions are in English. ## Dataset Structure ### Data Instances #### `unlabeled` Each instance in this configuration represents a single image with a caption: ``` { 'image_url': 'http://lh6.ggpht.com/-IvRtNLNcG8o/TpFyrudaT6I/AAAAAAAAM6o/_11MuAAKalQ/IMG_3422.JPG?imgmax=800', 'caption': 'a very typical bus station' } ``` #### `labeled` Each instance in this configuration represents a single image with a caption with addtional machine-generated image labels and confidence scores: ``` { 'image_url': 'https://thumb1.shutterstock.com/display_pic_with_logo/261388/223876810/stock-vector-christmas-tree-on-a-black-background-vector-223876810.jpg', 'caption': 'christmas tree on a black background .', 'labels': ['christmas tree', 'christmas decoration', 'font', 'text', 'graphic design', 'illustration','interior design', 'tree', 'christmas eve', 'ornament', 'fir', 'plant', 'pine', 'pine family', 'graphics'], 'MIDs': ['/m/025nd', '/m/05fc9mj', '/m/03gq5hm', '/m/07s6nbt', '/m/03c31', '/m/01kr8f', '/m/0h8nzzj', '/m/07j7r', '/m/014r1s', '/m/05ykl4', '/m/016x4z', '/m/05s2s', '/m/09t57', '/m/01tfm0', '/m/021sdg'], 'confidence_scores': [0.9818305373191833, 0.952756941318512, 0.9227379560470581, 0.8524878621101379, 0.7597672343254089, 0.7493422031402588, 0.7332468628883362, 0.6869218349456787, 0.6552258133888245, 0.6357356309890747, 0.5992692708969116, 0.585474967956543, 0.5222904086112976, 0.5113164782524109, 0.5036579966545105] } ``` ### Data Fields #### `unlabeled` - `image_url`: Static URL for downloading the image associated with the post. - `caption`: Textual description of the image. #### `labeled` - `image_url`: Static URL for downloading the image associated with the post. - `caption`: Textual description of the image. - `labels`: A sequence of machine-generated labels obtained using the [Google Cloud Vision API](https://cloud.google.com/vision). - `MIDs`: A sequence of machine-generated identifiers (MID) corresponding to the label's Google Knowledge Graph entry. - `confidence_scores`: A sequence of confidence scores denoting how likely the corresponing labels are present on the image. ### Data Splits #### `unlabeled` The basic version of the dataset split into Training and Validation splits. The Training split consists of 3,318,333 image-URL/caption pairs and the Validation split consists of 15,840 image-URL/caption pairs. #### `labeled` The labeled version of the dataset with a single. The entire data is contained in Training split, which is a subset of 2,007,090 image-URL/caption pairs from the Training set of the `unlabeled` config. ## Dataset Creation ### Curation Rationale From the paper: > In this paper, we make contributions to both the data and modeling categories. First, we present a new dataset of caption annotations Conceptual Captions (Fig. 1), which has an order of magnitude more images than the COCO dataset. Conceptual Captions consists of about 3.3M himage, descriptioni pairs. In contrast with the curated style of the COCO images, Conceptual Captions images and their raw descriptions are harvested from the web, and therefore represent a wider variety of styles. ### Source Data #### Initial Data Collection and Normalization From the homepage: >For Conceptual Captions, we developed a fully automatic pipeline that extracts, filters, and transforms candidate image/caption pairs, with the goal of achieving a balance of cleanliness, informativeness, fluency, and learnability of the resulting captions. Because no human annotators are involved, the Conceptual Captions dataset generation process is highly scalable. > >To generate this dataset, we started with a Flume pipeline that processes billions of Internet webpages, extracting, filtering, and processing candidate image and caption pairs, and keeping those that pass through several filters. > >We first screen for certain properties like size, aspect ratio, adult content scores. These filters discard more than 65% of the candidates. Next, we use Alt-Texts for text-based filtering, removing captions with non-descriptive text (such as SEO tags or hashtags); we also discard texts with high sentiment polarity or adult content scores, resulting in just 3% of the incoming candidates passing through. > >In the next step, we filter out candidates for which none of the text tokens can be mapped to the visual content of the image. We use image classifiers (e.g., Google Cloud Vision APIs) to assign class labels to images and match these labels against the candidate text (allowing morphological transformations), discarding >around 60% of the candidates that reach this stage. > >The candidates passing the above filters tend to be good Alt-text image descriptions. However, a large majority of these use proper names (for people, venues, locations, etc.), brands, dates, quotes, etc. This creates two distinct problems. First, some of these cannot be inferred based on the image pixels alone. This is problematic because unless the image has the necessary visual information it is not useful for training. Second, even if the proper names could be inferred from the image it is extremely difficult for a model to learn to perform both fine-grained classification and natural-language descriptions simultaneously. We posit that if automatic determination of names, locations, brands, etc. is needed, it should be done as a separate task that may leverage image meta-information (e.g. GPS info), or complementary techniques such as OCR. > >We address the above problems with the insight that proper names should be replaced by words that represent the same general notion, i.e., by their concept. For example, we remove locations (“Crowd at a concert in Los Angeles“ becomes “Crowd at a concert”), names (e.g., “Former Miss World Priyanka Chopra on the red carpet” becomes “actor on the red carpet”), proper noun modifiers (e.g., “Italian cuisine” becomes just “cuisine”) and noun phrases (e.g., “actor and actor” becomes “actors”). Around 20% of the samples are discarded during this transformation because it can leave sentences too short, or otherwise inconsistent. > >Finally, we perform another round of filtering to identify concepts with low-count. We cluster all resolved entities (e.g., “actor”, “dog”, “neighborhood”, etc.) and keep only the candidate types which have a count of over 100 mentions. This retains around 16K entity concepts such as: “person”, “actor”, “artist”, “player” and “illustration”. The less frequent ones that we dropped include “baguette”, “bridle”, “deadline”, “ministry” and “funnel”. #### Who are the source language producers? Not specified. ### Annotations #### Annotation process Annotations are extracted jointly with the images using the automatic pipeline. #### Who are the annotators? Not specified. ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators Piyush Sharma, Nan Ding, Sebastian Goodman and Radu Soricut. ### Licensing Information The dataset may be freely used for any purpose, although acknowledgement of Google LLC ("Google") as the data source would be appreciated. The dataset is provided "AS IS" without any warranty, express or implied. Google disclaims all liability for any damages, direct or indirect, resulting from the use of the dataset. ### Citation Information ```bibtex @inproceedings{sharma2018conceptual, title = {Conceptual Captions: A Cleaned, Hypernymed, Image Alt-text Dataset For Automatic Image Captioning}, author = {Sharma, Piyush and Ding, Nan and Goodman, Sebastian and Soricut, Radu}, booktitle = {Proceedings of ACL}, year = {2018}, } ``` ### Contributions Thanks to [@abhishekkrthakur](https://github.com/abhishekkrthakur) and [@mariosasko](https://github.com/mariosasko) for adding this dataset.
conceptual_captions
[ "task_categories:image-to-text", "task_ids:image-captioning", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:1M<n<10M", "source_datasets:original", "language:en", "license:other", "region:us" ]
2022-04-14T12:08:21+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["en"], "license": ["other"], "multilinguality": ["monolingual"], "size_categories": ["1M<n<10M"], "source_datasets": ["original"], "task_categories": ["image-to-text"], "task_ids": ["image-captioning"], "paperswithcode_id": "conceptual-captions", "pretty_name": "Conceptual Captions", "dataset_info": [{"config_name": "default", "features": [{"name": "id", "dtype": "string"}, {"name": "caption", "dtype": "string"}, {"name": "url", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 623230370, "num_examples": 3318333}, {"name": "validation", "num_bytes": 2846024, "num_examples": 15840}], "download_size": 0, "dataset_size": 626076394}, {"config_name": "unlabeled", "features": [{"name": "image_url", "dtype": "string"}, {"name": "caption", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 584520156, "num_examples": 3318333}, {"name": "validation", "num_bytes": 2698726, "num_examples": 15840}], "download_size": 567211172, "dataset_size": 587218882}, {"config_name": "labeled", "features": [{"name": "image_url", "dtype": "string"}, {"name": "caption", "dtype": "string"}, {"name": "labels", "sequence": "string"}, {"name": "MIDs", "sequence": "string"}, {"name": "confidence_scores", "sequence": "float64"}], "splits": [{"name": "train", "num_bytes": 1199330856, "num_examples": 2007090}], "download_size": 1282463277, "dataset_size": 1199330856}]}
2024-01-18T09:32:42+00:00
9c3742c2e077f17b9e6544910cfc4e23ae81db9d
# Dataset Card ## Disclaimer All rights belong to their owners. Models and datasets can be removed from the site at the request of the copyright holder. ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [How to use](#how-to-use) - [Dataset Structure](#dataset-structure) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [About](#about) ## Dataset Description - **Homepage:** [https://github.com/AlekseyKorshuk/huggingnft](https://github.com/AlekseyKorshuk/huggingnft) - **Repository:** [https://github.com/AlekseyKorshuk/huggingnft](https://github.com/AlekseyKorshuk/huggingnft) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Dataset Summary NFT images dataset for unconditional generation. NFT collection available [here](https://opensea.io/collection/cryptoskulls). Model is available [here](https://huggingface.co/huggingnft/cryptoskulls). Check Space: [link](https://huggingface.co/spaces/AlekseyKorshuk/huggingnft). ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## How to use How to load this dataset directly with the datasets library: ```python from datasets import load_dataset dataset = load_dataset("huggingnft/cryptoskulls") ``` ## Dataset Structure [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Data Fields The data fields are the same among all splits. - `image`: an `image` feature. - `id`: an `int` feature. - `token_metadata`: a `str` feature. - `image_original_url`: a `str` feature. ### Data Splits [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @InProceedings{huggingnft, author={Aleksey Korshuk} year=2022 } ``` ## About *Built by Aleksey Korshuk* [![Follow](https://img.shields.io/github/followers/AlekseyKorshuk?style=social)](https://github.com/AlekseyKorshuk) [![Follow](https://img.shields.io/twitter/follow/alekseykorshuk?style=social)](https://twitter.com/intent/follow?screen_name=alekseykorshuk) [![Follow](https://img.shields.io/badge/dynamic/json?color=blue&label=Telegram%20Channel&query=%24.result&url=https%3A%2F%2Fapi.telegram.org%2Fbot1929545866%3AAAFGhV-KKnegEcLiyYJxsc4zV6C-bdPEBtQ%2FgetChatMemberCount%3Fchat_id%3D-1001253621662&style=social&logo=telegram)](https://t.me/joinchat/_CQ04KjcJ-4yZTky) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/AlekseyKorshuk/huggingnft?style=social)](https://github.com/AlekseyKorshuk/huggingnft)
huggingnft/cryptoskulls
[ "license:mit", "huggingnft", "nft", "huggan", "gan", "image", "images", "region:us" ]
2022-04-14T13:16:34+00:00
{"license": "mit", "tags": ["huggingnft", "nft", "huggan", "gan", "image", "images"], "task": ["unconditional-image-generation"], "datasets": ["huggingnft/cryptoskulls"]}
2022-04-16T16:59:08+00:00
e57e10cf1f793a854486c7dc040ac18d59595199
# Dataset for project: kor_hate_eval(APEACH) ![](https://github.com/jason9693/APEACH/raw/master/resource/dist_topics.png) ## Sample Code <a href="https://colab.research.google.com/drive/1djd0fuoMYIaf7VCHaLQIziJi4_yBJruP#scrollTo=VPR24ysr5Q7k"><img src="https://colab.research.google.com/assets/colab-badge.svg" alt="base"/></a> ## Dataset Descritpion Korean Hate Speech Evaluation Datasets : trained with [BEEP!](https://huggingface.co/datasets/kor_hate) and evaluate with [APEACH](https://github.com/jason9693/APEACH) - **Repository: [Korean HateSpeech Evaluation Dataset](https://github.com/jason9693/APEACH)** - **Paper: [APEACH: Attacking Pejorative Expressions with Analysis on Crowd-Generated Hate Speech Evaluation Datasets](https://arxiv.org/abs/2202.12459)** - **Point of Contact: [Kichang Yang](ykcha9@gmail.com)** ### Languages ko-KR ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json {'text': ['(현재 호텔주인 심정) 아18 난 마른하늘에 날벼락맞고 호텔망하게생겼는데 누군 계속 추모받네....', '....한국적인 미인의 대표적인 분...너무나 곱고아름다운모습...그모습뒤의 슬픔을 미처 알지못했네요ㅠ'], 'class': ['Spoiled', 'Default']} ``` ### Dataset Fields The dataset has the following fields (also called "features"): ```json { "text": "Value(dtype='string', id=None)", "class": "ClassLabel(num_classes=2, names=['Default', 'Spoiled'], id=None)" } ``` ### Dataset Splits This dataset is split into a train and validation split. The split sizes are as follow: | Split name | Num samples | | ------------ | ------------------- | | train (binarized BEEP!) | 7896 | | valid (APEACH) | 3770 | ## Citation ``` @article{yang2022apeach, title={APEACH: Attacking Pejorative Expressions with Analysis on Crowd-Generated Hate Speech Evaluation Datasets}, author={Yang, Kichang and Jang, Wonjun and Cho, Won Ik}, journal={arXiv preprint arXiv:2202.12459}, year={2022} } ```
jason9693/APEACH
[ "task_categories:text-classification", "annotations_creators:crowdsourced", "annotations_creators:crowd-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:ko", "license:cc-by-sa-4.0", "arxiv:2202.12459", "region:us" ]
2022-04-14T13:27:43+00:00
{"annotations_creators": ["crowdsourced", "crowd-generated"], "language_creators": ["found"], "language": ["ko"], "license": ["cc-by-sa-4.0"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["binary-classification"], "paperswithcode_id": "apeach", "pretty_name": "APEACH"}
2022-07-05T03:18:07+00:00
716a41b2ec2e921f24b9b1df564a07f8643989a4
# AutoTrain Dataset for project: kor_hate_eval ## Dataset Descritpion This dataset has been automatically processed by AutoTrain for project kor_hate_eval. ### Languages The BCP-47 code for the dataset's language is unk. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json [ { "text": "(\ud604\uc7ac \ud638\ud154\uc8fc\uc778 \uc2ec\uc815) \uc54418 \ub09c \ub9c8\ub978\ud558\ub298\uc5d0 \ub0a0\ubcbc\ub77d\ub9de\uace0 \ud638\ud154\ub9dd\ud558\uac8c\uc0dd\uacbc\ub294\ub370 \ub204\uad70 \uacc4\uc18d \ucd94\ubaa8\ubc1b\ub124....", "target": 1 }, { "text": "....\ud55c\uad6d\uc801\uc778 \ubbf8\uc778\uc758 \ub300\ud45c\uc801\uc778 \ubd84...\ub108\ubb34\ub098 \uacf1\uace0\uc544\ub984\ub2e4\uc6b4\ubaa8\uc2b5...\uadf8\ubaa8\uc2b5\ub4a4\uc758 \uc2ac\ud514\uc744 \ubbf8\ucc98 \uc54c\uc9c0\ubabb\ud588\ub124\uc694\u3160", "target": 0 } ] ``` ### Dataset Fields The dataset has the following fields (also called "features"): ```json { "text": "Value(dtype='string', id=None)", "target": "ClassLabel(num_classes=2, names=['Default', 'Spoiled'], id=None)" } ``` ### Dataset Splits This dataset is split into a train and validation split. The split sizes are as follow: | Split name | Num samples | | ------------ | ------------------- | | train | 7896 | | valid | 3770 |
jason9693/autotrain-data-kor_hate_eval
[ "task_categories:text-classification", "region:us" ]
2022-04-14T14:42:28+00:00
{"task_categories": ["text-classification"]}
2022-04-14T14:44:07+00:00
ab7ad9330ec63b9652e8f091c76dfe4c549ba606
# Dataset Card ## Disclaimer All rights belong to their owners. Models and datasets can be removed from the site at the request of the copyright holder. ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [How to use](#how-to-use) - [Dataset Structure](#dataset-structure) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [About](#about) ## Dataset Description - **Homepage:** [https://github.com/AlekseyKorshuk/huggingnft](https://github.com/AlekseyKorshuk/huggingnft) - **Repository:** [https://github.com/AlekseyKorshuk/huggingnft](https://github.com/AlekseyKorshuk/huggingnft) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Dataset Summary NFT images dataset for unconditional generation. NFT collection available [here](https://opensea.io/collection/azuki). Model is available [here](https://huggingface.co/huggingnft/azuki). Check Space: [link](https://huggingface.co/spaces/AlekseyKorshuk/huggingnft). ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## How to use How to load this dataset directly with the datasets library: ```python from datasets import load_dataset dataset = load_dataset("huggingnft/azuki") ``` ## Dataset Structure [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Data Fields The data fields are the same among all splits. - `image`: an `image` feature. - `id`: an `int` feature. - `token_metadata`: a `str` feature. - `image_original_url`: a `str` feature. ### Data Splits [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @InProceedings{huggingnft, author={Aleksey Korshuk} year=2022 } ``` ## About *Built by Aleksey Korshuk* [![Follow](https://img.shields.io/github/followers/AlekseyKorshuk?style=social)](https://github.com/AlekseyKorshuk) [![Follow](https://img.shields.io/twitter/follow/alekseykorshuk?style=social)](https://twitter.com/intent/follow?screen_name=alekseykorshuk) [![Follow](https://img.shields.io/badge/dynamic/json?color=blue&label=Telegram%20Channel&query=%24.result&url=https%3A%2F%2Fapi.telegram.org%2Fbot1929545866%3AAAFGhV-KKnegEcLiyYJxsc4zV6C-bdPEBtQ%2FgetChatMemberCount%3Fchat_id%3D-1001253621662&style=social&logo=telegram)](https://t.me/joinchat/_CQ04KjcJ-4yZTky) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/AlekseyKorshuk/huggingnft?style=social)](https://github.com/AlekseyKorshuk/huggingnft)
huggingnft/azuki
[ "license:mit", "huggingnft", "nft", "huggan", "gan", "image", "images", "region:us" ]
2022-04-14T19:36:39+00:00
{"license": "mit", "tags": ["huggingnft", "nft", "huggan", "gan", "image", "images"], "task": ["unconditional-image-generation"], "datasets": ["huggingnft/azuki"]}
2022-04-16T16:59:08+00:00
8495090e5604bf7070abe230cb59090e24ab25ae
# Dataset Card ## Disclaimer All rights belong to their owners. Models and datasets can be removed from the site at the request of the copyright holder. ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [How to use](#how-to-use) - [Dataset Structure](#dataset-structure) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [About](#about) ## Dataset Description - **Homepage:** [https://github.com/AlekseyKorshuk/huggingnft](https://github.com/AlekseyKorshuk/huggingnft) - **Repository:** [https://github.com/AlekseyKorshuk/huggingnft](https://github.com/AlekseyKorshuk/huggingnft) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Dataset Summary NFT images dataset for unconditional generation. NFT collection available [here](https://opensea.io/collection/mutant-ape-yacht-club). Model is available [here](https://huggingface.co/huggingnft/mutant-ape-yacht-club). Check Space: [link](https://huggingface.co/spaces/AlekseyKorshuk/huggingnft). ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## How to use How to load this dataset directly with the datasets library: ```python from datasets import load_dataset dataset = load_dataset("huggingnft/mutant-ape-yacht-club") ``` ## Dataset Structure [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Data Fields The data fields are the same among all splits. - `image`: an `image` feature. - `id`: an `int` feature. - `token_metadata`: a `str` feature. - `image_original_url`: a `str` feature. ### Data Splits [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @InProceedings{huggingnft, author={Aleksey Korshuk} year=2022 } ``` ## About *Built by Aleksey Korshuk* [![Follow](https://img.shields.io/github/followers/AlekseyKorshuk?style=social)](https://github.com/AlekseyKorshuk) [![Follow](https://img.shields.io/twitter/follow/alekseykorshuk?style=social)](https://twitter.com/intent/follow?screen_name=alekseykorshuk) [![Follow](https://img.shields.io/badge/dynamic/json?color=blue&label=Telegram%20Channel&query=%24.result&url=https%3A%2F%2Fapi.telegram.org%2Fbot1929545866%3AAAFGhV-KKnegEcLiyYJxsc4zV6C-bdPEBtQ%2FgetChatMemberCount%3Fchat_id%3D-1001253621662&style=social&logo=telegram)](https://t.me/joinchat/_CQ04KjcJ-4yZTky) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/AlekseyKorshuk/huggingnft?style=social)](https://github.com/AlekseyKorshuk/huggingnft)
huggingnft/mutant-ape-yacht-club
[ "license:mit", "huggingnft", "nft", "huggan", "gan", "image", "images", "region:us" ]
2022-04-14T19:51:06+00:00
{"license": "mit", "tags": ["huggingnft", "nft", "huggan", "gan", "image", "images"], "task": ["unconditional-image-generation"], "datasets": ["huggingnft/mutant-ape-yacht-club"]}
2022-04-16T16:59:08+00:00
dbb125403842b8924d864f09f1c0eb357bd57435
# Dataset Card for [Dataset Name] ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@github-username](https://github.com/<github-username>) for adding this dataset.
XiangPan/waimai_10k
[ "region:us" ]
2022-04-14T21:14:23+00:00
{}
2022-04-14T21:38:31+00:00
b51fb34fb5ec42c55374863c3cbbf30f2c0f6f55
XiangPan/ChnSentiCorp_htl_8k
[ "license:other", "region:us" ]
2022-04-14T21:45:08+00:00
{"license": "other"}
2022-04-14T21:46:21+00:00
777a739ac615ec9c7015c36c86234533a5d50af2
# Configuration
awacke1/data.csv
[ "region:us" ]
2022-04-14T21:58:15+00:00
{"title": "Data.csv", "emoji": "\ud83d\udc28", "colorFrom": "pink", "colorTo": "gray", "sdk": "gradio", "sdk_version": "2.4.2", "app_file": "app.py", "pinned": false}
2023-09-22T12:12:17+00:00
966a1e77847f3e22de5fb961665331767082d571
[Needs More Information] # Dataset Card for squad_it_exp ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Languages](#languages) - [Additional Information](#additional-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** [Needs More Information] - **Repository:** [Needs More Information] - **Paper:** [Needs More Information] - **Leaderboard:** [Needs More Information] - **Point of Contact:** [Needs More Information] ### Dataset Summary SQuAD-it-exp is a dataset derived from the SQuAD-it dataset originally created by Croce et al. to 2018.<br/> SQuAD-it-exp has been enriched by adding new unanswerable questions in SQuAD v2 format.<br/> The dataset contains nearly 90,000 pairs of questions/answers in Italian. ### Languages The dataset is for the ITALIAN language ### Citation Information ``` @InProceedings{10.1007/978-3-030-03840-3_29, author="Croce, Danilo and Zelenanska, Alexandra and Basili, Roberto", editor="Ghidini, Chiara and Magnini, Bernardo and Passerini, Andrea and Traverso, Paolo", title="Neural Learning for Question Answering in Italian", booktitle="AI*IA 2018 -- Advances in Artificial Intelligence", year="2018", publisher="Springer International Publishing", address="Cham", pages="389--402", isbn="978-3-030-03840-3" } ```
bullmount/squad-it-exp
[ "region:us" ]
2022-04-15T04:03:58+00:00
{}
2022-04-17T17:30:50+00:00
4942ad98569e62b710c547f39f916724088ef520
### Dataset Summary This dataset is extracted from Climate Fever dataset (https://www.sustainablefinance.uzh.ch/en/research/climate-fever.html), pre-processed and ready to train and evaluate. The training objective is a text classification task - given a claim and evidence, predict if claim is related to evidence.
mwong/fever-claim-related
[ "task_categories:text-classification", "task_ids:fact-checking", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:extended|climate_fever", "language:en", "license:cc-by-sa-3.0", "license:gpl-3.0", "region:us" ]
2022-04-15T06:04:59+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["cc-by-sa-3.0", "gpl-3.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["extended|climate_fever"], "task_categories": ["text-classification"], "task_ids": ["fact-checking"], "paperswithcode_id": "fever", "pretty_name": "fever"}
2022-10-25T09:06:56+00:00
4c366c1882d27123f4aa640b824a29998f1c642d
### Dataset Summary This dataset is extracted from Climate Fever dataset (https://www.sustainablefinance.uzh.ch/en/research/climate-fever.html), pre-processed and, ready to train and evaluate. The training objective is a text classification task - given a claim and evidence, predict if claim is related to evidence.
mwong/climate-claim-related
[ "task_categories:text-classification", "task_ids:fact-checking", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:extended|climate_fever", "language:en", "license:cc-by-sa-3.0", "license:gpl-3.0", "region:us" ]
2022-04-15T06:09:18+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["cc-by-sa-3.0", "gpl-3.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["extended|climate_fever"], "task_categories": ["text-classification"], "task_ids": ["fact-checking"], "paperswithcode_id": "climate-fever", "pretty_name": "climate-fever"}
2022-10-25T09:06:59+00:00
3d6d4da7cc2b491448d172eebf397560fdded10a
The reddit_keywords.tsv file contains about 170k single word embeddings (scraped from reddit, filtering from an initial set of ~700k based on a minimum occurrence threshold) in this format: ```tsv temporary -0.276235,-0.181357,-0.325729,0.129826,0.016490,-0.230246,-0.039997,-0.990187,-0.014679,-0.044081,-0.120046,-0.250614,-0.303871,-0.264685,-0.010019,-0.158764,0.086107,-0.018172,0.003005,-0.383161,0.412182,0.104374,0.041335,-0.018206,0.085453,0.016297,-0.015680,0.047611,-0.267469,0.046825,-0.367247,-0.020667,-0.348124,0.055445,-0.303014,0.087954,0.077361,-0.052910,0.404438,-0.107339,-0.027286,-0.174772,0.287671,0.118175,0.224158,0.210142,0.071295,0.052860,0.235766,-0.140977,-0.355314,-0.421407,0.076506,-0.050502,0.334099,-0.090490,-0.109730,0.517465,0.057345,0.322140,0.217463,-0.218778,0.200798,0.140536,0.160337,-0.302322,-0.098611,-0.100849,-0.171952,-0.333828,0.143839,-0.010286,0.103448,0.046543,-0.094578,-0.083335,0.216615,-0.185091,0.028321,-0.251232,-0.021522,0.135202,-0.059559,0.513552,-0.156604,-0.426751,0.029338,-0.086346,-0.001045,-0.210324,-0.196247,-0.127054,-1.732658,0.172654,0.064660,0.051606,0.393296,0.132444,0.068706,-0.264383,0.083144,0.357062,0.501775,0.099174,-0.179929,-0.031447,0.077417,0.141482,-0.302417,0.160296,0.484913,0.070273,0.117609,-0.024784,0.086234,-0.164586,-0.211837,0.243161,0.118945,0.051511,0.225772,-0.207831,-0.132836,0.096240,-0.443813,-0.347750,0.192331,0.119417,-0.067559,-0.208074,-0.117854,0.078054,0.401030,6.348532,-0.012304,-0.099742,-0.065778,-0.299336,-0.164993,-0.089712,0.153861,0.244722,0.138961,0.231054,-0.296617,-0.129511,-0.021327,-0.005316,-0.187050,-0.073289,0.019646,0.458080,-0.027326,0.283158,0.137897,-0.196312,0.023471,0.342747,0.109227,-0.137838,-0.008336,-0.212090,-0.277437,-0.088123,-0.150103,0.030977,0.094198,-0.086804,0.260256,0.036756,0.118120,0.409172,-0.174826,0.454344,-0.333416,0.069056,-0.143509,-0.263730,0.016844,-0.069509,0.240573,0.104100,-0.138059,-0.037173,-0.189750,0.015344,0.034381,-0.243249,-0.052328,-0.111057,0.015412,-0.114713,-0.321371,-0.207981,0.037036,0.103251,-0.011858,-0.289237,0.111561,-0.170033,-0.178935,-0.072297,-0.042672,0.190604,0.174237,-0.095280,0.302311,0.024456,0.038216,-0.223006,0.372462,0.323767,0.078378,-0.297173,-0.195620,0.417219,-0.187052,-0.542408,-0.134892,-0.226160,-0.530608,-0.161821,0.120570,0.010190,0.011004,0.218169,0.322732,0.095584,0.424685,0.293537,-0.191970,0.038989,0.042194,-0.388086,0.496116,0.204738,-0.145585,0.463766,-0.227611,0.127603,-0.074332,-0.199442,-0.055274,-0.042825,-0.120296,0.017672,-0.450518,-0.314901,-0.045003,0.031523,-0.079665,-0.315374,0.305340,0.004655,-0.083071,0.191413,0.043845,-0.213311,0.129284,-0.218377,-0.282955,-0.066901,-0.068339,0.002564,-0.146045,0.056669,0.186583,-0.048750,-0.072946,-0.071184,-0.202749,-0.217035,0.276314,-0.282127,0.128067,0.097095,-0.246900,0.232340,-0.238046,-0.304384,0.067498,0.018847,-0.058201,-0.283596,-0.215553,0.035647,0.096342,-0.175125,0.026618,-0.319932,0.423662,-0.063089,0.251738,0.073425,-0.242309,-0.272967,-0.218592,-0.050702,0.091938,0.026258,0.141810,0.014719,-0.415617,0.102258,0.323665,0.213101,-0.219119,-0.074313,-0.075735,0.031039,-0.085159,0.187972,6.345324,0.043324,-0.220423,0.052132,-0.001249,-0.114997,0.001450,0.004655,0.365987,0.536724,0.394376,0.003819,0.262951,-0.065768,-0.087903,0.027754,-0.069572,-2.503358,0.097163,0.222208,0.032130,0.004387,0.129158,-0.238117,0.168215,-0.196026,-0.092511,-0.095957,0.519996,0.053166,-0.138281,-0.071842,-0.024337,-0.182440,0.207966,0.262904,0.325529,-0.087270,0.199483,-0.098656,0.097615,0.014249,-0.074579,0.351518,0.094744,0.148318,-0.173189,0.033593,0.027609,-0.045624,0.188491,0.203499,0.229421,0.050809,-0.222414,-0.016397,0.086318,0.116249,-0.242203,0.120892,0.042388,0.372276,-0.049954,-0.338517,-0.180879,0.083117,-0.284963,-0.178325,0.079176,0.019744,-0.023706,0.391955,-0.189259,-0.373736,0.149015,0.502598,-0.498027,-0.154271,-0.093499,-0.015292,-0.554516,0.355195,0.013390,0.475157,-0.366012,-0.138618,-0.045420,0.528353,0.134862,0.025135,0.141193,-0.075705,-0.265913,-0.227393,0.319143,-0.135606,-0.055334,-0.265537,0.124943,-0.176613,0.301410,0.243831,-0.190008,0.130851,0.057539,0.044628,0.205449,0.315888,-0.097760,-0.251490,-0.039288,-0.009690,-0.013857,0.292198,-0.114490,0.058920,0.032257,0.197568,-0.117429,-0.049549,-0.274646,-0.097156,-0.057420,0.261883,0.105485,-0.131978,-0.083086,0.492079,0.056150,0.163082,0.052169,-0.258462,0.164738,-0.121904,-0.349110,-0.399021,0.109116,0.108278,-0.102895,0.075380,0.120979,0.164346,-0.173332,0.038970,0.239190,0.404884,0.202795,0.021855,0.014958,0.220877,0.214221,-0.309071,0.157248,-0.182312,-0.069097,-0.271037,0.178052,-0.173829,0.410394,-0.023872,-0.118251,0.140042,-0.055087,0.269867,0.401690,0.251227,0.097262,0.225146,0.180279,-0.679833,0.014100,0.017635,-0.020673,0.288165,-0.162649,0.272822,0.118945,-0.178165,0.105399,0.076920,0.289865,0.479189,-0.379978,-0.074296,0.221087,0.110328,-0.434901,-0.009920,-0.329799,-0.326210,0.121444,-0.399424,0.131924,0.035093,-0.330143,-0.332781,-0.375134,-0.429944,-0.028793,-0.084496 permanent -0.125035,-0.234378,0.011184,0.196125,-0.178078,-0.278433,-0.169808,-0.477378,-0.091331,0.051704,0.052124,-0.342429,0.236901,-0.503706,-0.054427,0.378874,0.356929,0.098530,0.213484,-0.350122,0.476689,0.349297,-0.421352,0.131538,-0.037294,0.242601,0.110521,0.297674,-0.003884,-0.164057,-0.181568,-0.114656,-0.022335,-0.058460,-0.392774,0.592076,-0.037568,-0.093719,0.273190,0.031433,-0.276135,-0.129429,0.202552,0.247301,0.162464,0.331153,0.150925,0.103975,0.040481,-0.308759,-0.468749,-0.118056,-0.177642,0.071796,0.019445,-0.051476,0.051152,0.208523,0.207935,0.215263,0.240936,-0.260006,0.273524,-0.102152,0.086342,-0.583079,0.104273,0.052269,-0.079865,-0.353752,-0.042390,0.052536,0.373398,-0.083875,-0.085006,-0.094790,0.209163,0.116218,-0.000282,0.063966,-0.142604,0.170597,-0.014974,0.339414,-0.459107,-0.563759,0.073553,0.011647,0.132144,0.024776,-0.104373,-0.136440,-1.464302,0.560471,0.167517,0.387043,0.013425,0.354265,-0.273501,-0.138256,0.346923,0.277063,0.132669,-0.100053,-0.031133,-0.137729,-0.038392,0.127757,0.201051,0.122387,-0.091108,0.112959,-0.076981,-0.091213,0.259445,-0.250712,-0.086296,0.077766,-0.400991,-0.061569,0.295548,-0.546704,-0.181826,-0.145557,-0.003189,-0.065816,0.313023,-0.340320,-0.232408,0.108998,0.259111,0.151180,-0.166929,6.414488,0.501402,-0.091578,-0.057641,-0.482665,-0.142667,-0.264874,0.361437,0.394330,-0.229426,-0.091375,-0.243107,0.303489,-0.005123,-0.055163,0.015856,0.069838,0.031935,0.278514,0.166143,0.474343,0.105431,-0.076213,0.039309,-0.111546,0.012941,-0.164336,-0.017733,-0.281277,-0.086701,-0.025275,0.286478,-0.012244,0.419024,-0.218707,0.303495,-0.144674,-0.015870,0.211324,-0.125048,0.148710,-0.164560,0.090908,-0.009281,-0.350103,0.044986,0.231121,0.168193,-0.172223,-0.155072,-0.071494,-0.171293,0.057206,0.509076,-0.468795,-0.048402,-0.062685,-0.073230,-0.009878,-0.075013,-0.291539,0.082641,0.120893,-0.036523,-0.371523,-0.089427,-0.135797,0.039259,-0.154754,-0.454964,0.118403,0.345686,0.308087,0.189306,-0.186566,-0.052662,0.092485,0.443999,0.371476,0.544698,0.163462,0.211605,0.028551,-0.331050,-0.118373,-0.130023,-0.238063,-0.194845,-0.147683,0.269614,0.094254,0.080196,-0.016950,0.226205,0.251216,0.029845,0.241027,0.037793,0.250348,0.178878,-0.370625,0.021588,0.053517,0.089717,0.034888,-0.127252,-0.095143,-0.048264,-0.038292,-0.114234,0.081980,-0.344073,-0.199645,0.133945,0.046776,-0.164213,-0.125509,-0.078354,0.015241,0.525474,0.109172,-0.063231,-0.204491,-0.021912,-0.035685,-0.036702,-0.021009,-0.296292,-0.110856,-0.017419,-0.346117,0.123624,-0.022428,0.178189,-0.263868,-0.301899,-0.151516,0.189889,-0.468874,0.149208,-0.343711,-0.091530,0.159136,-0.205789,0.289440,0.010746,-0.458909,-0.003668,-0.154943,0.190147,-0.072661,-0.098433,-0.260481,-0.013620,0.239844,0.175296,0.013196,0.417082,0.388816,0.610390,0.218911,-0.285315,-0.397401,-0.407900,0.112988,-0.276133,-0.189056,0.077117,-0.106753,-0.315161,0.237132,0.145833,0.157616,-0.081629,-0.078093,0.011940,0.147423,-0.169398,0.207446,6.412528,0.098466,-0.073285,0.456039,-0.219336,0.225516,-0.126300,-0.085544,0.067576,0.480005,0.323118,-0.062557,0.029795,0.159936,0.207522,0.061824,0.081886,-2.257158,-0.030088,0.212595,0.034217,0.134851,0.060425,0.273878,0.141417,-0.372521,-0.256853,-0.449594,0.111124,-0.035848,0.246770,0.013314,0.095374,0.004721,0.067856,0.068077,0.466395,-0.224380,0.224336,-0.260102,-0.119188,-0.059211,-0.037218,0.257120,0.285553,-0.044297,0.090169,0.235599,0.122039,-0.496953,-0.158066,-0.082501,-0.067778,-0.129525,0.083023,0.095774,-0.055067,0.059558,-0.207958,-0.003060,-0.127451,0.016735,0.315940,0.089926,0.034565,0.117776,-0.771444,0.075713,0.224506,-0.118574,0.011695,-0.278467,0.103532,-0.355093,0.072645,0.554889,-0.227020,-0.206997,-0.082026,0.034534,-0.053694,-0.206349,0.212462,0.297205,-0.001969,-0.115018,-0.185391,0.439835,0.206251,0.275428,0.382456,0.143425,-0.285471,-0.118383,-0.042975,-0.127418,-0.086245,-0.013727,-0.001033,0.160316,0.266595,-0.023524,-0.233376,-0.004141,-0.194787,0.362530,0.329915,0.154557,-0.059378,-0.262286,-0.268349,0.006456,0.362081,-0.316577,-0.308787,-0.025319,-0.024902,0.180332,-0.309457,0.196798,-0.184453,0.050009,-0.142146,-0.352558,-0.272899,0.163149,-0.057184,0.129606,0.054385,0.049695,-0.017398,-0.508930,-0.402540,0.189464,-0.468858,-0.395760,0.130050,-0.095393,-0.081481,-0.148284,0.425590,0.208711,-0.175284,-0.026740,0.072050,0.488482,0.123898,0.185273,0.119057,-0.220553,0.085528,0.105237,0.326145,0.150759,0.081640,-0.403034,0.053875,-0.100664,0.156698,-0.125445,0.020690,0.407489,0.089345,0.201447,0.087196,0.215535,-0.358267,0.663917,0.102616,-0.911252,0.242669,0.106915,0.214653,0.051177,0.009364,-0.139282,-0.122035,0.074249,0.219813,-0.034759,0.089622,0.741162,0.248126,-0.458298,0.276372,-0.191041,-0.380611,0.273156,-0.160504,0.010753,-0.103899,0.379443,0.501672,-0.108174,-0.292099,0.073850,-0.201425,-0.711728,0.321330,0.047050 ```
rocca/clip-keyphrase-embeddings
[ "license:apache-2.0", "region:us" ]
2022-04-15T06:25:59+00:00
{"license": "apache-2.0"}
2022-04-15T07:44:51+00:00
cac1d6c26bc0c5266661019473ac0ffb33bcf9bc
# Dataset Card for C4 ## Table of Contents - [Dataset Card for C4](#dataset-card-for-c4) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Initial Data Collection and Normalization](#initial-data-collection-and-normalization) - [Who are the source language producers?](#who-are-the-source-language-producers) - [Annotations](#annotations) - [Annotation process](#annotation-process) - [Who are the annotators?](#who-are-the-annotators) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://huggingface.co/datasets/allenai/c4 - **Paper:** https://arxiv.org/abs/1910.10683 ### Dataset Summary A colossal, cleaned version of Common Crawl's web crawl corpus. Based on Common Crawl dataset: "https://commoncrawl.org". This is the version prepared by AllenAI, hosted at this address: https://huggingface.co/datasets/allenai/c4 It comes in four variants: - `en`: 305GB in JSON format - `en.noblocklist`: 380GB in JSON format - `en.noclean`: 2.3TB in JSON format - `realnewslike`: 15GB in JSON format The `en.noblocklist` variant is exactly the same as the `en` variant, except we turned off the so-called "badwords filter", which removes all documents that contain words from the lists at https://github.com/LDNOOBW/List-of-Dirty-Naughty-Obscene-and-Otherwise-Bad-Words. ### Supported Tasks and Leaderboards C4 is mainly intended to pretrain language models and word representations. ### Languages The dataset is in English. ## Dataset Structure ### Data Instances An example form the `en` config is: ``` { 'url': 'https://klyq.com/beginners-bbq-class-taking-place-in-missoula/', 'text': 'Beginners BBQ Class Taking Place in Missoula!\nDo you want to get better at making delicious BBQ? You will have the opportunity, put this on your calendar now. Thursday, September 22nd join World Class BBQ Champion, Tony Balay from Lonestar Smoke Rangers. He will be teaching a beginner level class for everyone who wants to get better with their culinary skills.\nHe will teach you everything you need to know to compete in a KCBS BBQ competition, including techniques, recipes, timelines, meat selection and trimming, plus smoker and fire information.\nThe cost to be in the class is $35 per person, and for spectators it is free. Included in the cost will be either a t-shirt or apron and you will be tasting samples of each meat that is prepared.', 'timestamp': '2019-04-25T12:57:54Z' } ``` ### Data Fields The data have several fields: - `url`: url of the source as a string - `text`: text content as a string - `timestamp`: timestamp as a string ### Data Splits | name | train |validation| |----------------|--------:|---------:| | en |364868892| 364608| | en.noblocklist |393391519| 393226| | en.noclean | ?| ?| | realnewslike | 13799838| 13863| ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization C4 dataset is a collection of about 750GB of English-language text sourced from the public Common Crawl web scrape. It includes heuristics to extract only natural language (as opposed to boilerplate and other gibberish) in addition to extensive deduplication. You can find the code that has been used to build this dataset in [c4.py](https://github.com/tensorflow/datasets/blob/5952d3d60d60e1727786fa7a9a23d24bb463d4d6/tensorflow_datasets/text/c4.py) by Tensorflow Datasets. The dataset was explicitly designed to be English only: any page that was not given a probability of at least 99% of being English by [langdetect](https://github.com/Mimino666/langdetect) was discarded. #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information AllenAI are releasing this dataset under the terms of ODC-BY. By using this, you are also bound by the Common Crawl terms of use in respect of the content contained in the dataset. ### Citation Information ``` @article{2019t5, author = {Colin Raffel and Noam Shazeer and Adam Roberts and Katherine Lee and Sharan Narang and Michael Matena and Yanqi Zhou and Wei Li and Peter J. Liu}, title = {Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer}, journal = {arXiv e-prints}, year = {2019}, archivePrefix = {arXiv}, eprint = {1910.10683}, } ``` ### Contributions Thanks to [@dirkgr](https://github.com/dirkgr) and [@lhoestq](https://github.com/lhoestq) for adding this dataset.
Peihao/test-dateset
[ "task_categories:text-generation", "task_categories:fill-mask", "task_ids:language-modeling", "task_ids:masked-language-modeling", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:multilingual", "size_categories:100M<n<1B", "source_datasets:original", "language:en", "license:odc-by", "arxiv:1910.10683", "region:us" ]
2022-04-15T06:45:58+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["found"], "language": ["en"], "license": ["odc-by"], "multilinguality": ["multilingual"], "size_categories": ["100M<n<1B"], "source_datasets": ["original"], "task_categories": ["text-generation", "fill-mask"], "task_ids": ["language-modeling", "masked-language-modeling"], "paperswithcode_id": "c4", "pretty_name": "C4"}
2022-10-25T09:08:29+00:00
56a5a8b0efca4529b0891c6a3cdfb35c2309dfc4
# Dataset Card for Conceptual 12M ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Dataset Preprocessing](#dataset-preprocessing) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Repository:** [Conceptual 12M repository](https://github.com/google-research-datasets/conceptual-12m) - **Paper:** [Conceptual 12M: Pushing Web-Scale Image-Text Pre-Training To Recognize Long-Tail Visual Concepts](https://arxiv.org/abs/2102.08981) - **Point of Contact:** [Conceptual Captions e-mail](mailto:conceptual-captions@google.com) ### Dataset Summary Conceptual 12M (CC12M) is a dataset with 12 million image-text pairs specifically meant to be used for visionand-language pre-training. Its data collection pipeline is a relaxed version of the one used in Conceptual Captions 3M (CC3M). ### Dataset Preprocessing This dataset doesn't download the images locally by default. Instead, it exposes URLs to the images. To fetch the images, use the following code: ```python from concurrent.futures import ThreadPoolExecutor from functools import partial import io import urllib import PIL.Image from datasets import load_dataset from datasets.utils.file_utils import get_datasets_user_agent USER_AGENT = get_datasets_user_agent() def fetch_single_image(image_url, timeout=None, retries=0): for _ in range(retries + 1): try: request = urllib.request.Request( image_url, data=None, headers={"user-agent": USER_AGENT}, ) with urllib.request.urlopen(request, timeout=timeout) as req: image = PIL.Image.open(io.BytesIO(req.read())) break except Exception: image = None return image def fetch_images(batch, num_threads, timeout=None, retries=0): fetch_single_image_with_args = partial(fetch_single_image, timeout=timeout, retries=retries) with ThreadPoolExecutor(max_workers=num_threads) as executor: batch["image"] = list(executor.map(fetch_single_image_with_args, batch["image_url"])) return batch num_threads = 20 dset = load_dataset("conceptual_12m") dset = dset.map(fetch_images, batched=True, batch_size=100, fn_kwargs={"num_threads": num_threads}) ``` ### Supported Tasks and Leaderboards - `image-captioning`: This dataset can be used to train model for the Image Captioning task. ### Languages All captions are in English. ## Dataset Structure ### Data Instances Each instance represents a single image with a caption: ``` { 'image_url': 'http://lh6.ggpht.com/-IvRtNLNcG8o/TpFyrudaT6I/AAAAAAAAM6o/_11MuAAKalQ/IMG_3422.JPG?imgmax=800', 'caption': 'a very typical bus station' } ``` ### Data Fields - `image_url`: Static URL for downloading the image associated with the post. - `caption`: Textual description of the image. ### Data Splits There is only training data, with a total of 12423374 rows ## Dataset Creation ### Curation Rationale Conceptual 12M shares the same pipeline with Conceptual Captions (CC3M), but relaxes some processing steps. ### Source Data #### Initial Data Collection and Normalization From the paper: > To arrive at CC12M, we keep the image-text filtering intact, and relax the unimodal filters only. First, for image-based filtering, we set the maximum ratio of larger to smaller dimension to 2.5 instead of 2. We still keep only JPEG images with size greater than 400 pixels, and still exclude images that trigger pornography detectors. Second, in text-based filtering, we allow text between 3 and 256 words in the alt-text. We still discard candidates with no noun or no determiner, but permit ones without prepositions. We discard the heuristics regarding high unique-word ratio covering various POS tags and word capitalization. We set the maximum fraction of word repetition allowed to 0.2. Given a larger pool of text due to the above relaxations, the threshold for counting a word type as rare is increased from 5 to 20 > The main motivation for CC3M to perform text transformation is that a majority of candidate captions contain ultrafine-grained entities such as proper names (people, venues, locations, etc.), making it extremely difficult to learn as part of the image captioning task. In contrast, we are not restricted by the end task of image caption generation. Our intuition is that relatively more difficult pre-training data would lead to better transferability. We thus do not perform hypernimization or digit substitution. [...] The only exception to the “keep alt-texts as raw as possible” rule is performing person-name substitutions, which we identify as necessary to protect the privacy of the individuals in these images. For this step, we use the Google Cloud Natural Language APIs to detect all named entities of type Person, and substitute them by a special token <PERSON>. Around 25% of all the alt-texts in CC12M are transformed in this fashion. #### Who are the source language producers? Not specified. ### Annotations #### Annotation process Annotations are extracted jointly with the images using the automatic pipeline. #### Who are the annotators? Not specified. ### Personal and Sensitive Information From the paper: > The only exception to the “keep alt-texts as raw as possible” rule is performing person-name substitutions, which we identify as necessary to protect the privacy of the individuals in these images. For this step, we use the Google Cloud Natural Language APIs to detect all named entities of type Person, and substitute them by a special token <PERSON>. Around 25% of all the alt-texts in CC12M are transformed in this fashion. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators Soravit Changpinyo, Piyush Sharma, Nan Ding and Radu Soricut. ### Licensing Information The dataset may be freely used for any purpose, although acknowledgement of Google LLC ("Google") as the data source would be appreciated. The dataset is provided "AS IS" without any warranty, express or implied. Google disclaims all liability for any damages, direct or indirect, resulting from the use of the dataset. ### Citation Information ```bibtex @inproceedings{changpinyo2021cc12m, title = {{Conceptual 12M}: Pushing Web-Scale Image-Text Pre-Training To Recognize Long-Tail Visual Concepts}, author = {Changpinyo, Soravit and Sharma, Piyush and Ding, Nan and Soricut, Radu}, booktitle = {CVPR}, year = {2021}, } ``` ### Contributions Thanks to [@thomasw21](https://github.com/thomasw21) for adding this dataset.
conceptual_12m
[ "task_categories:image-to-text", "task_ids:image-captioning", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:10M<n<100M", "source_datasets:original", "language:en", "license:other", "arxiv:2102.08981", "region:us" ]
2022-04-15T07:06:58+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["en"], "license": ["other"], "multilinguality": ["monolingual"], "size_categories": ["10M<n<100M"], "source_datasets": ["original"], "task_categories": ["image-to-text"], "task_ids": ["image-captioning"], "paperswithcode_id": "cc12m", "pretty_name": "Conceptual 12M", "dataset_info": {"features": [{"name": "image_url", "dtype": "string"}, {"name": "caption", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 2794168030, "num_examples": 12423374}], "download_size": 2707204412, "dataset_size": 2794168030}}
2024-01-18T09:31:48+00:00
d9553ef2399b048492e0ea67f5fa73d2bd24e55a
yumingh/course_project
[ "license:afl-3.0", "region:us" ]
2022-04-15T15:30:33+00:00
{"license": "afl-3.0"}
2022-04-15T15:33:27+00:00
15fe2e2ea5adada9197ba37af304eaecca32fd8b
ajanco/deep
[ "license:mit", "region:us" ]
2022-04-15T19:47:17+00:00
{"license": "mit"}
2022-04-18T19:41:12+00:00
35d54d53495778c09cabd7f86019cac79e578aed
FFHQ 70000张png图片 链接:https://pan.baidu.com/s/1XDfTKWOhtwAAQQJ0KBU4RQ 提取码:bowj ## Flickr-Faces-HQ Dataset (FFHQ) ![Python 3.6](https://img.shields.io/badge/python-3.6-green.svg?style=plastic) ![License CC](https://img.shields.io/badge/license-CC-green.svg?style=plastic) ![Format PNG](https://img.shields.io/badge/format-PNG-green.svg?style=plastic) ![Resolution 1024&times;1024](https://img.shields.io/badge/resolution-1024&times;1024-green.svg?style=plastic) ![Images 70000](https://img.shields.io/badge/images-70,000-green.svg?style=plastic) ![Teaser image](./ffhq-teaser.png) Flickr-Faces-HQ (FFHQ) is a high-quality image dataset of human faces, originally created as a benchmark for generative adversarial networks (GAN): > **A Style-Based Generator Architecture for Generative Adversarial Networks**<br> > Tero Karras (NVIDIA), Samuli Laine (NVIDIA), Timo Aila (NVIDIA)<br> > http://stylegan.xyz/paper The dataset consists of 70,000 high-quality PNG images at 1024&times;1024 resolution and contains considerable variation in terms of age, ethnicity and image background. It also has good coverage of accessories such as eyeglasses, sunglasses, hats, etc. The images were crawled from [Flickr](https://www.flickr.com/), thus inheriting all the biases of that website, and automatically aligned and cropped using [dlib](http://dlib.net/). Only images under permissive licenses were collected. Various automatic filters were used to prune the set, and finally [Amazon Mechanical Turk](https://www.mturk.com/) was used to remove the occasional statues, paintings, or photos of photos. For business inquiries, please contact [researchinquiries@nvidia.com](mailto:researchinquiries@nvidia.com) For press and other inquiries, please contact Hector Marinez at [hmarinez@nvidia.com](mailto:hmarinez@nvidia.com) ## Licenses The individual images were published in Flickr by their respective authors under either [Creative Commons BY 2.0](https://creativecommons.org/licenses/by/2.0/), [Creative Commons BY-NC 2.0](https://creativecommons.org/licenses/by-nc/2.0/), [Public Domain Mark 1.0](https://creativecommons.org/publicdomain/mark/1.0/), [Public Domain CC0 1.0](https://creativecommons.org/publicdomain/zero/1.0/), or [U.S. Government Works](http://www.usa.gov/copyright.shtml) license. All of these licenses allow **free use, redistribution, and adaptation for non-commercial purposes**. However, some of them require giving **appropriate credit** to the original author, as well as **indicating any changes** that were made to the images. The license and original author of each image are indicated in the metadata. * [https://creativecommons.org/licenses/by/2.0/](https://creativecommons.org/licenses/by/2.0/) * [https://creativecommons.org/licenses/by-nc/2.0/](https://creativecommons.org/licenses/by-nc/2.0/) * [https://creativecommons.org/publicdomain/mark/1.0/](https://creativecommons.org/publicdomain/mark/1.0/) * [https://creativecommons.org/publicdomain/zero/1.0/](https://creativecommons.org/publicdomain/zero/1.0/) * [http://www.usa.gov/copyright.shtml](http://www.usa.gov/copyright.shtml) The dataset itself (including JSON metadata, download script, and documentation) is made available under [Creative Commons BY-NC-SA 4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/) license by NVIDIA Corporation. You can **use, redistribute, and adapt it for non-commercial purposes**, as long as you (a) give appropriate credit by **citing our paper**, (b) **indicate any changes** that you've made, and (c) distribute any derivative works **under the same license**. * [https://creativecommons.org/licenses/by-nc-sa/4.0/](https://creativecommons.org/licenses/by-nc-sa/4.0/) ## Overview All data is hosted on Google Drive: | Path | Size | Files | Format | Description | :--- | :--: | ----: | :----: | :---------- | [ffhq-dataset](https://drive.google.com/open?id=1u2xu7bSrWxrbUxk-dT-UvEJq8IjdmNTP) | 2.56 TB | 210,014 | | Main folder | &boxvr;&nbsp; [ffhq-dataset-v1.json](https://drive.google.com/open?id=1IB0BFbN_eRZx9UkJqLHSgJiQhqX-PrI6) | 254 MB | 1 | JSON | Metadata including copyright info, URLs, etc. | &boxvr;&nbsp; [images1024x1024](https://drive.google.com/open?id=1u3Hbfn3Q6jsTlte3BY85CGwId77H-OOu) | 89.1 GB | 70,000 | PNG | Aligned and cropped images at 1024&times;1024 | &boxvr;&nbsp; [thumbnails128x128](https://drive.google.com/open?id=1uJkWCpLUM-BnXW3H_IgVMdfENeNDFNmC) | 1.95 GB | 70,000 | PNG | Thumbnails at 128&times;128 | &boxvr;&nbsp; [in-the-wild-images](https://drive.google.com/open?id=1YyuocbwILsHAjTusSUG-_zL343jlVBhf) | 955 GB | 70,000 | PNG | Original images from Flickr | &boxvr;&nbsp; [tfrecords](https://drive.google.com/open?id=1LTBpJ0W_WLjqza3zdayligS8Dh1V1gA6) | 273 GB | 9 | tfrecords | Multi-resolution data for [StyleGAN](http://stylegan.xyz/code) and [ProGAN](https://github.com/tkarras/progressive_growing_of_gans) | &boxur;&nbsp; [zips](https://drive.google.com/open?id=1WocxvZ4GEZ1DI8dOz30aSj2zT6pkATYS) | 1.28 TB | 4 | ZIP | Contents of each folder as a ZIP archive. High-level statistics: ![Pie charts](./ffhq-piecharts.png) For use cases that require separate training and validation sets, we have appointed the first 60,000 images to be used for training and the remaining 10,000 for validation. In the [StyleGAN paper](http://stylegan.xyz/paper), however, we used all 70,000 images for training. We have explicitly made sure that there are no duplicate images in the dataset itself. However, please note that the `in-the-wild` folder may contain multiple copies of the same image in cases where we extracted several different faces from the same image. ## Download script You can either grab the data directly from Google Drive or use the provided [download script](./download_ffhq.py). The script makes things considerably easier by automatically downloading all the requested files, verifying their checksums, retrying each file several times on error, and employing multiple concurrent connections to maximize bandwidth. ``` > python download_ffhq.py -h usage: download_ffhq.py [-h] [-j] [-s] [-i] [-t] [-w] [-r] [-a] [--num_threads NUM] [--status_delay SEC] [--timing_window LEN] [--chunk_size KB] [--num_attempts NUM] Download Flickr-Face-HQ (FFHQ) dataset to current working directory. optional arguments: -h, --help show this help message and exit -j, --json download metadata as JSON (254 MB) -s, --stats print statistics about the dataset -i, --images download 1024x1024 images as PNG (89.1 GB) -t, --thumbs download 128x128 thumbnails as PNG (1.95 GB) -w, --wilds download in-the-wild images as PNG (955 GB) -r, --tfrecords download multi-resolution TFRecords (273 GB) -a, --align recreate 1024x1024 images from in-the-wild images --num_threads NUM number of concurrent download threads (default: 32) --status_delay SEC time between download status prints (default: 0.2) --timing_window LEN samples for estimating download eta (default: 50) --chunk_size KB chunk size for each download thread (default: 128) --num_attempts NUM number of download attempts per file (default: 10) ``` ``` > python ..\download_ffhq.py --json --images Downloading JSON metadata... \ 100.00% done 1/1 files 0.25/0.25 GB 43.21 MB/s ETA: done Parsing JSON metadata... Downloading 70000 files... | 100.00% done 70000/70000 files 89.19 GB/89.19 GB 59.87 MB/s ETA: done ``` The script also serves as a reference implementation of the automated scheme that we used to align and crop the images. Once you have downloaded the in-the-wild images with `python download_ffhq.py --wilds`, you can run `python download_ffhq.py --align` to reproduce exact replicas of the aligned 1024&times;1024 images using the facial landmark locations included in the metadata. ## Metadata The `ffhq-dataset-v1.json` file contains the following information for each image in a machine-readable format: ``` { "0": { # Image index "category": "training", # Training or validation "metadata": { # Info about the original Flickr photo: "photo_url": "https://www.flickr.com/photos/...", # - Flickr URL "photo_title": "DSCF0899.JPG", # - File name "author": "Jeremy Frumkin", # - Author "country": "", # - Country where the photo was taken "license": "Attribution-NonCommercial License", # - License name "license_url": "https://creativecommons.org/...", # - License detail URL "date_uploaded": "2007-08-16", # - Date when the photo was uploaded to Flickr "date_crawled": "2018-10-10" # - Date when the photo was crawled from Flickr }, "image": { # Info about the aligned 1024x1024 image: "file_url": "https://drive.google.com/...", # - Google Drive URL "file_path": "images1024x1024/00000.png", # - Google Drive path "file_size": 1488194, # - Size of the PNG file in bytes "file_md5": "ddeaeea6ce59569643715759d537fd1b", # - MD5 checksum of the PNG file "pixel_size": [1024, 1024], # - Image dimensions "pixel_md5": "47238b44dfb87644460cbdcc4607e289", # - MD5 checksum of the raw pixel data "face_landmarks": [...] # - 68 face landmarks reported by dlib }, "thumbnail": { # Info about the 128x128 thumbnail: "file_url": "https://drive.google.com/...", # - Google Drive URL "file_path": "thumbnails128x128/00000.png", # - Google Drive path "file_size": 29050, # - Size of the PNG file in bytes "file_md5": "bd3e40b2ba20f76b55dc282907b89cd1", # - MD5 checksum of the PNG file "pixel_size": [128, 128], # - Image dimensions "pixel_md5": "38d7e93eb9a796d0e65f8c64de8ba161" # - MD5 checksum of the raw pixel data }, "in_the_wild": { # Info about the in-the-wild image: "file_url": "https://drive.google.com/...", # - Google Drive URL "file_path": "in-the-wild-images/00000.png", # - Google Drive path "file_size": 3991569, # - Size of the PNG file in bytes "file_md5": "1dc0287e73e485efb0516a80ce9d42b4", # - MD5 checksum of the PNG file "pixel_size": [2016, 1512], # - Image dimensions "pixel_md5": "86b3470c42e33235d76b979161fb2327", # - MD5 checksum of the raw pixel data "face_rect": [667, 410, 1438, 1181], # - Axis-aligned rectangle of the face region "face_landmarks": [...], # - 68 face landmarks reported by dlib "face_quad": [...] # - Aligned quad of the face region } }, ... } ``` ## Acknowledgements We thank Jaakko Lehtinen, David Luebke, and Tuomas Kynk&auml;&auml;nniemi for in-depth discussions and helpful comments; Janne Hellsten, Tero Kuosmanen, and Pekka J&auml;nis for compute infrastructure and help with the code release. We also thank Vahid Kazemi and Josephine Sullivan for their work on automatic face detection and alignment that enabled us to collect the data in the first place: > **One Millisecond Face Alignment with an Ensemble of Regression Trees**<br> > Vahid Kazemi, Josephine Sullivan<br> > Proc. CVPR 2014<br> > https://www.cv-foundation.org/openaccess/content_cvpr_2014/papers/Kazemi_One_Millisecond_Face_2014_CVPR_paper.pdf
student/FFHQ
[ "region:us" ]
2022-04-16T04:21:25+00:00
{}
2022-04-16T05:24:36+00:00
c31370fdaed32ac5f64fed5ee6d7dd6397f5e47a
# Malay-TTS-Yasmin All notebooks and code related at https://github.com/huseinzol05/malaya-speech/tree/master/data/azure-tts ## Attributes ### Wiki and News - 24000 sample rate, super clean. - narrator `ms-MY-YasminNeural`. - approximate 99.4 hours. - Texts from Malay Wikipedia and News. - Sentences between 2 words and 20 words. ### Parliament - 24000 sample rate, super clean. - narrator `ms-MY-YasminNeural`. - approximate 142 hours. - Texts from Malaysia Malay Parliament. - Sentences between 2 words and 25 words. ## how-to ### Wiki and News 1. Download [populated-text.json](populated-text.json) and [tts-malay-yasmin.tar.gz](tts-malay-yasmin.tar.gz). 2. To get wav and transcript, ```python import json import soundfile as sf with open('populated-text.json') as fopen: texts = json.load(fopen) index = 0 text = texts[index] y, sr = sf.read(f'female/{index}.wav') ``` ### Parliament 1. Download [populated-parliament.json](populated-parliament.json) and [tts-malay-yasmin-parliament.tar.gz](tts-malay-yasmin-parliament.tar.gz). 2. To get wav and transcript, ```python import json import soundfile as sf with open('populated-parliament.json') as fopen: texts = json.load(fopen) index = 0 text = texts[index] y, sr = sf.read(f'female-parliament/{index}.wav') ```
huseinzol05/Malay-TTS-Yasmin
[ "region:us" ]
2022-04-16T08:28:21+00:00
{}
2022-04-25T05:21:15+00:00
61ff60cb76c1fabef90bc8013587f0fb6a4fa142
# Dataset Card for Top Quark Tagging ## Table of Contents - [Dataset Card for Top Quark Tagging](#dataset-card-for-top-quark-tagging) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://zenodo.org/record/2603256 - **Paper:** https://arxiv.org/abs/1902.09914 - **Point of Contact:** [Gregor Kasieczka](gregor.kasieczka@uni-hamburg.de) ### Dataset Summary Top Quark Tagging is a dataset of Monte Carlo simulated events produced by proton-proton collisions at the Large Hadron Collider. The top-quark signal and mixed quark-gluon background jets are produced with Pythia8 with its default tune for a center-of-mass energy of 14 TeV. Multiple interactions and pile-up are ignored. The leading 200 jet constituent four-momenta \\( (E, p_x, p_y, p_z) \\) are stored, with zero-padding applied to jets with fewer than 200 constituents. ### Supported Tasks and Leaderboards - `tabular-classification`: The dataset can be used to train a model for tabular binary classification, which consists in predicting whether an event is produced from a top signal or quark-gluon background. Success on this task is typically measured by achieving a *high* [accuracy](https://huggingface.co/metrics/accuracy) and AUC score. ## Dataset Structure ### Data Instances Each instance in the dataset consists of the four-momenta of the leading 200 jet constituents, sorted by \\(p_T\\). For jets with fewer than 200 constituents, zero-padding is applied. The four-momenta of the top-quark are also provided, along with a label in the `is_signal_new` column to indicate whether the event stems from a top-quark (1) or QCD background (0). An example instance looks as follows: ``` {'E_0': 474.0711364746094, 'PX_0': -250.34703063964844, 'PY_0': -223.65196228027344, 'PZ_0': -334.73809814453125, ... 'E_199': 0.0, 'PX_199': 0.0, 'PY_199': 0.0, 'PZ_199': 0.0, 'truthE': 0.0, 'truthPX': 0.0, 'truthPY': 0.0, 'truthPZ': 0.0, 'ttv': 0, 'is_signal_new': 0} ``` ### Data Fields The fields in the dataset have the following meaning: - `E_i`: the energy of jet constituent \\(i\\). - `PX_i`: the \\(x\\) component of the jet constituent's momentum - `PY_i`: the \\(y\\) component of the jet constituent's momentum - `PZ_i`: the \\(z\\) component of the jet constituent's momentum - `truthE`: the energy of the top-quark - `truthPX`: the \\(x\\) component of the top quark's momentum - `truthPY`: the \\(y\\) component of the top quark's momentum - `truthPZ`: the \\(z\\) component of the top quark's momentum - `ttv`: a flag that indicates which split (train, validation, or test) that a jet belongs to. Redundant since each split is provided as a separate dataset - `is_signal_new`: the label for each jet. A 1 indicates a top-quark, while a 0 indicates QCD background. ### Data Splits | | train | validation | test | |------------------|--------:|-----------:|-------:| | Number of events | 1211000 | 403000 | 404000 | ### Licensing Information This dataset is released under the [Creative Commons Attribution 4.0 International](https://creativecommons.org/licenses/by/4.0/legalcode) license. ### Citation Information ``` @dataset{kasieczka_gregor_2019_2603256, author = {Kasieczka, Gregor and Plehn, Tilman and Thompson, Jennifer and Russel, Michael}, title = {Top Quark Tagging Reference Dataset}, month = mar, year = 2019, publisher = {Zenodo}, version = {v0 (2018\_03\_27)}, doi = {10.5281/zenodo.2603256}, url = {https://doi.org/10.5281/zenodo.2603256} } ``` ### Contributions Thanks to [@lewtun](https://github.com/lewtun) for adding this dataset.
dl4phys/top_tagging
[ "license:cc-by-4.0", "arxiv:1902.09914", "region:us" ]
2022-04-16T08:53:34+00:00
{"license": "cc-by-4.0"}
2022-04-18T06:43:02+00:00
bb8e41b794e8da834d39efba3c090c9a1d30cbaa
dl4phys/top_tagging_nsubjettiness
[ "license:cc-by-4.0", "region:us" ]
2022-04-16T10:27:21+00:00
{"license": "cc-by-4.0"}
2022-04-16T15:27:05+00:00
69e53d567fb37a42b0756132dee8b7c11cbddc55
dl4phys/top_tagging_images
[ "license:cc-by-4.0", "region:us" ]
2022-04-16T12:08:03+00:00
{"license": "cc-by-4.0"}
2022-04-17T09:33:58+00:00
8a852d571fd838da91f2beb879f489f382d1462b
# PLOD: An Abbreviation Detection Dataset This is the repository for PLOD Dataset published at LREC 2022. The dataset can help build sequence labelling models for the task Abbreviation Detection. ### Dataset We provide two variants of our dataset - Filtered and Unfiltered. They are described in our paper here. 1. The Filtered version can be accessed via [Huggingface Datasets here](https://huggingface.co/datasets/surrey-nlp/PLOD-filtered) and a [CONLL format is present here](https://github.com/surrey-nlp/PLOD-AbbreviationDetection).<br/> 2. The Unfiltered version can be accessed via [Huggingface Datasets here](https://huggingface.co/datasets/surrey-nlp/PLOD-unfiltered) and a [CONLL format is present here](https://github.com/surrey-nlp/PLOD-AbbreviationDetection).<br/> 3. The [SDU Shared Task](https://sites.google.com/view/sdu-aaai22/home) data we use for zero-shot testing is [available here](https://huggingface.co/datasets/surrey-nlp/SDU-test). # Dataset Card for PLOD-filtered ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** [Needs More Information] - **Repository:** https://github.com/surrey-nlp/PLOD-AbbreviationDetection - **Paper:** https://arxiv.org/abs/2204.12061 - **Leaderboard:** https://paperswithcode.com/sota/abbreviationdetection-on-plod-filtered - **Point of Contact:** [Diptesh Kanojia](mailto:d.kanojia@surrey.ac.uk) ### Dataset Summary This PLOD Dataset is an English-language dataset of abbreviations and their long-forms tagged in text. The dataset has been collected for research from the PLOS journals indexing of abbreviations and long-forms in the text. This dataset was created to support the Natural Language Processing task of abbreviation detection and covers the scientific domain. ### Supported Tasks and Leaderboards This dataset primarily supports the Abbreviation Detection Task. It has also been tested on a train+dev split provided by the Acronym Detection Shared Task organized as a part of the Scientific Document Understanding (SDU) workshop at AAAI 2022. ### Languages English ## Dataset Structure ### Data Instances A typical data point comprises an ID, a set of `tokens` present in the text, a set of `pos_tags` for the corresponding tokens obtained via Spacy NER, and a set of `ner_tags` which are limited to `AC` for `Acronym` and `LF` for `long-forms`. An example from the dataset: {'id': '1', 'tokens': ['Study', '-', 'specific', 'risk', 'ratios', '(', 'RRs', ')', 'and', 'mean', 'BW', 'differences', 'were', 'calculated', 'using', 'linear', 'and', 'log', '-', 'binomial', 'regression', 'models', 'controlling', 'for', 'confounding', 'using', 'inverse', 'probability', 'of', 'treatment', 'weights', '(', 'IPTW', ')', 'truncated', 'at', 'the', '1st', 'and', '99th', 'percentiles', '.'], 'pos_tags': [8, 13, 0, 8, 8, 13, 12, 13, 5, 0, 12, 8, 3, 16, 16, 0, 5, 0, 13, 0, 8, 8, 16, 1, 8, 16, 0, 8, 1, 8, 8, 13, 12, 13, 16, 1, 6, 0, 5, 0, 8, 13], 'ner_tags': [0, 0, 0, 3, 4, 0, 1, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 3, 4, 4, 4, 4, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0] } ### Data Fields - id: the row identifier for the dataset point. - tokens: The tokens contained in the text. - pos_tags: the Part-of-Speech tags obtained for the corresponding token above from Spacy NER. - ner_tags: The tags for abbreviations and long-forms. ### Data Splits | | Train | Valid | Test | | ----- | ------ | ----- | ---- | | Filtered | 112652 | 24140 | 24140| | Unfiltered | 113860 | 24399 | 24399| ## Dataset Creation ### Source Data #### Initial Data Collection and Normalization Extracting the data from PLOS Journals online and then tokenization, normalization. #### Who are the source language producers? PLOS Journal ## Additional Information ### Dataset Curators The dataset was initially created by Leonardo Zilio, Hadeel Saadany, Prashant Sharma, Diptesh Kanojia, Constantin Orasan. ### Licensing Information CC-BY-SA 4.0 ### Citation Information [Needs More Information] ### Installation We use the custom NER pipeline in the [spaCy transformers](https://spacy.io/universe/project/spacy-transformers) library to train our models. This library supports training via any pre-trained language models available at the :rocket: [HuggingFace repository](https://huggingface.co/).<br/> Please see the instructions at these websites to setup your own custom training with our dataset to reproduce the experiments using Spacy. OR<br/> However, you can also reproduce the experiments via the Python notebook we [provide here](https://github.com/surrey-nlp/PLOD-AbbreviationDetection/blob/main/nbs/fine_tuning_abbr_det.ipynb) which uses HuggingFace Trainer class to perform the same experiments. The exact hyperparameters can be obtained from the models readme cards linked below. Before starting, please perform the following steps: ```bash git clone https://github.com/surrey-nlp/PLOD-AbbreviationDetection cd PLOD-AbbreviationDetection pip install -r requirements.txt ``` Now, you can use the notebook to reproduce the experiments. ### Model(s) Our best performing models are hosted on the HuggingFace models repository | Models | [`PLOD - Unfiltered`](https://huggingface.co/datasets/surrey-nlp/PLOD-unfiltered) | [`PLOD - Filtered`](https://huggingface.co/datasets/surrey-nlp/PLOD-filtered) | Description | | --- | :---: | :---: | --- | | [RoBERTa<sub>large</sub>](https://huggingface.co/roberta-large) | [RoBERTa<sub>large</sub>-finetuned-abbr](https://huggingface.co/surrey-nlp/roberta-large-finetuned-abbr) | -soon- | Fine-tuning on the RoBERTa<sub>large</sub> language model | | [RoBERTa<sub>base</sub>](https://huggingface.co/roberta-base) | -soon- | [RoBERTa<sub>base</sub>-finetuned-abbr](https://huggingface.co/surrey-nlp/roberta-large-finetuned-abbr) | Fine-tuning on the RoBERTa<sub>base</sub> language model | | [AlBERT<sub>large-v2</sub>](https://huggingface.co/albert-large-v2) | [AlBERT<sub>large-v2</sub>-finetuned-abbDet](https://huggingface.co/surrey-nlp/albert-large-v2-finetuned-abbDet) | -soon- | Fine-tuning on the AlBERT<sub>large-v2</sub> language model | On the link provided above, the model(s) can be used with the help of the Inference API via the web-browser itself. We have placed some examples with the API for testing.<br/> ### Usage You can use the HuggingFace Model link above to find the instructions for using this model in Python locally using the notebook provided in the Git repo.
surrey-nlp/PLOD-filtered
[ "task_categories:token-classification", "annotations_creators:Leonardo Zilio, Hadeel Saadany, Prashant Sharma, Diptesh Kanojia, Constantin Orasan", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:en", "license:cc-by-sa-4.0", "abbreviation-detection", "arxiv:2204.12061", "region:us" ]
2022-04-16T13:50:15+00:00
{"annotations_creators": ["Leonardo Zilio, Hadeel Saadany, Prashant Sharma, Diptesh Kanojia, Constantin Orasan"], "language_creators": ["found"], "language": ["en"], "license": "cc-by-sa-4.0", "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["token-classification"], "task_ids": [], "paperswithcode_id": "plod-filtered", "pretty_name": "PLOD: An Abbreviation Detection Dataset", "tags": ["abbreviation-detection"]}
2023-01-14T23:30:12+00:00
9caa35b94fe9c220e003d581e3b37e55012993b5
# PLOD: An Abbreviation Detection Dataset This is the repository for PLOD Dataset published at LREC 2022. The dataset can help build sequence labelling models for the task Abbreviation Detection. ### Dataset We provide two variants of our dataset - Filtered and Unfiltered. They are described in our paper here. 1. The Filtered version can be accessed via [Huggingface Datasets here](https://huggingface.co/datasets/surrey-nlp/PLOD-filtered) and a [CONLL format is present here](https://github.com/surrey-nlp/PLOD-AbbreviationDetection).<br/> 2. The Unfiltered version can be accessed via [Huggingface Datasets here](https://huggingface.co/datasets/surrey-nlp/PLOD-unfiltered) and a [CONLL format is present here](https://github.com/surrey-nlp/PLOD-AbbreviationDetection).<br/> 3. The [SDU Shared Task](https://sites.google.com/view/sdu-aaai22/home) data we use for zero-shot testing is [available here](https://huggingface.co/datasets/surrey-nlp/SDU-test). # Dataset Card for PLOD-unfiltered ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** [Needs More Information] - **Repository:** https://github.com/surrey-nlp/PLOD-AbbreviationDetection - **Paper:** https://arxiv.org/abs/2204.12061 - **Leaderboard:** https://paperswithcode.com/sota/abbreviationdetection-on-plod-an-abbreviation - **Point of Contact:** [Diptesh Kanojia](mailto:d.kanojia@surrey.ac.uk) ### Dataset Summary This PLOD Dataset is an English-language dataset of abbreviations and their long-forms tagged in text. The dataset has been collected for research from the PLOS journals indexing of abbreviations and long-forms in the text. This dataset was created to support the Natural Language Processing task of abbreviation detection and covers the scientific domain. ### Supported Tasks and Leaderboards This dataset primarily supports the Abbreviation Detection Task. It has also been tested on a train+dev split provided by the Acronym Detection Shared Task organized as a part of the Scientific Document Understanding (SDU) workshop at AAAI 2022. ### Languages English ## Dataset Structure ### Data Instances A typical data point comprises an ID, a set of `tokens` present in the text, a set of `pos_tags` for the corresponding tokens obtained via Spacy NER, and a set of `ner_tags` which are limited to `AC` for `Acronym` and `LF` for `long-forms`. An example from the dataset: {'id': '1', 'tokens': ['Study', '-', 'specific', 'risk', 'ratios', '(', 'RRs', ')', 'and', 'mean', 'BW', 'differences', 'were', 'calculated', 'using', 'linear', 'and', 'log', '-', 'binomial', 'regression', 'models', 'controlling', 'for', 'confounding', 'using', 'inverse', 'probability', 'of', 'treatment', 'weights', '(', 'IPTW', ')', 'truncated', 'at', 'the', '1st', 'and', '99th', 'percentiles', '.'], 'pos_tags': [8, 13, 0, 8, 8, 13, 12, 13, 5, 0, 12, 8, 3, 16, 16, 0, 5, 0, 13, 0, 8, 8, 16, 1, 8, 16, 0, 8, 1, 8, 8, 13, 12, 13, 16, 1, 6, 0, 5, 0, 8, 13], 'ner_tags': [0, 0, 0, 3, 4, 0, 1, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 3, 4, 4, 4, 4, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0] } ### Data Fields - id: the row identifier for the dataset point. - tokens: The tokens contained in the text. - pos_tags: the Part-of-Speech tags obtained for the corresponding token above from Spacy NER. - ner_tags: The tags for abbreviations and long-forms. ### Data Splits | | Train | Valid | Test | | ----- | ------ | ----- | ---- | | Filtered | 112652 | 24140 | 24140| | Unfiltered | 113860 | 24399 | 24399| ## Dataset Creation ### Source Data #### Initial Data Collection and Normalization Extracting the data from PLOS Journals online and then tokenization, normalization. #### Who are the source language producers? PLOS Journal ## Additional Information ### Dataset Curators The dataset was initially created by Leonardo Zilio, Hadeel Saadany, Prashant Sharma, Diptesh Kanojia, Constantin Orasan. ### Licensing Information CC-BY-SA 4.0 ### Citation Information [Needs More Information] ### Installation We use the custom NER pipeline in the [spaCy transformers](https://spacy.io/universe/project/spacy-transformers) library to train our models. This library supports training via any pre-trained language models available at the :rocket: [HuggingFace repository](https://huggingface.co/).<br/> Please see the instructions at these websites to setup your own custom training with our dataset to reproduce the experiments using Spacy. OR<br/> However, you can also reproduce the experiments via the Python notebook we [provide here](https://github.com/surrey-nlp/PLOD-AbbreviationDetection/blob/main/nbs/fine_tuning_abbr_det.ipynb) which uses HuggingFace Trainer class to perform the same experiments. The exact hyperparameters can be obtained from the models readme cards linked below. Before starting, please perform the following steps: ```bash git clone https://github.com/surrey-nlp/PLOD-AbbreviationDetection cd PLOD-AbbreviationDetection pip install -r requirements.txt ``` Now, you can use the notebook to reproduce the experiments. ### Model(s) Our best performing models are hosted on the HuggingFace models repository: | Models | [`PLOD - Unfiltered`](https://huggingface.co/datasets/surrey-nlp/PLOD-unfiltered) | [`PLOD - Filtered`](https://huggingface.co/datasets/surrey-nlp/PLOD-filtered) | Description | | --- | :---: | :---: | --- | | [RoBERTa<sub>large</sub>](https://huggingface.co/roberta-large) | [RoBERTa<sub>large</sub>-finetuned-abbr](https://huggingface.co/surrey-nlp/roberta-large-finetuned-abbr) | -soon- | Fine-tuning on the RoBERTa<sub>large</sub> language model | | [RoBERTa<sub>base</sub>](https://huggingface.co/roberta-base) | -soon- | [RoBERTa<sub>base</sub>-finetuned-abbr](https://huggingface.co/surrey-nlp/roberta-large-finetuned-abbr) | Fine-tuning on the RoBERTa<sub>base</sub> language model | | [AlBERT<sub>large-v2</sub>](https://huggingface.co/albert-large-v2) | [AlBERT<sub>large-v2</sub>-finetuned-abbDet](https://huggingface.co/surrey-nlp/albert-large-v2-finetuned-abbDet) | -soon- | Fine-tuning on the AlBERT<sub>large-v2</sub> language model | On the link provided above, the model(s) can be used with the help of the Inference API via the web-browser itself. We have placed some examples with the API for testing.<br/> ### Usage You can use the HuggingFace Model link above to find the instructions for using this model in Python locally using the notebook provided in the Git repo.
surrey-nlp/PLOD-unfiltered
[ "task_categories:token-classification", "annotations_creators:Leonardo Zilio, Hadeel Saadany, Prashant Sharma, Diptesh Kanojia, Constantin Orasan", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:en", "license:cc-by-sa-4.0", "abbreviation-detection", "arxiv:2204.12061", "region:us" ]
2022-04-16T17:49:49+00:00
{"annotations_creators": ["Leonardo Zilio, Hadeel Saadany, Prashant Sharma, Diptesh Kanojia, Constantin Orasan"], "language_creators": ["found"], "language": ["en"], "license": ["cc-by-sa-4.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["token-classification"], "task_ids": [], "paperswithcode_id": "plod-an-abbreviation-detection-dataset-for", "pretty_name": "PLOD: An Abbreviation Detection Dataset", "tags": ["abbreviation-detection"]}
2023-01-14T23:31:04+00:00
8158095eeaed506a0f8bd650eaf665d1cb30e33d
necm77/negotiation_data
[ "region:us" ]
2022-04-16T20:31:13+00:00
{}
2022-04-19T19:50:08+00:00
cc19b6f9f276ad7fc1d5d76870cdc605a2805e8b
Paercky/Tweets
[ "license:afl-3.0", "region:us" ]
2022-04-16T20:31:54+00:00
{"license": "afl-3.0"}
2022-04-16T20:33:23+00:00
1bb2f5816caf50002da4e7bd5ec845fec22eb4cd
# AutoTrain Dataset for project: Tweets ## Dataset Descritpion This dataset has been automatically processed by AutoTrain for project Tweets. ### Languages The BCP-47 code for the dataset's language is en. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json [ { "text": "So the mask mandate goes away the day after #Furnal2022 ends, and you know what will happen after th[...]", "target": 0 }, { "text": "@EwanMacKenna Also does anyone know whether Margaret Buttimer of Bandon is still in prison for the '[...]", "target": 1 } ] ``` ### Dataset Fields The dataset has the following fields (also called "features"): ```json { "text": "Value(dtype='string', id=None)", "target": "ClassLabel(num_classes=3, names=['1', '2', '3'], id=None)" } ``` ### Dataset Splits This dataset is split into a train and validation split. The split sizes are as follow: | Split name | Num samples | | ------------ | ------------------- | | train | 1679 | | valid | 420 |
Paercky/autotrain-data-Tweets
[ "task_categories:text-classification", "language:en", "region:us" ]
2022-04-16T20:46:11+00:00
{"language": ["en"], "task_categories": ["text-classification"]}
2022-10-25T09:08:35+00:00
a4d324be68761bd614dd3be85ccaac497001fabb
# Malay-TTS-Osman All notebooks and code related at https://github.com/huseinzol05/malaya-speech/tree/master/data/azure-tts ## Attributes ### Wiki and News - 24000 sample rate, super clean. - narrator `ms-MY-OsmanNeural`. - approximate 94.5 hours - Texts from Malay Wikipedia and News. - Sentences between 2 words and 20 words. ### Parliament - 24000 sample rate, super clean. - narrator `ms-MY-OsmanNeural`. - approximate 133.2 hours. - Texts from Malaysia Malay Parliament. - Sentences between 2 words and 25 words. ## how-to ### Wiki and News 1. Download [populated-text.json](populated-text.json) and [tts-malay-osman.tar.gz](tts-malay-osman.tar.gz). 2. To get wav and transcript, ```python import json import soundfile as sf with open('populated-text.json') as fopen: texts = json.load(fopen) index = 0 text = texts[index] y, sr = sf.read(f'male/{index}.wav') ``` ### Parliament 1. Download [populated-parliament.json](populated-parliament.json) and [tts-malay-osman-parliament.tar.gz](tts-malay-osman-parliament.tar.gz). 2. To get wav and transcript, ```python import json import soundfile as sf with open('populated-parliament.json') as fopen: texts = json.load(fopen) index = 0 text = texts[index] y, sr = sf.read(f'male-parliament/{index}.wav') ```
huseinzol05/Malay-TTS-Osman
[ "region:us" ]
2022-04-16T22:15:43+00:00
{}
2022-04-17T04:39:21+00:00
53997f64314d8360a2b2d23b7445ba47d4c66b9a
enimai/MuST-C-ru
[ "license:afl-3.0", "region:us" ]
2022-04-17T04:20:19+00:00
{"license": "afl-3.0"}
2022-04-17T04:23:38+00:00
7092c27872e919f31d0496fb8b9c47bd2cba3f6c
# Dataset Card for "IndicXNLI" ## Table of Contents - [Dataset Card for "IndicXNLI"](#dataset-card-for-indicxnli) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) ## Dataset Description - **Homepage:** <https://github.com/divyanshuaggarwal/IndicXNLI> - **Paper:** [IndicXNLI: Evaluating Multilingual Inference for Indian Languages](https://arxiv.org/abs/2204.08776) - **Point of Contact:** [Divyanshu Aggarwal](mailto:divyanshuggrwl@gmail.com) ### Dataset Summary INDICXNLI is similar to existing XNLI dataset in shape/form, but focusses on Indic language family. INDICXNLI include NLI data for eleven major Indic languages that includes Assamese (‘as’), Gujarat (‘gu’), Kannada (‘kn’), Malayalam (‘ml’), Marathi (‘mr’), Odia (‘or’), Punjabi (‘pa’), Tamil (‘ta’), Telugu (‘te’), Hindi (‘hi’), and Bengali (‘bn’). ### Supported Tasks and Leaderboards **Tasks:** Natural Language Inference **Leaderboards:** Currently there is no Leaderboard for this dataset. ### Languages - `Assamese (as)` - `Bengali (bn)` - `Gujarati (gu)` - `Kannada (kn)` - `Hindi (hi)` - `Malayalam (ml)` - `Marathi (mr)` - `Oriya (or)` - `Punjabi (pa)` - `Tamil (ta)` - `Telugu (te)` ## Dataset Structure ### Data Instances One example from the `hi` dataset is given below in JSON format. ```python {'premise': 'अवधारणात्मक रूप से क्रीम स्किमिंग के दो बुनियादी आयाम हैं-उत्पाद और भूगोल।', 'hypothesis': 'उत्पाद और भूगोल क्रीम स्किमिंग का काम करते हैं।', 'label': 1 (neutral) } ``` ### Data Fields - `premise (string)`: Premise Sentence - `hypothesis (string)`: Hypothesis Sentence - `label (integer)`: Integer label `0` if hypothesis `entails` the premise, `2` if hypothesis `negates` the premise and `1` otherwise. ### Data Splits <!-- Below is the dataset split given for `hi` dataset. ```python DatasetDict({ train: Dataset({ features: ['premise', 'hypothesis', 'label'], num_rows: 392702 }) test: Dataset({ features: ['premise', 'hypothesis', 'label'], num_rows: 5010 }) validation: Dataset({ features: ['premise', 'hypothesis', 'label'], num_rows: 2490 }) }) ``` --> Language | ISO 639-1 Code |Train | Test | Dev | --------------|----------------|-------|-----|------| Assamese | as | 392,702 | 5,010 | 2,490 | Bengali | bn | 392,702 | 5,010 | 2,490 | Gujarati | gu | 392,702 | 5,010 | 2,490 | Hindi | hi | 392,702 | 5,010 | 2,490 | Kannada | kn | 392,702 | 5,010 | 2,490 | Malayalam | ml |392,702 | 5,010 | 2,490 | Marathi | mr |392,702 | 5,010 | 2,490 | Oriya | or | 392,702 | 5,010 | 2,490 | Punjabi | pa | 392,702 | 5,010 | 2,490 | Tamil | ta | 392,702 | 5,010 | 2,490 | Telugu | te | 392,702 | 5,010 | 2,490 | <!-- The dataset split remains same across all languages. --> ## Dataset usage Code snippet for using the dataset using datasets library. ```python from datasets import load_dataset dataset = load_dataset("Divyanshu/indicxnli") ``` ## Dataset Creation Machine translation of XNLI english dataset to 11 listed Indic Languages. ### Curation Rationale [More information needed] ### Source Data [XNLI dataset](https://cims.nyu.edu/~sbowman/xnli/) #### Initial Data Collection and Normalization [Detailed in the paper](https://arxiv.org/abs/2204.08776) #### Who are the source language producers? [Detailed in the paper](https://arxiv.org/abs/2204.08776) #### Human Verification Process [Detailed in the paper](https://arxiv.org/abs/2204.08776) ## Considerations for Using the Data ### Social Impact of Dataset [Detailed in the paper](https://arxiv.org/abs/2204.08776) ### Discussion of Biases [Detailed in the paper](https://arxiv.org/abs/2204.08776) ### Other Known Limitations [Detailed in the paper](https://arxiv.org/abs/2204.08776) ### Dataset Curators Divyanshu Aggarwal, Vivek Gupta, Anoop Kunchukuttan ### Licensing Information Contents of this repository are restricted to only non-commercial research purposes under the [Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0)](https://creativecommons.org/licenses/by-nc/4.0/). Copyright of the dataset contents belongs to the original copyright holders. ### Citation Information If you use any of the datasets, models or code modules, please cite the following paper: ``` @misc{https://doi.org/10.48550/arxiv.2204.08776, doi = {10.48550/ARXIV.2204.08776}, url = {https://arxiv.org/abs/2204.08776}, author = {Aggarwal, Divyanshu and Gupta, Vivek and Kunchukuttan, Anoop}, keywords = {Computation and Language (cs.CL), Artificial Intelligence (cs.AI), FOS: Computer and information sciences, FOS: Computer and information sciences}, title = {IndicXNLI: Evaluating Multilingual Inference for Indian Languages}, publisher = {arXiv}, year = {2022}, copyright = {Creative Commons Attribution 4.0 International} } ``` <!-- ### Contributions -->
Divyanshu/indicxnli
[ "task_categories:text-classification", "task_ids:natural-language-inference", "annotations_creators:machine-generated", "language_creators:machine-generated", "multilinguality:multilingual", "size_categories:1M<n<10M", "source_datasets:original", "language:as", "language:bn", "language:gu", "language:hi", "language:kn", "language:ml", "language:mr", "language:or", "language:pa", "language:ta", "language:te", "license:cc0-1.0", "arxiv:2204.08776", "region:us" ]
2022-04-17T16:48:10+00:00
{"annotations_creators": ["machine-generated"], "language_creators": ["machine-generated"], "language": ["as", "bn", "gu", "hi", "kn", "ml", "mr", "or", "pa", "ta", "te"], "license": ["cc0-1.0"], "multilinguality": ["multilingual"], "size_categories": ["1M<n<10M"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["natural-language-inference"], "pretty_name": "IndicXNLI"}
2022-10-06T14:26:00+00:00
539315f30d0cfd1aa5765a65a4dcd6d93d168d20
ዛጎል ዜና- መንግስት አምስት ሺህ የሚጠጉ እስረኞችን “ተመራቂዎች” በሚል መፍታቱን ይፋ ባደረገበት ቀን በተመሳሳይ አምቦ ተማሪዎች ተቃውሞ ማሰማታቸው ተሰማ። ተማሪዎቹ የአስቸኳይ አዋጁን በመጣስ ” መረራ ይፈታ” እያሉ ተቃውሞ መጀመራቸው ነው የተሰማው። ከትምህርት ቤት ወደ ትምህርት ቤት የሰፋው ተቃውሞ ብህይወት ላይ አደጋ ባያስከትልም በንብረት ላይ ግን ጉዳት አድርሷል። መኪና ሲቃጠል ያዩ የአይን ምስክሮች ተቃውሞውን በጀመሩት ላይም ሆነ ዘግይተው በተቀላቀሉት ላይ እንደ ቀደሞው ያለ የሃይል እርምጃ አልተወሰደም። የኦሮሚያ ሚዲያ ኔት ወርክ እንዳለው ደግሞ በርካታ ሰዎች ታስረዋል። ለወትሮው ህገ መንግስቱን በሃይል ለመናድ የተነሱ፣ የነውጥ ሃይሎች፣ አተራማሾች፣ የጥፋት ሃይል ተላላኪዎች በሚል ተጠርጥረው በቁጥጥር ስር ከዋሉት መካከል 4035 የሚሆኑት ሲፈቱ እስረኞቹ “ስድስት ኮርስ ወስደው ተመረቁ” ነው የተባለው። የኦሮሚያ ማረሚያ ቤቶች አስተዳደር ኮሚሽነር ፀሃይ በላይን ጠቅሶ ፋና እንደዘገበው ጦላይ ተሃድሶ ማዕከል ከገቡ 5 ሺህ 600 ሰልጣኞች መካከል 4035 ያህሉ በስድስት ዋና ዋና ጉዳዮች ሥልጠና ወስደው ተመርቀዋል። ኮርሶቹም በፍፁም፣ አይደገምም፣ የቀለም አብዮት፣ የኢትዮጰያ ህገ–መንግስት እና የኢትዮጵያ ህዳሴ የሚሉ ርዕሰ ጉዳዮችን የተካተቱባቸው ነው። አበምርቃቱ ላይ ጠቅላይ ሚኒስትር ሃይለማርያም ተገኝተው “ ሽኝት” አደርጉላቸው ተብሏል። በርካታ ቃል ተገብቶላቸዋል። መስመርም ተሰምሮላቸዋል። “በደምና በአጥንት የተጻፈውን ሕገመንግስት፣ ዋጋ የተከፈለበትን ህገመንግስት” በማለት አቶ ሃይለማርያም በሃይል ለመናድ መሞከር አይቻልም በለዋል። “ ልክ እናንተ አይደገምም እንዳላችሁት፣ እኛም አይደገም እንላለን” ብለዋል። የፋና ዘገባ እንዲህ ይነበባል። አዲስ አበባ ፣ ታህሳስ 12 ፣ 2009 (ኤፍ ቢ ሲ) በሃገሪቱ የተለያዩ አካባቢዎች በተፈጠረው ሁከት ውስጥ ተሳትፈው በማሰልጠኛ ጣቢያዎች የተሃድሶ ስልጠና ሲወስዱ የነበሩ ዜጎች ወደ መጡበት እየተመለሱ ነው። በአዋሽ፣ አላጌና ብር ሸለቆ ማዕከላት የተሃድሶ ስልጠና የወሰዱ ዜጎች ናቸው ወደ አካባቢያቸው እየተመለሱ ያሉት። በጦላይ ለአንድ ወር የተሃድሶ ስልጠና የወሰዱ 4 ሺህ 35 ዜጎችም ሥልጠናቸውን አጠናቀው ነገ ወደ መጡበት አካባቢ ይመለሳሉ ተብሏል። በጦላይ የተሃድሶ ማዕከል የተገኙት ጠቅላይ ሚኒስትር ኃይለማርያም ደሳለኝ በዚሁ ጊዜ ባስተላለፉት መልዕክት ሰልጣኞች ወደ መደበኛ ህይወታቸው እንዲመለሱ መንግሥት ድጋፍ ያደርጋል ብለዋል። ሠራተኞች ወደ ሥራ ገበታቸው እንዲመለሱ የሚደረግ ሲሆን ተማሪዎች ደግሞ ትምህርታቸው እንዲቀጥሉ ይደረጋልም ነው ያሉት ጠቅላይ ሚኒስትር ኃይለማርያም። ሥራ አጥ የሆኑ ወጣቶችም በራሳቸው መንገድ ሥራ እንዲፈጥሩ ድጋፍ እንደሚደረግላቸው ጠቅላይ ሚኒስትሩ ገልጸዋል። ሠላም፣ ልማትና ዴሞክራሲ የማይነጣጡ የአንድ አገር ህልውና መሰረት መሆናቸውን ወጣቱ ተገንዝቦ እነዚህን እሴቶች የመጠበቅ ኃላፊነቱን እንዲወጣ ጠይቀዋል። ወጣቱ ጥያቄ እንኳ ቢኖረው ሕገ-መንግሥቱ በሚፈቅደው መሰረት የማቅረብና መልስ የማግኘት መብት እንዳለው ገልጸዋል። ባለፉት ወራት እንደታየው ጥያቄውን በአመጽና ግርግር መጠየቁ ዋጋ እንዳስከፈለ ለማሳያነት በማንሳት። እንዲህ ዓይነት ሁኔታ እንዳይደገም መንግሥትም የራሱን ስህተት ለማረም ጥልቅ ተሃድሶ እያደረገ መሆኑን ገልጸው ወጣቱም የራሱን ስህተት በማረም ከመንግሥት ጋር በመሆን ሠላሙን እንዲጠብቅ መልዕክት አስተላልፈዋል። የኦሮሚያ ክልል ርዕሰ መስተዳደር አቶ ለማ መገርሳ በበኩላቸው በክልሉ የሰፈነውን ሠላም ለማስቀጠል ከሁሉም የህብረተሰብ ክፍል ጋር በቅንጅት ሥራዎች ይሰራሉ ብለዋል። ከወራት በፊት በተፈጠረው ሁከትና ግርግር ህይወት የጠፋ መሆኑን ገልጸው ለዘመናት የተለፋባቸው የህዝብ ኃብቶችም መውደማቸው አግባብ አለመሆኑን ተናግረዋል። ክልሉ ሊለወጥና ሊለማ የሚችለው የክልሉ ወጣቶች ለሠላም በጋራ ዘብ ሲቆሙ እንደሆነም አስምረውበታል። አሁን ወደ
surafelkindu/Amharic_corpus
[ "license:mit", "region:us" ]
2022-04-17T17:06:43+00:00
{"license": "mit"}
2022-04-17T17:19:47+00:00
7cffe68258589932209921818b9f9e56324850e3
oLMpics README
KevinZ/oLMpics
[ "region:us" ]
2022-04-18T01:14:53+00:00
{}
2022-04-19T17:08:06+00:00
976943672aec93d411e31de606fc103e4aa6073b
鸟类400.物种图像分类 58388训练集,2000测试测试集,2000验证图像224X224X3 jpg格式 400种鸟类的数据集。58388张训练图像、2000张测试图像(每种5张图像)和2000张验证图像(每种5张图像)。这是一个非常高质量的数据集,每张图像中只有一只鸟,鸟通常占据图像中至少50%的像素。因此,即使是一个中等复杂的模型也能在90%的范围内实现训练和测试精度。 所有图像均为jpg格式的224 X 224 X 3彩色图像。数据集包括列车集、测试集和验证集。每套包含400个子目录,每种鸟类一个。如果使用Keras ImageDataGenerator,则数据结构非常方便。flowfromdirectory创建列车、测试和有效数据生成器。数据集还包括一个鸟类物种档案。csv。此cvs文件包含三列。“文件路径”列包含图像文件的文件路径。“标签”列包含与图像文件关联的类名。鸟类种类。如果使用df=pandas读入csv文件。birdscsv(Bird Species.csv)将创建一个pandas数据帧,然后可以将其拆分为traindf、testdf和validdf数据帧,以创建您自己的数据划分为train、test和validdf数据集。 注:数据集中的测试和验证图像是手工选择的“最佳”图像,因此使用这些数据集与创建自己的测试和验证集相比,您的模型可能会获得最高的准确度分数。然而,就看不见的图像上的模型性能而言,后一种情况更为准确。 这些图片是通过网络搜索按物种名称收集的。下载一个物种的图像文件后,使用我开发的python duplicate image detector程序检查其重复图像。删除所有检测到的重复项,以防止它们在训练集、测试集和验证集之间成为共同的图像。 之后,对图像进行裁剪,使鸟占据图像中至少50%的像素。然后,这些图像以jpg格式调整为224x224 X3。裁剪确保了当CNN对其进行处理时,图像中有足够的信息来创建高度准确的分类器。即使是一个中等稳健的模型,也应在高90%的范围内实现训练、验证和测试精度。由于数据集很大,我建议您尝试使用150 X 150 X3的模型和图像大小进行训练,以减少训练时间。所有文件也从每个物种的一个开始按顺序编号。所以测试图像被命名为1。jpg至5。jpg。对于验证图像也是如此。训练图像也用“零”填充顺序编号。例如001。jpg,002。jpg…010。jpg,011。jpg…。。099.jpg,100jpg,102。当与python文件函数和目录中的Keras流一起使用时,zero的填充保留了文件顺序。 训练集是不平衡的,每个物种有不同数量的文件。然而,每个物种至少有120个训练图像文件。这种不平衡并没有影响我的内核分类器,因为它在测试集上达到了98%以上的准确率。 数据集中一个显著的不平衡是雄性物种图像与雌性物种图像的比例。大约85%的图片是男性的,15%是女性的。典型的雄性动物的肤色要多样化得多,而一个物种的雌性动物通常是平淡无奇的。因此,男性和女性的形象可能看起来完全不同。几乎所有的测试和验证图像都来自该物种的雄性。因此,分类器可能无法在雌性物种图像上表现良好。
student/birds_400
[ "region:us" ]
2022-04-18T02:13:37+00:00
{}
2022-04-18T02:15:55+00:00
254dd05ce1dd064b434436fb491836b3b489fd9b
CUB200-2011数据集介绍: 该数据集由加州理工学院再2010年提出的细粒度数据集,也是目前细粒度分类识别研究的基准图像数据集。 该数据集共有11788张鸟类图像,包含200类鸟类子类,其中训练数据集有5994张图像,测试集有5794张图像,每张图像均提供了图像类标记信息,图像中鸟的bounding box,鸟的关键part信息,以及鸟类的属性信息,数据集如下图所示。 下载的数据集中,包含了如下文件: bounding_boxes.txt;classes.txt;image_class_labels.txt; images.txt; train_test_split.txt. 其中,bounding_boxes.txt为图像中鸟类的边界框信息;classes.txt为鸟类的类别信息,共有200类; image_class_labels.txt为图像标签和所属类别标签信息;images.txt为图像的标签和图像路径信息;train_test_split.txt为训练集和测试集划分。 本博客主要是根据train_test_split.txt文件和images.txt文件将原始下载的CUB200-2011数据集划分为训练集和测试集。在深度学习Pytorch框架下采用ImageFolder和DataLoader读取数据集较为方便。相关的python代码如下: (1) CUB200-2011训练集和测试集划分代码 # *_*coding: utf-8 *_* # author --liming-- """ 读取images.txt文件,获得每个图像的标签 读取train_test_split.txt文件,获取每个图像的train, test标签.其中1为训练,0为测试. """ import os import shutil import numpy as np import config import time time_start = time.time() # 文件路径 path_images = config.path + 'images.txt' path_split = config.path + 'train_test_split.txt' trian_save_path = config.path + 'dataset/train/' test_save_path = config.path + 'dataset/test/' # 读取images.txt文件 images = [] with open(path_images,'r') as f: for line in f: images.append(list(line.strip('\n').split(','))) # 读取train_test_split.txt文件 split = [] with open(path_split, 'r') as f_: for line in f_: split.append(list(line.strip('\n').split(','))) # 划分 num = len(images) # 图像的总个数 for k in range(num): file_name = images[k][0].split(' ')[1].split('/')[0] aaa = int(split[k][0][-1]) if int(split[k][0][-1]) == 1: # 划分到训练集 #判断文件夹是否存在 if os.path.isdir(trian_save_path + file_name): shutil.copy(config.path + 'images/' + images[k][0].split(' ')[1], trian_save_path+file_name+'/'+images[k][0].split(' ')[1].split('/')[1]) else: os.makedirs(trian_save_path + file_name) shutil.copy(config.path + 'images/' + images[k][0].split(' ')[1], trian_save_path + file_name + '/' + images[k][0].split(' ')[1].split('/')[1]) print('%s处理完毕!' % images[k][0].split(' ')[1].split('/')[1]) else: #判断文件夹是否存在 if os.path.isdir(test_save_path + file_name): aaaa = config.path + 'images/' + images[k][0].split(' ')[1] bbbb = test_save_path+file_name+'/'+images[k][0].split(' ')[1] shutil.copy(config.path + 'images/' + images[k][0].split(' ')[1], test_save_path+file_name+'/'+images[k][0].split(' ')[1].split('/')[1]) else: os.makedirs(test_save_path + file_name) shutil.copy(config.path + 'images/' + images[k][0].split(' ')[1], test_save_path + file_name + '/' + images[k][0].split(' ')[1].split('/')[1]) print('%s处理完毕!' % images[k][0].split(' ')[1].split('/')[1]) time_end = time.time() print('CUB200训练集和测试集划分完毕, 耗时%s!!' % (time_end - time_start)) config文件 # *_*coding: utf-8 *_* # author --liming-- path = '/media/lm/C3F680DFF08EB695/细粒度数据集/birds/CUB200/CUB_200_2011/' ROOT_TRAIN = path + 'images/train/' ROOT_TEST = path + 'images/test/' BATCH_SIZE = 16 (2) 利用Pytorch方式读取数据 # *_*coding: utf-8 *_* # author --liming-- """ 用于已下载数据集的转换,便于pytorch的读取 """ import torch import torchvision import config from torchvision import datasets, transforms data_transform = transforms.Compose([ transforms.Resize(224), transforms.ToTensor(), transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]) ]) def train_data_load(): # 训练集 root_train = config.ROOT_TRAIN train_dataset = torchvision.datasets.ImageFolder(root_train, transform=data_transform) CLASS = train_dataset.class_to_idx print('训练数据label与文件名的关系:', CLASS) train_loader = torch.utils.data.DataLoader(train_dataset, batch_size=config.BATCH_SIZE, shuffle=True) return CLASS, train_loader def test_data_load(): # 测试集 root_test = config.ROOT_TEST test_dataset = torchvision.datasets.ImageFolder(root_test, transform=data_transform) CLASS = test_dataset.class_to_idx print('测试数据label与文件名的关系:',CLASS) test_loader = torch.utils.data.DataLoader(test_dataset, batch_size=config.BATCH_SIZE, shuffle=True) return CLASS, test_loader if __name__ == '__main___': train_data_load() test_data_load()
student/CUB_birds_200_2011
[ "region:us" ]
2022-04-18T02:19:41+00:00
{}
2022-04-18T02:21:03+00:00
a125caa5d8ea8d27952f7fbe1b7f187ddf6a39bb
# Citation If you use the dataset, please cite the paper: @article{10.1007/s10579-021-09568-y, year = {2022}, title = {{Abstractive text summarization and new large-scale datasets for agglutinative languages Turkish and Hungarian}}, author = {Baykara, Batuhan and Güngör, Tunga}, journal = {Language Resources and Evaluation}, issn = {1574-020X}, doi = {10.1007/s10579-021-09568-y}, pages = {1--35}}
batubayk/TR-News
[ "task_categories:summarization", "task_categories:text-classification", "task_categories:text-generation", "task_categories:text2text-generation", "size_categories:100K<n<1M", "language:tr", "region:us" ]
2022-04-18T16:23:02+00:00
{"language": ["tr"], "size_categories": ["100K<n<1M"], "task_categories": ["summarization", "text-classification", "text-generation", "text2text-generation"], "pretty_name": "TR-News"}
2023-03-04T22:39:35+00:00
08d5a00cd764b7116bbc8422818fdfa57c7b413b
# Citation If you use the dataset, please cite the paper: @article{10.1007/s10579-021-09568-y, year = {2022}, title = {{Abstractive text summarization and new large-scale datasets for agglutinative languages Turkish and Hungarian}}, author = {Baykara, Batuhan and Güngör, Tunga}, journal = {Language Resources and Evaluation}, issn = {1574-020X}, doi = {10.1007/s10579-021-09568-y}, pages = {1--35}}
batubayk/HU-News
[ "task_categories:summarization", "task_categories:text-classification", "task_categories:text-generation", "task_categories:text2text-generation", "size_categories:100K<n<1M", "language:hu", "region:us" ]
2022-04-18T16:23:27+00:00
{"language": ["hu"], "size_categories": ["100K<n<1M"], "task_categories": ["summarization", "text-classification", "text-generation", "text2text-generation"], "pretty_name": "HU-News"}
2023-03-04T22:40:26+00:00
13e24921a5d4e04d6ab4bf21fada26c530e3db1f
# Tweet Emotion Intensity Dataset ## Papers: * Emotion Intensities in Tweets. Saif M. Mohammad and Felipe Bravo-Marquez. In Proceedings of the sixth joint conference on lexical and computational semantics (*Sem), August 2017, Vancouver, Canada. * WASSA-2017 Shared Task on Emotion Intensity. Saif M. Mohammad and Felipe Bravo-Marquez. In Proceedings of the EMNLP 2017 Workshop on Computational Approaches to Subjectivity, Sentiment, and Social Media (WASSA), September 2017, Copenhagen, Denmark.
stepp1/tweet_emotion_intensity
[ "region:us" ]
2022-04-18T16:32:33+00:00
{}
2022-04-18T19:49:56+00:00
e078606bba4ff4735ffac758f8fb5e9d9045e4ba
# Dataset Card for the-reddit-irl-dataset ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Licensing Information](#licensing-information) ## Dataset Description - **Homepage:** [https://socialgrep.com/datasets](https://socialgrep.com/datasets/the-reddit-irl-dataset?utm_source=huggingface&utm_medium=link&utm_campaign=theredditirldataset) - **Point of Contact:** [Website](https://socialgrep.com/contact?utm_source=huggingface&utm_medium=link&utm_campaign=theredditirldataset) ### Dataset Summary Data from the humour subreddits /r/meirl and /r/me_irl, up to Apr 1 2022. ### Languages Mainly English. ## Dataset Structure ### Data Instances A data point is a post or a comment. Due to the separate nature of the two, those exist in two different files - even though many fields are shared. ### Data Fields - 'type': the type of the data point. Can be 'post' or 'comment'. - 'id': the base-36 Reddit ID of the data point. Unique when combined with type. - 'subreddit.id': the base-36 Reddit ID of the data point's host subreddit. Unique. - 'subreddit.name': the human-readable name of the data point's host subreddit. - 'subreddit.nsfw': a boolean marking the data point's host subreddit as NSFW or not. - 'created_utc': a UTC timestamp for the data point. - 'permalink': a reference link to the data point on Reddit. - 'score': score of the data point on Reddit. - 'domain': (Post only) the domain of the data point's link. - 'url': (Post only) the destination of the data point's link, if any. - 'selftext': (Post only) the self-text of the data point, if any. - 'title': (Post only) the title of the post data point. - 'body': (Comment only) the body of the comment data point. - 'sentiment': (Comment only) the result of an in-house sentiment analysis pipeline. Used for exploratory analysis. ## Additional Information ### Licensing Information CC-BY v4.0
SocialGrep/the-reddit-irl-dataset
[ "annotations_creators:lexyr", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:10M<n<100M", "source_datasets:original", "language:en", "license:cc-by-4.0", "region:us" ]
2022-04-18T18:18:54+00:00
{"annotations_creators": ["lexyr"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10M<n<100M"], "source_datasets": ["original"]}
2022-07-01T16:52:22+00:00
4544b33bcc8077384d02b422f91b5723b890f53c
# Dataset Card for "squad" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://rajpurkar.github.io/SQuAD-explorer/](https://rajpurkar.github.io/SQuAD-explorer/) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 33.51 MB - **Size of the generated dataset:** 85.75 MB - **Total amount of disk used:** 119.27 MB ### Dataset Summary Stanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by crowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span, from the corresponding reading passage, or the question might be unanswerable. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure We show detailed information for up to 5 configurations of the dataset. ### Data Instances #### plain_text - **Size of downloaded dataset files:** 33.51 MB - **Size of the generated dataset:** 85.75 MB - **Total amount of disk used:** 119.27 MB An example of 'train' looks as follows. ``` { "answers": { "answer_start": [1], "text": ["This is a test text"] }, "context": "This is a test context.", "id": 1, "question": "Is this a test?", } ``` ### Data Fields The data fields are the same among all splits. #### plain_text - `id`: a `int32` feature. - `context`: a `string` feature. - `question`: a `string` feature. - `answers`: a dictionary feature containing: - `text`: a `string` feature. - `answer_start`: a `int32` feature. ### Data Splits | name |train|validation| |----------|----:|---------:| |plain_text|---| ---| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ``` ### Contributions Thanks to [@lewtun](https://github.com/lewtun), [@albertvillanova](https://github.com/albertvillanova), [@patrickvonplaten](https://github.com/patrickvonplaten), [@thomwolf](https://github.com/thomwolf) for adding this dataset.
Lexi/spanextract
[ "task_categories:question-answering", "task_ids:extractive-qa", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:extended|wikipedia", "language:en", "license:cc-by-4.0", "region:us" ]
2022-04-18T19:06:21+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced", "found"], "language": ["en"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["extended|wikipedia"], "task_categories": ["question-answering"], "task_ids": ["extractive-qa"], "paperswithcode_id": "squad", "pretty_name": "SQuAD"}
2022-10-25T09:08:42+00:00
f549503ef55b76632d02538edb6987ea9f96f82c
Paper: [Read, Revise, Repeat: A System Demonstration for Human-in-the-loop Iterative Text Revision](https://arxiv.org/abs/2204.03685) Authors: Wanyu Du*, Zae Myung Kim*, Vipul Raheja, Dhruv Kumar, Dongyeop Kang Github repo: https://github.com/vipulraheja/IteraTeR Watch our system demonstration below! [![demo](https://yt-embed.herokuapp.com/embed?v=lK08tIpEoaE)](https://www.youtube.com/watch?v=lK08tIpEoaE)
wanyu/IteraTeR_v2
[ "task_categories:text2text-generation", "annotations_creators:crowdsourced", "language_creators:found", "multilinguality:monolingual", "source_datasets:original", "language:en", "license:apache-2.0", "conditional-text-generation", "text-editing", "arxiv:2204.03685", "region:us" ]
2022-04-18T19:09:17+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["found"], "language": ["en"], "license": ["apache-2.0"], "multilinguality": ["monolingual"], "source_datasets": ["original"], "task_categories": ["text2text-generation"], "task_ids": [], "pretty_name": "IteraTeR_v2", "language_bcp47": ["en-US"], "tags": ["conditional-text-generation", "text-editing"]}
2022-10-24T17:58:08+00:00
6572d608a26a8f4fa2715ae788cb52b343e1e64d
ciroy/mlcommons-test
[ "license:cc-by-4.0", "region:us" ]
2022-04-18T19:36:47+00:00
{"license": "cc-by-4.0"}
2022-04-18T19:36:47+00:00
77396f1031920d1c116ec6d63ac397ff6aa492d3
## Dataset Description - **Homepage:** [Cat and Dog](https://www.kaggle.com/datasets/tongpython/cat-and-dog) - **Download Size** 217.30 MiB - **Generated Size** 198.89 MiB - **Total Size** 416.20 MiB ### Dataset Summary A dataset from [kaggle](https://www.kaggle.com/datasets/tongpython/cat-and-dog) with duplicate data removed. ### Data Fields The data instances have the following fields: - `image`: A `PIL.Image.Image` object containing the image. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]`. - `labels`: an `int` classification label. ### Class Label Mappings: ``` { "cat": 0, "dog": 1, } ``` ### Data Splits | | train | test | |---------------|-------|-----:| | # of examples | 8000 | 2000 | ```python >>> from datasets import load_dataset >>> dataset = load_dataset("Bingsu/Cat_and_Dog") >>> dataset DatasetDict({ train: Dataset({ features: ['image', 'labels'], num_rows: 8000 }) test: Dataset({ features: ['image', 'labels'], num_rows: 2000 }) }) >>> dataset["train"].features {'image': Image(decode=True, id=None), 'labels': ClassLabel(num_classes=2, names=['cat', 'dog'], id=None)} ```
Bingsu/Cat_and_Dog
[ "task_categories:image-classification", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:cc0-1.0", "region:us" ]
2022-04-19T01:23:06+00:00
{"language": ["en"], "license": ["cc0-1.0"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["image-classification"], "pretty_name": "Cat and Dog", "dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "labels", "dtype": {"class_label": {"names": {"0": "cat", "1": "dog"}}}}], "splits": [{"name": "train", "num_bytes": 166451650.0, "num_examples": 8000}, {"name": "test", "num_bytes": 42101650.0, "num_examples": 2000}], "download_size": 227859268, "dataset_size": 208553300.0, "size_in_bytes": 436412568.0}}
2023-01-26T10:48:25+00:00
48fdfd7ab1dbc1a62e4e8a8b9f4c360259d51d3c
## Dataset Description - **Homepage:** [Korean Single Speaker Speech Dataset](https://www.kaggle.com/datasets/bryanpark/korean-single-speaker-speech-dataset) - **Repository:** [Kyubyong/kss](https://github.com/Kyubyong/kss) - **Paper:** N/A - **Leaderboard:** N/A - **Point of Contact:** N/A # Description of the original author ### KSS Dataset: Korean Single speaker Speech Dataset KSS Dataset is designed for the Korean text-to-speech task. It consists of audio files recorded by a professional female voice actoress and their aligned text extracted from my books. As a copyright holder, by courtesy of the publishers, I release this dataset to the public. To my best knowledge, this is the first publicly available speech dataset for Korean. ### File Format Each line in `transcript.v.1.3.txt` is delimited by `|` into six fields. - A. Audio file path - B. Original script - C. Expanded script - D. Decomposed script - E. Audio duration (seconds) - F. English translation e.g., 1/1_0470.wav|저는 보통 20분 정도 낮잠을 잡니다.|저는 보통 이십 분 정도 낮잠을 잡니다.|저는 보통 이십 분 정도 낮잠을 잡니다.|4.1|I usually take a nap for 20 minutes. ### Specification - Audio File Type: wav - Total Running Time: 12+ hours - Sample Rate: 44,100 KHZ - Number of Audio Files: 12,853 - Sources - |1| [Kyubyong Park, 500 Basic Korean Verbs, Tuttle Publishing, 2015.](https://www.amazon.com/500-Basic-Korean-Verbs-Comprehensive/dp/0804846057/ref=sr_1_1?s=books&ie=UTF8&qid=1522911616&sr=1-1&keywords=kyubyong+park)| - |2| [Kyubyong Park, 500 Basic Korean Adjectives 2nd Ed., Youkrak, 2015.](http://www.hanbooks.com/500bakoad.html)| - |3| [Kyubyong Park, Essential Korean Vocabulary, Tuttle Publishing, 2015.](https://www.amazon.com/Essential-Korean-Vocabulary-Phrases-Fluently/dp/0804843252/ref=sr_1_3?s=books&ie=UTF8&qid=1522911806&sr=1-3&keywords=kyubyong+park)| - |4| [Kyubyong Park, Tuttle Learner's Korean-English Dictionary, Tuttle Publishing, 2012.](https://www.amazon.com/Tuttle-Learners-Korean-English-Dictionary-Essential/dp/0804841500/ref=sr_1_8?s=books&ie=UTF8&qid=1522911806&sr=1-8&keywords=kyubyong+park)| ### License NC-SA 4.0. You CANNOT use this dataset for ANY COMMERCIAL purpose. Otherwise, you can freely use this. ### Citation If you want to cite KSS Dataset, please refer to this: Kyubyong Park, KSS Dataset: Korean Single speaker Speech Dataset, https://kaggle.com/bryanpark/korean-single-speaker-speech-dataset, 2018 ### Reference Check out [this](https://github.com/Kyubyong/kss) for a project using this KSS Dataset. ### Contact You can contact me at kbpark.linguist@gmail.com. April, 2018. Kyubyong Park ### Dataset Summary 12,853 Korean audio files with transcription. ### Supported Tasks and Leaderboards text-to-speech ### Languages korean ## Dataset Structure ### Data Instances ```python >>> from datasets import load_dataset >>> dataset = load_dataset("Bingsu/KSS_Dataset") >>> dataset["train"].features {'audio': Audio(sampling_rate=44100, mono=True, decode=True, id=None), 'original_script': Value(dtype='string', id=None), 'expanded_script': Value(dtype='string', id=None), 'decomposed_script': Value(dtype='string', id=None), 'duration': Value(dtype='float32', id=None), 'english_translation': Value(dtype='string', id=None)} ``` ```python >>> dataset["train"][0] {'audio': {'path': None, 'array': array([ 0.00000000e+00, 3.05175781e-05, -4.57763672e-05, ..., 0.00000000e+00, -3.05175781e-05, -3.05175781e-05]), 'sampling_rate': 44100}, 'original_script': '그는 괜찮은 척하려고 애쓰는 것 같았다.', 'expanded_script': '그는 괜찮은 척하려고 애쓰는 것 같았다.', 'decomposed_script': '그는 괜찮은 척하려고 애쓰는 것 같았다.', 'duration': 3.5, 'english_translation': 'He seemed to be pretending to be okay.'} ``` ### Data Splits | | train | |---------------|------:| | # of examples | 12853 |
Bingsu/KSS_Dataset
[ "task_categories:text-to-speech", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:ko", "license:cc-by-nc-sa-4.0", "region:us" ]
2022-04-19T05:59:21+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["expert-generated"], "language": ["ko"], "license": ["cc-by-nc-sa-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-to-speech"], "task_ids": [], "pretty_name": "Korean Single Speaker Speech Dataset"}
2022-07-01T23:10:10+00:00
547fc965adf45751ee1fb0881b0c921e66f250ab
hnchen/Session-search
[ "license:afl-3.0", "region:us" ]
2022-04-19T06:25:19+00:00
{"license": "afl-3.0"}
2022-05-07T08:21:10+00:00
ec3653e28f1d3429a0977a7fcac32f308de33e02
samwell/en_twi
[ "license:mit", "region:us" ]
2022-04-19T09:16:56+00:00
{"license": "mit"}
2022-04-19T09:20:26+00:00
199e4ae37915137c555b1765c01477c216287d34
# FLEURS ## Dataset Description - **Fine-Tuning script:** [pytorch/speech-recognition](https://github.com/huggingface/transformers/tree/main/examples/pytorch/speech-recognition) - **Paper:** [FLEURS: Few-shot Learning Evaluation of Universal Representations of Speech](https://arxiv.org/abs/2205.12446) - **Total amount of disk used:** ca. 350 GB Fleurs is the speech version of the [FLoRes machine translation benchmark](https://arxiv.org/abs/2106.03193). We use 2009 n-way parallel sentences from the FLoRes dev and devtest publicly available sets, in 102 languages. Training sets have around 10 hours of supervision. Speakers of the train sets are different than speakers from the dev/test sets. Multilingual fine-tuning is used and ”unit error rate” (characters, signs) of all languages is averaged. Languages and results are also grouped into seven geographical areas: - **Western Europe**: *Asturian, Bosnian, Catalan, Croatian, Danish, Dutch, English, Finnish, French, Galician, German, Greek, Hungarian, Icelandic, Irish, Italian, Kabuverdianu, Luxembourgish, Maltese, Norwegian, Occitan, Portuguese, Spanish, Swedish, Welsh* - **Eastern Europe**: *Armenian, Belarusian, Bulgarian, Czech, Estonian, Georgian, Latvian, Lithuanian, Macedonian, Polish, Romanian, Russian, Serbian, Slovak, Slovenian, Ukrainian* - **Central-Asia/Middle-East/North-Africa**: *Arabic, Azerbaijani, Hebrew, Kazakh, Kyrgyz, Mongolian, Pashto, Persian, Sorani-Kurdish, Tajik, Turkish, Uzbek* - **Sub-Saharan Africa**: *Afrikaans, Amharic, Fula, Ganda, Hausa, Igbo, Kamba, Lingala, Luo, Northern-Sotho, Nyanja, Oromo, Shona, Somali, Swahili, Umbundu, Wolof, Xhosa, Yoruba, Zulu* - **South-Asia**: *Assamese, Bengali, Gujarati, Hindi, Kannada, Malayalam, Marathi, Nepali, Oriya, Punjabi, Sindhi, Tamil, Telugu, Urdu* - **South-East Asia**: *Burmese, Cebuano, Filipino, Indonesian, Javanese, Khmer, Lao, Malay, Maori, Thai, Vietnamese* - **CJK languages**: *Cantonese and Mandarin Chinese, Japanese, Korean* ## How to use & Supported Tasks ### How to use The `datasets` library allows you to load and pre-process your dataset in pure Python, at scale. The dataset can be downloaded and prepared in one call to your local drive by using the `load_dataset` function. For example, to download the Hindi config, simply specify the corresponding language config name (i.e., "hi_in" for Hindi): ```python from datasets import load_dataset fleurs = load_dataset("google/fleurs", "hi_in", split="train") ``` Using the datasets library, you can also stream the dataset on-the-fly by adding a `streaming=True` argument to the `load_dataset` function call. Loading a dataset in streaming mode loads individual samples of the dataset at a time, rather than downloading the entire dataset to disk. ```python from datasets import load_dataset fleurs = load_dataset("google/fleurs", "hi_in", split="train", streaming=True) print(next(iter(fleurs))) ``` *Bonus*: create a [PyTorch dataloader](https://huggingface.co/docs/datasets/use_with_pytorch) directly with your own datasets (local/streamed). Local: ```python from datasets import load_dataset from torch.utils.data.sampler import BatchSampler, RandomSampler fleurs = load_dataset("google/fleurs", "hi_in", split="train") batch_sampler = BatchSampler(RandomSampler(fleurs), batch_size=32, drop_last=False) dataloader = DataLoader(fleurs, batch_sampler=batch_sampler) ``` Streaming: ```python from datasets import load_dataset from torch.utils.data import DataLoader fleurs = load_dataset("google/fleurs", "hi_in", split="train") dataloader = DataLoader(fleurs, batch_size=32) ``` To find out more about loading and preparing audio datasets, head over to [hf.co/blog/audio-datasets](https://huggingface.co/blog/audio-datasets). ### Example scripts Train your own CTC or Seq2Seq Automatic Speech Recognition models on FLEURS with `transformers` - [here](https://github.com/huggingface/transformers/tree/main/examples/pytorch/speech-recognition). Fine-tune your own Language Identification models on FLEURS with `transformers` - [here](https://github.com/huggingface/transformers/tree/main/examples/pytorch/audio-classification) ### 1. Speech Recognition (ASR) ```py from datasets import load_dataset fleurs_asr = load_dataset("google/fleurs", "af_za") # for Afrikaans # to download all data for multi-lingual fine-tuning uncomment following line # fleurs_asr = load_dataset("google/fleurs", "all") # see structure print(fleurs_asr) # load audio sample on the fly audio_input = fleurs_asr["train"][0]["audio"] # first decoded audio sample transcription = fleurs_asr["train"][0]["transcription"] # first transcription # use `audio_input` and `transcription` to fine-tune your model for ASR # for analyses see language groups all_language_groups = fleurs_asr["train"].features["lang_group_id"].names lang_group_id = fleurs_asr["train"][0]["lang_group_id"] all_language_groups[lang_group_id] ``` ### 2. Language Identification LangID can often be a domain classification, but in the case of FLEURS-LangID, recordings are done in a similar setting across languages and the utterances correspond to n-way parallel sentences, in the exact same domain, making this task particularly relevant for evaluating LangID. The setting is simple, FLEURS-LangID is splitted in train/valid/test for each language. We simply create a single train/valid/test for LangID by merging all. ```py from datasets import load_dataset fleurs_langID = load_dataset("google/fleurs", "all") # to download all data # see structure print(fleurs_langID) # load audio sample on the fly audio_input = fleurs_langID["train"][0]["audio"] # first decoded audio sample language_class = fleurs_langID["train"][0]["lang_id"] # first id class language = fleurs_langID["train"].features["lang_id"].names[language_class] # use audio_input and language_class to fine-tune your model for audio classification ``` ### 3. Retrieval Retrieval provides n-way parallel speech and text data. Similar to how XTREME for text leverages Tatoeba to evaluate bitext mining a.k.a sentence translation retrieval, we use Retrieval to evaluate the quality of fixed-size representations of speech utterances. Our goal is to incentivize the creation of fixed-size speech encoder for speech retrieval. The system has to retrieve the English "key" utterance corresponding to the speech translation of "queries" in 15 languages. Results have to be reported on the test sets of Retrieval whose utterances are used as queries (and keys for English). We augment the English keys with a large number of utterances to make the task more difficult. ```py from datasets import load_dataset fleurs_retrieval = load_dataset("google/fleurs", "af_za") # for Afrikaans # to download all data for multi-lingual fine-tuning uncomment following line # fleurs_retrieval = load_dataset("google/fleurs", "all") # see structure print(fleurs_retrieval) # load audio sample on the fly audio_input = fleurs_retrieval["train"][0]["audio"] # decoded audio sample text_sample_pos = fleurs_retrieval["train"][0]["transcription"] # positive text sample text_sample_neg = fleurs_retrieval["train"][1:20]["transcription"] # negative text samples # use `audio_input`, `text_sample_pos`, and `text_sample_neg` to fine-tune your model for retrieval ``` Users can leverage the training (and dev) sets of FLEURS-Retrieval with a ranking loss to build better cross-lingual fixed-size representations of speech. ## Dataset Structure We show detailed information the example configurations `af_za` of the dataset. All other configurations have the same structure. ### Data Instances **af_za** - Size of downloaded dataset files: 1.47 GB - Size of the generated dataset: 1 MB - Total amount of disk used: 1.47 GB An example of a data instance of the config `af_za` looks as follows: ``` {'id': 91, 'num_samples': 385920, 'path': '/home/patrick/.cache/huggingface/datasets/downloads/extracted/310a663d52322700b3d3473cbc5af429bd92a23f9bc683594e70bc31232db39e/home/vaxelrod/FLEURS/oss2_obfuscated/af_za/audio/train/17797742076841560615.wav', 'audio': {'path': '/home/patrick/.cache/huggingface/datasets/downloads/extracted/310a663d52322700b3d3473cbc5af429bd92a23f9bc683594e70bc31232db39e/home/vaxelrod/FLEURS/oss2_obfuscated/af_za/audio/train/17797742076841560615.wav', 'array': array([ 0.0000000e+00, 0.0000000e+00, 0.0000000e+00, ..., -1.1205673e-04, -8.4638596e-05, -1.2731552e-04], dtype=float32), 'sampling_rate': 16000}, 'raw_transcription': 'Dit is nog nie huidiglik bekend watter aantygings gemaak sal word of wat owerhede na die seun gelei het nie maar jeugmisdaad-verrigtinge het in die federale hof begin', 'transcription': 'dit is nog nie huidiglik bekend watter aantygings gemaak sal word of wat owerhede na die seun gelei het nie maar jeugmisdaad-verrigtinge het in die federale hof begin', 'gender': 0, 'lang_id': 0, 'language': 'Afrikaans', 'lang_group_id': 3} ``` ### Data Fields The data fields are the same among all splits. - **id** (int): ID of audio sample - **num_samples** (int): Number of float values - **path** (str): Path to the audio file - **audio** (dict): Audio object including loaded audio array, sampling rate and path ot audio - **raw_transcription** (str): The non-normalized transcription of the audio file - **transcription** (str): Transcription of the audio file - **gender** (int): Class id of gender - **lang_id** (int): Class id of language - **lang_group_id** (int): Class id of language group ### Data Splits Every config only has the `"train"` split containing of *ca.* 1000 examples, and a `"validation"` and `"test"` split each containing of *ca.* 400 examples. ## Dataset Creation We collect between one and three recordings for each sentence (2.3 on average), and buildnew train-dev-test splits with 1509, 150 and 350 sentences for train, dev and test respectively. ## Considerations for Using the Data ### Social Impact of Dataset This dataset is meant to encourage the development of speech technology in a lot more languages of the world. One of the goal is to give equal access to technologies like speech recognition or speech translation to everyone, meaning better dubbing or better access to content from the internet (like podcasts, streaming or videos). ### Discussion of Biases Most datasets have a fair distribution of gender utterances (e.g. the newly introduced FLEURS dataset). While many languages are covered from various regions of the world, the benchmark misses many languages that are all equally important. We believe technology built through FLEURS should generalize to all languages. ### Other Known Limitations The dataset has a particular focus on read-speech because common evaluation benchmarks like CoVoST-2 or LibriSpeech evaluate on this type of speech. There is sometimes a known mismatch between performance obtained in a read-speech setting and a more noisy setting (in production for instance). Given the big progress that remains to be made on many languages, we believe better performance on FLEURS should still correlate well with actual progress made for speech understanding. ## Additional Information All datasets are licensed under the [Creative Commons license (CC-BY)](https://creativecommons.org/licenses/). ### Citation Information You can access the FLEURS paper at https://arxiv.org/abs/2205.12446. Please cite the paper when referencing the FLEURS corpus as: ``` @article{fleurs2022arxiv, title = {FLEURS: Few-shot Learning Evaluation of Universal Representations of Speech}, author = {Conneau, Alexis and Ma, Min and Khanuja, Simran and Zhang, Yu and Axelrod, Vera and Dalmia, Siddharth and Riesa, Jason and Rivera, Clara and Bapna, Ankur}, journal={arXiv preprint arXiv:2205.12446}, url = {https://arxiv.org/abs/2205.12446}, year = {2022}, ``` ### Contributions Thanks to [@patrickvonplaten](https://github.com/patrickvonplaten) and [@aconneau](https://github.com/aconneau) for adding this dataset.
google/fleurs
[ "task_categories:automatic-speech-recognition", "annotations_creators:expert-generated", "annotations_creators:crowdsourced", "annotations_creators:machine-generated", "language_creators:crowdsourced", "language_creators:expert-generated", "multilinguality:multilingual", "size_categories:10K<n<100K", "language:afr", "language:amh", "language:ara", "language:asm", "language:ast", "language:azj", "language:bel", "language:ben", "language:bos", "language:cat", "language:ceb", "language:cmn", "language:ces", "language:cym", "language:dan", "language:deu", "language:ell", "language:eng", "language:spa", "language:est", "language:fas", "language:ful", "language:fin", "language:tgl", "language:fra", "language:gle", "language:glg", "language:guj", "language:hau", "language:heb", "language:hin", "language:hrv", "language:hun", "language:hye", "language:ind", "language:ibo", "language:isl", "language:ita", "language:jpn", "language:jav", "language:kat", "language:kam", "language:kea", "language:kaz", "language:khm", "language:kan", "language:kor", "language:ckb", "language:kir", "language:ltz", "language:lug", "language:lin", "language:lao", "language:lit", "language:luo", "language:lav", "language:mri", "language:mkd", "language:mal", "language:mon", "language:mar", "language:msa", "language:mlt", "language:mya", "language:nob", "language:npi", "language:nld", "language:nso", "language:nya", "language:oci", "language:orm", "language:ory", "language:pan", "language:pol", "language:pus", "language:por", "language:ron", "language:rus", "language:bul", "language:snd", "language:slk", "language:slv", "language:sna", "language:som", "language:srp", "language:swe", "language:swh", "language:tam", "language:tel", "language:tgk", "language:tha", "language:tur", "language:ukr", "language:umb", "language:urd", "language:uzb", "language:vie", "language:wol", "language:xho", "language:yor", "language:yue", "language:zul", "license:cc-by-4.0", "speech-recognition", "arxiv:2205.12446", "arxiv:2106.03193", "region:us" ]
2022-04-19T09:25:58+00:00
{"annotations_creators": ["expert-generated", "crowdsourced", "machine-generated"], "language_creators": ["crowdsourced", "expert-generated"], "language": ["afr", "amh", "ara", "asm", "ast", "azj", "bel", "ben", "bos", "cat", "ceb", "cmn", "ces", "cym", "dan", "deu", "ell", "eng", "spa", "est", "fas", "ful", "fin", "tgl", "fra", "gle", "glg", "guj", "hau", "heb", "hin", "hrv", "hun", "hye", "ind", "ibo", "isl", "ita", "jpn", "jav", "kat", "kam", "kea", "kaz", "khm", "kan", "kor", "ckb", "kir", "ltz", "lug", "lin", "lao", "lit", "luo", "lav", "mri", "mkd", "mal", "mon", "mar", "msa", "mlt", "mya", "nob", "npi", "nld", "nso", "nya", "oci", "orm", "ory", "pan", "pol", "pus", "por", "ron", "rus", "bul", "snd", "slk", "slv", "sna", "som", "srp", "swe", "swh", "tam", "tel", "tgk", "tha", "tur", "ukr", "umb", "urd", "uzb", "vie", "wol", "xho", "yor", "yue", "zul"], "license": ["cc-by-4.0"], "multilinguality": ["multilingual"], "size_categories": ["10K<n<100K"], "task_categories": ["automatic-speech-recognition"], "task_ids": [], "pretty_name": "The Cross-lingual TRansfer Evaluation of Multilingual Encoders for Speech (XTREME-S) benchmark is a benchmark designed to evaluate speech representations across languages, tasks, domains and data regimes. It covers 102 languages from 10+ language families, 3 different domains and 4 task families: speech recognition, translation, classification and retrieval.", "tags": ["speech-recognition"]}
2023-02-07T20:51:01+00:00
5a8256d0dff9c4bd3be3ba3e67e4e70173f802ea
mteb/sprintduplicatequestions-pairclassification
[ "language:en", "region:us" ]
2022-04-19T09:40:02+00:00
{"language": ["en"]}
2022-09-27T18:15:57+00:00
d55ad9aa644a3b7cea98e5b0dc408792eb3369e0
# Open Medieval French Source: [https://github.com/OpenMedFr/texts](https://github.com/OpenMedFr/texts)
bigscience-historical-texts/Open_Medieval_French
[ "language:fro", "region:us" ]
2022-04-19T09:49:03+00:00
{"language": ["fro"]}
2022-12-12T08:50:28+00:00
4d853f94cd57d85ec13805aeeac3ae3e5eb4c49c
mteb/askubuntudupquestions-reranking
[ "language:en", "region:us" ]
2022-04-19T10:08:05+00:00
{"language": ["en"]}
2022-09-27T18:11:08+00:00
40abf43ceb2f64198e063e981f46d58ef07c304d
# Wikinews-fr-100 Benchmark Dataset for Keyphrase Generation ## About Wikinews-fr-100 is a dataset for benchmarking keyphrase extraction and generation models. The dataset is composed of 100 news articles in French collected from [wikinews](https://fr.wikinews.org/wiki/Accueil). Keyphrases were annotated by readers (students in computer science) in an uncontrolled setting (that is, not limited to thesaurus entries). Details about the dataset can be found in the original paper [(Bougouin et al., 2013)][bougouin-2013]. Reference (indexer-assigned) keyphrases are also categorized under the PRMU (<u>P</u>resent-<u>R</u>eordered-<u>M</u>ixed-<u>U</u>nseen) scheme as proposed in [(Boudin and Gallina, 2021)][boudin-2021]. Present reference keyphrases are also ordered by their order of apparition in the concatenation of title and abstract. Text pre-processing (tokenization) is carried out using `spacy` (`fr_core_news_sm` model) with a special rule to avoid splitting words with hyphens (e.g. graph-based is kept as one token). Stemming (Snowball stemmer implementation for french provided in `nltk`) is applied before reference keyphrases are matched against the source text. Details about the process can be found in `prmu.py`. ## Content and statistics The dataset contains the following test split: | Split | # documents | #words | # keyphrases | % Present | % Reordered | % Mixed | % Unseen | | :--------- | ----------: | -----: | -----------: | --------: | ----------: | ------: | -------: | | Test | 100 | 306.9 | 9.64 | 95.91 | 1.40 | 0.85 | 1.84 | The following data fields are available : - **id**: unique identifier of the document. - **title**: title of the document. - **abstract**: abstract of the document. - **keyphrases**: list of reference keyphrases. - **prmu**: list of <u>P</u>resent-<u>R</u>eordered-<u>M</u>ixed-<u>U</u>nseen categories for reference keyphrases. ## References - (Bougouin et al., 2013) Adrien Bougouin, Florian Boudin, and Béatrice Daille. 2013. [TopicRank: Graph-Based Topic Ranking for Keyphrase Extraction][bougouin-2013]. In Proceedings of the Sixth International Joint Conference on Natural Language Processing, pages 543–551, Nagoya, Japan. Asian Federation of Natural Language Processing. - (Boudin and Gallina, 2021) Florian Boudin and Ygor Gallina. 2021. [Redefining Absent Keyphrases and their Effect on Retrieval Effectiveness][boudin-2021]. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 4185–4193, Online. Association for Computational Linguistics. [bougouin-2013]: https://aclanthology.org/I13-1062/ [boudin-2021]: https://aclanthology.org/2021.naacl-main.330/
taln-ls2n/wikinews-fr-100
[ "task_categories:text-generation", "annotations_creators:unknown", "language_creators:unknown", "multilinguality:monolingual", "size_categories:n<1K", "language:fr", "license:cc-by-4.0", "region:us" ]
2022-04-19T10:55:39+00:00
{"annotations_creators": ["unknown"], "language_creators": ["unknown"], "language": ["fr"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["n<1K"], "task_categories": ["text-mining", "text-generation"], "task_ids": ["keyphrase-generation", "keyphrase-extraction"], "pretty_name": "Wikinews-fr-100"}
2022-09-23T06:38:18+00:00
56a6d0140cf6356659e2a7c1413286a774468d44
mteb/scidocs-reranking
[ "language:en", "region:us" ]
2022-04-19T11:15:26+00:00
{"language": ["en"]}
2022-09-27T18:11:31+00:00
ef807ea29a75ec4f91b50fd4191cb4ee4589a9f9
mteb/stackoverflowdupquestions-reranking
[ "language:en", "region:us" ]
2022-04-19T11:24:06+00:00
{"language": ["en"]}
2022-09-27T18:13:01+00:00
3d3a799e2c93779d80e316fb5f2991332ef55fe3
JEFFREY-VERDIERE/train_X_PJT_NLP
[ "region:us" ]
2022-04-19T12:30:30+00:00
{}
2022-04-19T12:34:19+00:00
986fd2f2865cc142a9177e27e11b5585bd0c885a
# TALN-Archives Benchmark Dataset for Keyphrase Generation ## About TALN-Archives is a dataset for benchmarking keyphrase extraction and generation models. The dataset is composed of 1207 abstracts of scientific papers in French collected from the [TALN Archives](http://talnarchives.atala.org/). Keyphrases were annotated by authors in an uncontrolled setting (that is, not limited to thesaurus entries). English translations of title/abstract/keyphrases are also available for a subset of the documents (456 fully- and 719 partially-translated documents), allowing to experiment with cross-lingual / multilingual keyphrase generation. Details about the dataset can be found in the original paper [(Boudin, 2013)][boudin-2013]. Reference (indexer-assigned) keyphrases are also categorized under the PRMU (<u>P</u>resent-<u>R</u>eordered-<u>M</u>ixed-<u>U</u>nseen) scheme as proposed in [(Boudin and Gallina, 2021)][boudin-2021]. <u>P</u>resent reference keyphrases are also ordered by their order of apparition in the concatenation of title and abstract. Text pre-processing (tokenization) is carried out using `spacy` (`fr_core_news_sm` model) with a special rule to avoid splitting words with hyphens (e.g. graph-based is kept as one token). Stemming (Snowball stemmer implementation for french provided in `nltk`) is applied before reference keyphrases are matched against the source text. Details about the process can be found in `prmu.py`. ## Content and statistics The dataset contains the following test split: | Split | # documents | #words | # keyphrases | % Present | % Reordered | % Mixed | % Unseen | | :--------- | ----------: | -----: | -----------: | --------: | ----------: | ------: | -------: | | Test | 1207 | 138.3 | 4.12 | 53.83 | 12.32 | 21.69 | 12.16 | The following data fields are available : - **id**: unique identifier of the document. - **title**: title of the document. - **abstract**: abstract of the document. - **keyphrases**: list of reference keyphrases. - **prmu**: list of <u>P</u>resent-<u>R</u>eordered-<u>M</u>ixed-<u>U</u>nseen categories for reference keyphrases. - **translation**: translations of title, abstract and keyphrases in English if available. ## References - (Boudin, 2013) Florian Boudin. 2013. [TALN Archives : a digital archive of French research articles in Natural Language Processing (TALN Archives : une archive numérique francophone des articles de recherche en Traitement Automatique de la Langue) [in French]][boudin-2013]. In Proceedings of TALN 2013 (Volume 2: Short Papers), pages 507–514, Les Sables d’Olonne, France. ATALA. - (Boudin and Gallina, 2021) Florian Boudin and Ygor Gallina. 2021. [Redefining Absent Keyphrases and their Effect on Retrieval Effectiveness][boudin-2021]. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 4185–4193, Online. Association for Computational Linguistics. [boudin-2013]: https://aclanthology.org/F13-2001/ [boudin-2021]: https://aclanthology.org/2021.naacl-main.330/
taln-ls2n/taln-archives
[ "task_categories:text-generation", "annotations_creators:unknown", "language_creators:unknown", "multilinguality:multilingual", "size_categories:1K<n<10K", "language:fr", "language:en", "license:cc-by-4.0", "region:us" ]
2022-04-19T12:45:33+00:00
{"annotations_creators": ["unknown"], "language_creators": ["unknown"], "language": ["fr", "en"], "license": ["cc-by-4.0"], "multilinguality": ["multilingual"], "size_categories": ["1K<n<10K"], "task_categories": ["text-mining", "text-generation"], "task_ids": ["keyphrase-generation", "keyphrase-extraction"], "pretty_name": "TALN-Archives"}
2022-09-23T06:58:07+00:00
20a6d6f312dd54037fe07a32d58e5e168867909d
mteb/sickr-sts
[ "language:en", "region:us" ]
2022-04-19T13:28:03+00:00
{"language": ["en"]}
2022-09-27T18:13:22+00:00
803ef9f038eb29517b85712d839188e2daf7629e
# 🏴‍☠️ **_This dataset is depricated! Please see NENA Speech_** 🏴‍☠️ # Dataset Card for urmi_assyrian_voice ## Dataset Description The Urmi Assyrian Voice dataset is parsed from the research and fieldwork of [Geoffrey Khan](https://cambridge.academia.edu/GeoffreyKhan) which is made public through the [North-Eastern Neo-Aramaic Database Project](https://nena.ames.cam.ac.uk/dialects/225/audio). Annotation corrections as well as parsing was performed by Matthew Nazari. ## Dataset Summary This dataset contains labelled audio examples of the Urmi dialect of North-Eastern Neo-Aramaic. The dataset only consists of one female speaker in her late seventies. Note that you will need to normalize the utterances for machine learning tasks (clean punctuation, remove accents, etc).
mnazari/urmi-assyrian-voice
[ "task_categories:automatic-speech-recognition", "annotations_creators:Geoffrey Khan", "annotations_creators:Matthew Nazari", "language:aii", "license:cc0-1.0", "region:us" ]
2022-04-19T13:29:14+00:00
{"annotations_creators": ["Geoffrey Khan", "Matthew Nazari"], "language": "aii", "license": "cc0-1.0", "task_categories": ["automatic-speech-recognition"], "pretty_name": "Assyrian", "size_category": "n<1K"}
2023-09-22T04:31:05+00:00
9ee918f184421b6bd48b78f6c714d86546106103
mteb/biosses-sts
[ "language:en", "region:us" ]
2022-04-19T13:47:25+00:00
{"language": ["en"]}
2022-09-27T18:13:38+00:00
8913289635987208e6e7c72789e4be2fe94b6abd
mteb/stsbenchmark-sts
[ "language:en", "region:us" ]
2022-04-19T13:53:43+00:00
{"language": ["en"]}
2022-09-27T18:11:21+00:00
47054de4458827ac3fb5136f5f953ddf3deb3c53
# Dataset Card for Multi-Document ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Repository:** [Multi-Document repository](https://github.com/arka0821/multi_document_summarization) - **Paper:** [Multi-Document: A Large-scale Dataset for Extreme Multi-document Summarization of Scientific Articles](https://arxiv.org/abs/2010.14235) ### Dataset Summary Multi-Document, a large-scale multi-document summarization dataset created from scientific articles. Multi-Document introduces a challenging multi-document summarization task: writing the related-work section of a paper based on its abstract and the articles it references. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The text in the dataset is in English ## Dataset Structure ### Data Instances {"id": "n3ByHGrxH3bvfrvF", "docs": [{"id": "1394519630182457344", "text": "Clover Bio's COVID-19 vaccine candidate shows immune response against SARS-CoV-2 variants in mouse model https://t.co/wNWa9GQux5"}, {"id": "1398154482463170561", "text": "The purpose of the Vaccine is not to stop you from catching COVID 19. The vaccine introduces the immune system to an inactivated form of the SARS-CoV-2 coronavirus or a small part of it. This then equips the body with the ability to fight the virus better in case you get it. https://t.co/Cz9OU6Zi7P"}, {"id": "1354844652520792071", "text": "The Moderna mRNA COVID-19 vaccine appears to be effective against the novel, rapidly spreading variants of SARS-CoV-2.\nResearchers analysed blood samples from vaccinated people and monkeys- Both contained neutralising antibodies against the virus. \nPT1/2\n#COVID19vaccines #biotech https://t.co/ET1maJznot"}, {"id": "1340189698107518976", "text": "@KhandaniM Pfizer vaccine introduces viral surface protein which is constant accross SARS COV 2 variants into the body. Body builds antibodies against this protein, not any virus. These antibodies instructs macrophages &amp; T-Cells to attack &amp; destroy any COVID-19 v variant at infection point"}, {"id": "1374368989581778945", "text": "@DelthiaRicks \" Pfizer and BioNTech\u2019s COVID-19 vaccine is an mRNA vaccine, which does not use the live virus but rather a small portion of the viral sequence of the SARS-CoV-2 virus to instruct the body to produce the spike protein displayed on the surface of the virus.\""}, {"id": "1353354819315126273", "text": "Pfizer and BioNTech Publish Results of Study Showing COVID-19 Vaccine Elicits Antibodies that Neutralize Pseudovirus Bearing the SARS-CoV-2 U.K. Strain Spike Protein in Cell Culture | Pfizer https://t.co/YXcSnjLt8C"}, {"id": "1400821856362401792", "text": "Pfizer-BioNTech's covid-19 vaccine elicits lower levels of antibodies against the SARS-CoV-2\u00a0Delta variant\u00a0(B.1.617.2), first discovered in India, in comparison to other variants, said a research published in\u00a0Lancet\u00a0journal.\n https://t.co/IaCMX81X3b"}, {"id": "1367252963190665219", "text": "New research from UNC-Chapel Hill suggests that those who have previously experienced a SARS-CoV-2 infection develop a significant antibody response to the first dose of mRNA-based COVID-19 vaccine.\nhttps://t.co/B4vR1KUQ0w"}, {"id": "1375949502461394946", "text": "Mechanism of a COVID-19 nanoparticle vaccine candidate that elicits a broadly neutralizing antibody response to SARS-CoV-2 variants https://t.co/nc1L0uvtlI #bioRxiv"}, {"id": "1395428608349548550", "text": "JCI - Efficient maternal to neonatal transfer of antibodies against SARS-CoV-2 and BNT162b2 mRNA COVID-19 vaccine https://t.co/vIBcpPaKFZ"}], "summary": "The COVID-19 vaccine appears to be effective against the novel, rapidly spreading variants of SARS-CoV-2. Pfizer-BioNTech's COVID-19 vaccine use small portion of the viral sequence of the SARS-CoV-2 virus to equip the body with the ability to fight the virus better in case you get it."} ### Data Fields {'id': text of paper abstract \ 'docs': document id \ [ 'id': id of text \ 'text': text data \ ] 'summary': summary text } ### Data Splits The data is split into a training, validation and test. | train | validation | test | |------:|-----------:|-----:| | 50 | 10 | 5 | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ``` @article{lu2020multi, title={Multi-Document: A Large-scale Dataset for Extreme Multi-document Summarization of Scientific Articles}, author={Arka Das, India}, journal={arXiv preprint arXiv:2010.14235}, year={2022} } ``` ### Contributions Thanks to [@arka0821] (https://github.com/arka0821/multi_document_summarization) for adding this dataset.
arka0821/multi_document_summarization
[ "task_categories:summarization", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:unknown", "arxiv:2010.14235", "region:us" ]
2022-04-19T14:34:53+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["summarization"], "task_ids": ["summarization-other-paper-abstract-generation"], "paperswithcode_id": "multi-document", "pretty_name": "Multi-Document"}
2022-10-20T18:13:26+00:00
f0bf0ee2765b1df5e8dc9bc0f3bd661fd880fda5
asdAD222/segment_T
[ "region:us" ]
2022-04-19T14:53:17+00:00
{}
2022-04-19T14:56:02+00:00
2a0ac924abdb53e4ba4a7c1ae943fc8ef3b296d3
danrambado/SODA
[ "license:mit", "region:us" ]
2022-04-19T17:57:01+00:00
{"license": "mit"}
2022-04-19T17:57:01+00:00
4265bc69c85c8f1d21b223a8e6cc61cdf57fda95
# Dataset Card for ID Word2Phoneme ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-splits) - [Additional Information](#additional-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** [Github](https://github.com/open-dict-data/ipa-dict/blob/master/data/ma.txt) - **Repository:** [Github](https://github.com/open-dict-data/ipa-dict/blob/master/data/ma.txt) - **Point of Contact:** - **Size of downloaded dataset files:** - **Size of the generated dataset:** - **Total amount of disk used:** ### Dataset Summary Originally a [Malay/Indonesian Lexicon](https://github.com/open-dict-data/ipa-dict/blob/master/data/ma.txt) retrieved from [ipa-dict](https://github.com/open-dict-data/ipa-dict). We removed the accented letters (because Indonesian graphemes do not use accents), separated homographs, and removed backslashes in phonemes -- resulting in a word-to-phoneme dataset. ### Languages - Indonesian - Malay ## Dataset Structure ### Data Instances | word | phoneme | | ----- | ------- | | aba | aba | | ab | ab | | ab’ad | abʔad | | abad | abad | | abadi | abadi | | ... | ... | ### Data Fields - `word`: Word (grapheme) as a string. - `phoneme`: Phoneme (IPA) as a string. ### Data Splits | train | | ----- | | 27553 | ## Additional Information ### Citation Information ``` @misc{open-dict-data-no-date, author = {{Open-Dict-Data}}, title = {{GitHub - open-dict-data/ipa-dict: Monolingual wordlists with pronunciation information in IPA}}, url = {https://github.com/open-dict-data/ipa-dict}, } ```
bookbot/id_word2phoneme
[ "task_categories:text2text-generation", "annotations_creators:no-annotation", "language_creators:found", "source_datasets:original", "language:id", "language:ms", "region:us" ]
2022-04-20T06:37:29+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["found"], "language": ["id", "ms"], "source_datasets": ["original"], "task_categories": ["text2text-generation"], "task_ids": [], "pretty_name": "ID Word2Phoneme"}
2023-03-20T10:00:22+00:00
6128e88e317eb20bb3c429b09db8eecd915bbd41
Khalsuu/filipino_dataset_script
[ "license:apache-2.0", "region:us" ]
2022-04-20T07:18:10+00:00
{"license": "apache-2.0"}
2022-04-28T14:25:10+00:00
a064f3d34057b27eb27a12bf81165f5fad9a09f1
# Dataset Card for "CrossSum" ## Table of Contents - [Dataset Card Creation Guide](#dataset-card-creation-guide) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Initial Data Collection and Normalization](#initial-data-collection-and-normalization) - [Who are the source language producers?](#who-are-the-source-language-producers) - [Annotations](#annotations) - [Annotation process](#annotation-process) - [Who are the annotators?](#who-are-the-annotators) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Repository:** [https://github.com/csebuetnlp/CrossSum](https://github.com/csebuetnlp/CrossSum) - **Paper:** [CrossSum: Beyond English-Centric Cross-Lingual Abstractive Text Summarization for 1500+ Language Pairs](https://arxiv.org/abs/2112.08804) - **Point of Contact:** [Tahmid Hasan](mailto:tahmidhasan@cse.buet.ac.bd) ### Dataset Summary We present CrossSum, a large-scale dataset comprising 1.70 million cross-lingual article summary samples in 1500+ language-pairs constituting 45 languages. We use the multilingual XL-Sum dataset and align identical articles written in different languages via crosslingual retrieval using a language-agnostic representation model. ### Supported Tasks and Leaderboards [More information needed](https://github.com/csebuetnlp/CrossSum) ### Languages - `amharic` - `arabic` - `azerbaijani` - `bengali` - `burmese` - `chinese_simplified` - `chinese_traditional` - `english` - `french` - `gujarati` - `hausa` - `hindi` - `igbo` - `indonesian` - `japanese` - `kirundi` - `korean` - `kyrgyz` - `marathi` - `nepali` - `oromo` - `pashto` - `persian` - `pidgin` - `portuguese` - `punjabi` - `russian` - `scottish_gaelic` - `serbian_cyrillic` - `serbian_latin` - `sinhala` - `somali` - `spanish` - `swahili` - `tamil` - `telugu` - `thai` - `tigrinya` - `turkish` - `ukrainian` - `urdu` - `uzbek` - `vietnamese` - `welsh` - `yoruba` ## Loading the dataset ```python from datasets import load_dataset # for available language names, see above src_lang = "english" tgt_lang = "bengali" ds = load_dataset(f"csebuetnlp/CrossSum", "{}-{}".format(src_lang, tgt_lang)) ``` ## Dataset Structure ### Data Instances One example from the `English` dataset is given below in JSON format. ``` { "source_url": "https://www.bbc.com/japanese/53074000", "target_url": "https://www.bbc.com/bengali/news-53064712", "summary": "বিজ্ঞানীরা বলছেন ডেক্সামেথাসোন নামে সস্তা ও সহজলভ্য একটি ওষুধ করোনাভাইরাসে গুরুতর অসুস্থ রোগীদের জীবন রক্ষা করতে সাহায্য করবে।", "text": "ミシェル・ロバーツ、BBCニュースオンライン健康担当編集長 英オックスフォード大学の研究チームによると、低用量のデキサメタゾンは新型ウイルスとの戦いで画期的な突破口になる。 新型コロナウイルスに対し、様々な既存の治療法の効果を試す世界的規模の臨床試験の一貫として、デキサメタゾンが試された。 その結果、人工呼吸器を必要とする重症患者の致死率が3割下がり、酸素供給を必要とする患者の場合は2割下がった。 新型ウイルスのパンデミック(世界的流行)の初期からイギリスでデキサメタゾンを治療に使用していた場合、最大5000人の命が救えたはずだと研究者たちは言う。 さらに、新型コロナウイルスによる感染症「COVID-19」の患者が多く出ている貧しい国にとっても、安価なデキサメタゾンを使う治療は大いに役立つと期待される。 重症者の致死率が大幅に下がる イギリス政府は20万人分の投与量を備蓄しており、国民医療制度の国民保健サービス(NHS)で患者への使用を開始する方針を示した。 ボリス・ジョンソン英首相は「イギリス科学界の素晴らしい成果」を歓迎し、「たとえ感染の第2波が来ても備蓄が足りるよう、数を確保するための措置をとった」と述べた。 イングランド首席医務官クリス・ウィッティー教授は、「COVID-19にとってこれまでで一番重要な臨床試験結果だ。手に入りやすく安全でなじみのある薬によって、酸素供給や人工呼吸器が必要な人の致死率が大幅に下がった。(中略)この発見が世界中で人命を救う」と評価した。 <関連記事> 新型コロナウイルスに20人が感染した場合、19人は入院しないまま回復する。入院する人もほとんどは回復するものの、重症化して酸素供給や人工呼吸器を必要とする人もいる。 デキサメタゾンはこうした重症患者の治療に効果があるもよう。 新型ウイルスに感染した患者の体内では、ウイルスと戦う免疫系が暴走することがある。その免疫系の過剰反応による体の損傷を、デキサメタゾンが緩和するものとみられる。 「サイトカイン・ストーム」と呼ばれる免疫系の過剰反応が、患者の命を奪うこともある。 デキサメタゾンはすでに抗炎症剤として、ぜんそくや皮膚炎など様々な症状の治療に使われている。 初めて致死率を下げる薬 オックスフォード大学が主導する臨床試験は、約2000人の入院患者にデキサメタゾンを投与。それ以外の4000人以上の患者と容体を比較した。 人工呼吸器を使用する患者については、死亡リスクが40%から28%に下がった。 酸素供給する患者は、死亡リスクが25%から20%に下がった。 研究チームのピーター・ホービー教授は、「今のところ、致死率を実際に下げる結果が出たのは、この薬だけだ。しかも、致死率をかなり下げる。画期的な突破口だ」と話した。 研究を主導するマーティン・ランドレイ教授によると、人工呼吸器を使う患者の8人に1人、ならびに酸素供給治療を受ける患者の20-25人に1人が、デキサメタゾンで救えることが分かったという。 「これはきわめて明確なメリットだ」と教授は言う。 「最大10日間、デキサメタゾンを投与するという治療法で、費用は患者1人あたり1日約5ポンド(約670円)。つまり、35ポンド(約4700円)で人ひとりの命が救える」 「しかもこれは、世界中で手に入る薬だ」 状況が許す限り、新型コロナウイルスで入院中の患者にはただちに投与を開始すべきだと、ランドレイ教授は促した。 ただし、自宅で自己治療するために薬局に買いに行くべきではないと言う。 デキサメタゾンは、呼吸補助を必要としない軽症の患者には効果がないもよう。 3月に始動した新型コロナウイルス治療薬の無作為化臨床試験「リカバリー・トライアル」は、抗マラリア薬「ヒドロキシクロロキン」も調べたものの、心臓疾患や致死率の悪化につながるという懸念から、ヒドロキシクロロキンについては試験を中止した。 一方で、感染者の回復にかかる時間を短縮するとみられるレムデシビルは、すでにNHSの保険対象になり治療現場で使われている。 <解説> ファーガス・ウォルシュBBC健康担当編集委員 COVID-19の死者を減らすと初めて立証された薬は、高価な新しい薬ではなく、古くからずっと使われてきた、きわめて安いステロイド剤だった。 世界中の患者が直ちにその恩恵を受けることになるので、これは歓迎すべき発見だ。 この臨床試験の最新成果がこれほど急いで発表されたのは、そのためだ。とてつもない影響を世界中にもたらすので。 デキサメタゾンは1960年代初めから、関節リウマチやぜんそくなど、幅広い症状の治療に使われてきた。 これまでは、人工呼吸器を必要とするCOVID-19患者の半数が亡くなってきた。その致死率を3割減らすというのは、絶大な効果だ。 集中治療室では点滴で投与する。もう少し軽症な患者には、錠剤で与える。 これまでのところ、COVID-19患者に効果があると証明された薬は、エボラ治療薬のレムデシビルだけだった。 レムデシビルは症状の回復期間を15日から11日に短縮する。しかし、致死率を下げると言えるだけの証拠は出ていなかった。 デキサメタゾンと異なり、レムデシビルは数の少ない新薬で、薬価もまだ公表されていない。" } ``` ### Data Fields - 'source_url': A string representing the source article URL. - 'target_url': A string representing the target article URL. - 'summary': A string containing the article summary. - 'text' : A string containing the article text. ### Data Splits No. of total examples for each language pair are as follows: Language (ISO 639-1-Code) | am | ar | az | bn | my | zh-CN | zh-TW | en | fr | gu | ha | hi | ig | id | ja | rn | ko | ky | mr | np | om | ps | fa | pcm | pt | pa | ru | gd | sr | sr | si | so | es | sw | ta | te | th | ti | tr | uk | ur | uz | vi | cy | yo ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- | ----- am | -- | 667 | 100 | 272 | 95 | 179 | 167 | 1456 | 358 | 173 | 221 | 377 | 26 | 494 | 264 | 423 | 244 | 92 | 221 | 301 | 21 | 192 | 431 | 209 | 307 | 189 | 347 | 0 | 357 | 365 | 62 | 309 | 351 | 378 | 390 | 329 | 124 | 131 | 435 | 345 | 409 | 41 | 285 | 1 | 67 ar | 667 | -- | 787 | 804 | 652 | 2968 | 2843 | 9653 | 989 | 475 | 747 | 3665 | 86 | 6084 | 1188 | 876 | 707 | 299 | 559 | 854 | 9 | 2161 | 4186 | 436 | 2539 | 547 | 5564 | 1 | 1109 | 1145 | 315 | 1049 | 3654 | 1186 | 1311 | 877 | 367 | 27 | 4147 | 3457 | 4935 | 388 | 2666 | 38 | 141 az | 100 | 787 | -- | 277 | 84 | 371 | 334 | 1317 | 208 | 192 | 126 | 748 | 28 | 1111 | 231 | 188 | 155 | 221 | 194 | 242 | 1 | 252 | 817 | 91 | 678 | 190 | 2238 | 4 | 289 | 283 | 124 | 367 | 704 | 539 | 515 | 245 | 140 | 2 | 1495 | 1383 | 966 | 199 | 725 | 30 | 42 bn | 272 | 804 | 277 | -- | 139 | 318 | 284 | 1549 | 317 | 559 | 231 | 1396 | 35 | 1076 | 342 | 298 | 352 | 154 | 586 | 668 | 2 | 300 | 790 | 135 | 764 | 580 | 838 | 0 | 562 | 564 | 151 | 412 | 701 | 471 | 919 | 793 | 245 | 6 | 860 | 688 | 1382 | 98 | 527 | 37 | 61 my | 95 | 652 | 84 | 139 | -- | 356 | 314 | 685 | 90 | 96 | 74 | 528 | 12 | 761 | 144 | 100 | 112 | 58 | 89 | 152 | 1 | 234 | 426 | 39 | 230 | 86 | 535 | 0 | 115 | 123 | 87 | 79 | 431 | 86 | 185 | 147 | 71 | 4 | 449 | 350 | 591 | 62 | 447 | 4 | 12 zh-CN | 179 | 2968 | 371 | 318 | 356 | -- | 47101 | 4975 | 348 | 201 | 159 | 1379 | 38 | 2851 | 1017 | 240 | 412 | 139 | 240 | 275 | 14 | 559 | 1111 | 149 | 1371 | 250 | 2572 | 2 | 504 | 530 | 166 | 323 | 2002 | 412 | 511 | 353 | 269 | 11 | 1511 | 1619 | 1651 | 176 | 1858 | 33 | 39 zh-TW | 167 | 2843 | 334 | 284 | 314 | 47101 | -- | 4884 | 331 | 174 | 150 | 1213 | 35 | 2588 | 953 | 209 | 382 | 131 | 213 | 252 | 16 | 501 | 967 | 141 | 1271 | 226 | 2286 | 1 | 453 | 494 | 150 | 302 | 1873 | 383 | 465 | 335 | 250 | 12 | 1294 | 1464 | 1444 | 158 | 1663 | 31 | 38 en | 1456 | 9653 | 1317 | 1549 | 685 | 4975 | 4884 | -- | 1889 | 978 | 913 | 4728 | 144 | 10040 | 3040 | 1878 | 1673 | 490 | 1181 | 1614 | 38 | 1522 | 4680 | 1074 | 4744 | 1330 | 9080 | 128 | 3760 | 3809 | 532 | 2141 | 6910 | 2701 | 3156 | 2121 | 1020 | 58 | 5676 | 6562 | 6320 | 450 | 4574 | 2655 | 229 fr | 358 | 989 | 208 | 317 | 90 | 348 | 331 | 1889 | -- | 242 | 477 | 616 | 106 | 1018 | 274 | 735 | 264 | 124 | 241 | 323 | 4 | 196 | 602 | 439 | 921 | 247 | 849 | 2 | 555 | 569 | 98 | 502 | 990 | 872 | 425 | 380 | 185 | 10 | 829 | 721 | 766 | 76 | 438 | 40 | 159 gu | 173 | 475 | 192 | 559 | 96 | 201 | 174 | 978 | 242 | -- | 147 | 5170 | 34 | 710 | 228 | 183 | 268 | 106 | 2091 | 561 | 1 | 246 | 522 | 101 | 529 | 2210 | 582 | 0 | 331 | 345 | 125 | 261 | 540 | 300 | 1762 | 2066 | 164 | 5 | 631 | 508 | 1619 | 80 | 450 | 21 | 54 ha | 221 | 747 | 126 | 231 | 74 | 159 | 150 | 913 | 477 | 147 | -- | 460 | 202 | 901 | 157 | 485 | 135 | 61 | 159 | 239 | 5 | 229 | 487 | 529 | 375 | 157 | 525 | 1 | 258 | 258 | 49 | 391 | 463 | 568 | 299 | 260 | 87 | 9 | 519 | 400 | 526 | 59 | 352 | 30 | 362 hi | 377 | 3665 | 748 | 1396 | 528 | 1379 | 1213 | 4728 | 616 | 5170 | 460 | -- | 65 | 5627 | 623 | 489 | 520 | 234 | 3831 | 1357 | 4 | 1519 | 5351 | 192 | 6563 | 4052 | 4622 | 1 | 809 | 807 | 449 | 747 | 2931 | 893 | 3711 | 3762 | 378 | 7 | 3694 | 3935 | 15666 | 352 | 3738 | 77 | 79 ig | 26 | 86 | 28 | 35 | 12 | 38 | 35 | 144 | 106 | 34 | 202 | 65 | -- | 113 | 24 | 107 | 32 | 16 | 51 | 36 | 3 | 11 | 49 | 255 | 61 | 39 | 79 | 0 | 51 | 51 | 13 | 77 | 91 | 151 | 52 | 54 | 18 | 5 | 91 | 83 | 61 | 15 | 65 | 6 | 296 id | 494 | 6084 | 1111 | 1076 | 761 | 2851 | 2588 | 10040 | 1018 | 710 | 901 | 5627 | 113 | -- | 1274 | 994 | 774 | 347 | 745 | 1104 | 8 | 1430 | 3892 | 367 | 4409 | 725 | 7588 | 7 | 1387 | 1379 | 470 | 1312 | 4547 | 1873 | 1886 | 1131 | 599 | 9 | 5663 | 4829 | 6476 | 432 | 4810 | 145 | 174 ja | 264 | 1188 | 231 | 342 | 144 | 1017 | 953 | 3040 | 274 | 228 | 157 | 623 | 24 | 1274 | -- | 372 | 654 | 140 | 302 | 424 | 2 | 266 | 1014 | 152 | 706 | 269 | 1517 | 2 | 550 | 571 | 109 | 387 | 950 | 425 | 641 | 425 | 305 | 5 | 1242 | 1013 | 797 | 49 | 908 | 25 | 33 rn | 423 | 876 | 188 | 298 | 100 | 240 | 209 | 1878 | 735 | 183 | 485 | 489 | 107 | 994 | 372 | -- | 283 | 106 | 242 | 369 | 18 | 228 | 684 | 398 | 526 | 206 | 711 | 0 | 443 | 450 | 77 | 584 | 607 | 1186 | 521 | 363 | 149 | 13 | 724 | 610 | 617 | 59 | 631 | 20 | 180 ko | 244 | 707 | 155 | 352 | 112 | 412 | 382 | 1673 | 264 | 268 | 135 | 520 | 32 | 774 | 654 | 283 | -- | 99 | 319 | 445 | 1 | 150 | 596 | 130 | 587 | 264 | 649 | 0 | 522 | 543 | 81 | 234 | 613 | 324 | 541 | 452 | 197 | 5 | 680 | 616 | 532 | 54 | 530 | 12 | 45 ky | 92 | 299 | 221 | 154 | 58 | 139 | 131 | 490 | 124 | 106 | 61 | 234 | 16 | 347 | 140 | 106 | 99 | -- | 107 | 167 | 4 | 102 | 252 | 59 | 251 | 118 | 1013 | 1 | 206 | 211 | 45 | 145 | 279 | 150 | 206 | 174 | 109 | 3 | 346 | 508 | 270 | 113 | 201 | 12 | 23 mr | 221 | 559 | 194 | 586 | 89 | 240 | 213 | 1181 | 241 | 2091 | 159 | 3831 | 51 | 745 | 302 | 242 | 319 | 107 | -- | 630 | 1 | 232 | 608 | 138 | 524 | 1797 | 675 | 0 | 419 | 436 | 129 | 270 | 603 | 332 | 1776 | 1886 | 196 | 11 | 706 | 596 | 1395 | 79 | 473 | 16 | 48 np | 301 | 854 | 242 | 668 | 152 | 275 | 252 | 1614 | 323 | 561 | 239 | 1357 | 36 | 1104 | 424 | 369 | 445 | 167 | 630 | -- | 1 | 303 | 916 | 134 | 706 | 545 | 849 | 2 | 553 | 538 | 164 | 420 | 687 | 513 | 994 | 741 | 217 | 7 | 930 | 741 | 1156 | 84 | 719 | 39 | 65 om | 21 | 9 | 1 | 2 | 1 | 14 | 16 | 38 | 4 | 1 | 5 | 4 | 3 | 8 | 2 | 18 | 1 | 4 | 1 | 1 | -- | 2 | 3 | 11 | 4 | 6 | 8 | 0 | 2 | 3 | 0 | 6 | 7 | 5 | 2 | 2 | 1 | 103 | 5 | 10 | 1 | 4 | 2 | 0 | 7 ps | 192 | 2161 | 252 | 300 | 234 | 559 | 501 | 1522 | 196 | 246 | 229 | 1519 | 11 | 1430 | 266 | 228 | 150 | 102 | 232 | 303 | 2 | -- | 2815 | 94 | 594 | 249 | 1246 | 0 | 235 | 242 | 156 | 304 | 766 | 314 | 441 | 314 | 92 | 8 | 1049 | 818 | 2833 | 156 | 657 | 7 | 32 fa | 431 | 4186 | 817 | 790 | 426 | 1111 | 967 | 4680 | 602 | 522 | 487 | 5351 | 49 | 3892 | 1014 | 684 | 596 | 252 | 608 | 916 | 3 | 2815 | -- | 186 | 5512 | 541 | 4328 | 0 | 1028 | 1023 | 276 | 812 | 2512 | 1002 | 1250 | 797 | 364 | 8 | 3695 | 3567 | 6752 | 313 | 3190 | 66 | 74 pcm | 209 | 436 | 91 | 135 | 39 | 149 | 141 | 1074 | 439 | 101 | 529 | 192 | 255 | 367 | 152 | 398 | 130 | 59 | 138 | 134 | 11 | 94 | 186 | -- | 227 | 112 | 322 | 0 | 234 | 246 | 28 | 219 | 314 | 436 | 232 | 162 | 85 | 28 | 287 | 280 | 232 | 18 | 170 | 9 | 462 pt | 307 | 2539 | 678 | 764 | 230 | 1371 | 1271 | 4744 | 921 | 529 | 375 | 6563 | 61 | 4409 | 706 | 526 | 587 | 251 | 524 | 706 | 4 | 594 | 5512 | 227 | -- | 579 | 4452 | 7 | 1371 | 1341 | 231 | 602 | 7112 | 983 | 1042 | 820 | 468 | 3 | 3483 | 4421 | 6759 | 186 | 3754 | 110 | 97 pa | 189 | 547 | 190 | 580 | 86 | 250 | 226 | 1330 | 247 | 2210 | 157 | 4052 | 39 | 725 | 269 | 206 | 264 | 118 | 1797 | 545 | 6 | 249 | 541 | 112 | 579 | -- | 629 | 0 | 410 | 404 | 128 | 283 | 585 | 357 | 1726 | 1892 | 200 | 10 | 643 | 570 | 1515 | 73 | 431 | 16 | 44 ru | 347 | 5564 | 2238 | 838 | 535 | 2572 | 2286 | 9080 | 849 | 582 | 525 | 4622 | 79 | 7588 | 1517 | 711 | 649 | 1013 | 675 | 849 | 8 | 1246 | 4328 | 322 | 4452 | 629 | -- | 5 | 1495 | 1460 | 373 | 1166 | 4864 | 1672 | 1628 | 892 | 595 | 7 | 6223 | 22241 | 5309 | 809 | 3963 | 134 | 125 gd | 0 | 1 | 4 | 0 | 0 | 2 | 1 | 128 | 2 | 0 | 1 | 1 | 0 | 7 | 2 | 0 | 0 | 1 | 0 | 2 | 0 | 0 | 0 | 0 | 7 | 0 | 5 | -- | 2 | 3 | 2 | 1 | 3 | 1 | 0 | 0 | 1 | 0 | 6 | 5 | 2 | 1 | 3 | 36 | 2 sr | 357 | 1109 | 289 | 562 | 115 | 504 | 453 | 3760 | 555 | 331 | 258 | 809 | 51 | 1387 | 550 | 443 | 522 | 206 | 419 | 553 | 2 | 235 | 1028 | 234 | 1371 | 410 | 1495 | 2 | -- | 9041 | 127 | 377 | 1235 | 574 | 761 | 691 | 340 | 6 | 1247 | 1512 | 1021 | 109 | 685 | 42 | 69 sr | 365 | 1145 | 283 | 564 | 123 | 530 | 494 | 3809 | 569 | 345 | 258 | 807 | 51 | 1379 | 571 | 450 | 543 | 211 | 436 | 538 | 3 | 242 | 1023 | 246 | 1341 | 404 | 1460 | 3 | 9041 | -- | 137 | 382 | 1260 | 568 | 775 | 699 | 347 | 10 | 1229 | 1498 | 1009 | 112 | 639 | 45 | 79 si | 62 | 315 | 124 | 151 | 87 | 166 | 150 | 532 | 98 | 125 | 49 | 449 | 13 | 470 | 109 | 77 | 81 | 45 | 129 | 164 | 0 | 156 | 276 | 28 | 231 | 128 | 373 | 2 | 127 | 137 | -- | 137 | 260 | 189 | 348 | 173 | 69 | 7 | 301 | 306 | 510 | 38 | 216 | 5 | 15 so | 309 | 1049 | 367 | 412 | 79 | 323 | 302 | 2141 | 502 | 261 | 391 | 747 | 77 | 1312 | 387 | 584 | 234 | 145 | 270 | 420 | 6 | 304 | 812 | 219 | 602 | 283 | 1166 | 1 | 377 | 382 | 137 | -- | 689 | 1020 | 723 | 384 | 178 | 19 | 968 | 875 | 1000 | 75 | 724 | 20 | 116 es | 351 | 3654 | 704 | 701 | 431 | 2002 | 1873 | 6910 | 990 | 540 | 463 | 2931 | 91 | 4547 | 950 | 607 | 613 | 279 | 603 | 687 | 7 | 766 | 2512 | 314 | 7112 | 585 | 4864 | 3 | 1235 | 1260 | 260 | 689 | -- | 1047 | 1073 | 827 | 469 | 10 | 3645 | 3130 | 3060 | 290 | 2330 | 59 | 133 sw | 378 | 1186 | 539 | 471 | 86 | 412 | 383 | 2701 | 872 | 300 | 568 | 893 | 151 | 1873 | 425 | 1186 | 324 | 150 | 332 | 513 | 5 | 314 | 1002 | 436 | 983 | 357 | 1672 | 1 | 574 | 568 | 189 | 1020 | 1047 | -- | 929 | 492 | 261 | 10 | 1348 | 1309 | 1253 | 90 | 936 | 37 | 219 ta | 390 | 1311 | 515 | 919 | 185 | 511 | 465 | 3156 | 425 | 1762 | 299 | 3711 | 52 | 1886 | 641 | 521 | 541 | 206 | 1776 | 994 | 2 | 441 | 1250 | 232 | 1042 | 1726 | 1628 | 0 | 761 | 775 | 348 | 723 | 1073 | 929 | -- | 2278 | 400 | 14 | 1486 | 1423 | 2404 | 134 | 1092 | 32 | 68 te | 329 | 877 | 245 | 793 | 147 | 353 | 335 | 2121 | 380 | 2066 | 260 | 3762 | 54 | 1131 | 425 | 363 | 452 | 174 | 1886 | 741 | 2 | 314 | 797 | 162 | 820 | 1892 | 892 | 0 | 691 | 699 | 173 | 384 | 827 | 492 | 2278 | -- | 306 | 11 | 893 | 832 | 1748 | 107 | 644 | 21 | 61 th | 124 | 367 | 140 | 245 | 71 | 269 | 250 | 1020 | 185 | 164 | 87 | 378 | 18 | 599 | 305 | 149 | 197 | 109 | 196 | 217 | 1 | 92 | 364 | 85 | 468 | 200 | 595 | 1 | 340 | 347 | 69 | 178 | 469 | 261 | 400 | 306 | -- | 5 | 477 | 480 | 414 | 37 | 357 | 10 | 26 ti | 131 | 27 | 2 | 6 | 4 | 11 | 12 | 58 | 10 | 5 | 9 | 7 | 5 | 9 | 5 | 13 | 5 | 3 | 11 | 7 | 103 | 8 | 8 | 28 | 3 | 10 | 7 | 0 | 6 | 10 | 7 | 19 | 10 | 10 | 14 | 11 | 5 | -- | 8 | 8 | 4 | 2 | 5 | 0 | 6 tr | 435 | 4147 | 1495 | 860 | 449 | 1511 | 1294 | 5676 | 829 | 631 | 519 | 3694 | 91 | 5663 | 1242 | 724 | 680 | 346 | 706 | 930 | 5 | 1049 | 3695 | 287 | 3483 | 643 | 6223 | 6 | 1247 | 1229 | 301 | 968 | 3645 | 1348 | 1486 | 893 | 477 | 8 | -- | 4108 | 4340 | 370 | 2981 | 126 | 130 uk | 345 | 3457 | 1383 | 688 | 350 | 1619 | 1464 | 6562 | 721 | 508 | 400 | 3935 | 83 | 4829 | 1013 | 610 | 616 | 508 | 596 | 741 | 10 | 818 | 3567 | 280 | 4421 | 570 | 22241 | 5 | 1512 | 1498 | 306 | 875 | 3130 | 1309 | 1423 | 832 | 480 | 8 | 4108 | -- | 4290 | 442 | 3017 | 108 | 89 ur | 409 | 4935 | 966 | 1382 | 591 | 1651 | 1444 | 6320 | 766 | 1619 | 526 | 15666 | 61 | 6476 | 797 | 617 | 532 | 270 | 1395 | 1156 | 1 | 2833 | 6752 | 232 | 6759 | 1515 | 5309 | 2 | 1021 | 1009 | 510 | 1000 | 3060 | 1253 | 2404 | 1748 | 414 | 4 | 4340 | 4290 | -- | 389 | 3723 | 72 | 88 uz | 41 | 388 | 199 | 98 | 62 | 176 | 158 | 450 | 76 | 80 | 59 | 352 | 15 | 432 | 49 | 59 | 54 | 113 | 79 | 84 | 4 | 156 | 313 | 18 | 186 | 73 | 809 | 1 | 109 | 112 | 38 | 75 | 290 | 90 | 134 | 107 | 37 | 2 | 370 | 442 | 389 | -- | 257 | 10 | 15 vi | 285 | 2666 | 726 | 527 | 447 | 1858 | 1663 | 4575 | 438 | 450 | 352 | 3738 | 65 | 4810 | 908 | 631 | 530 | 201 | 473 | 719 | 2 | 657 | 3190 | 170 | 3755 | 431 | 3963 | 3 | 685 | 639 | 216 | 724 | 2330 | 936 | 1092 | 644 | 357 | 5 | 2982 | 3017 | 3723 | 257 | -- | 106 | 76 cy | 1 | 38 | 30 | 37 | 4 | 33 | 31 | 2655 | 40 | 21 | 30 | 77 | 6 | 145 | 25 | 20 | 12 | 12 | 16 | 39 | 0 | 7 | 66 | 9 | 110 | 16 | 134 | 36 | 42 | 45 | 5 | 20 | 59 | 37 | 32 | 21 | 10 | 0 | 126 | 108 | 72 | 10 | 106 | -- | 8 yo | 67 | 141 | 42 | 61 | 12 | 39 | 38 | 229 | 159 | 54 | 362 | 79 | 296 | 174 | 33 | 180 | 45 | 23 | 48 | 65 | 7 | 32 | 74 | 462 | 97 | 44 | 125 | 2 | 69 | 79 | 15 | 116 | 133 | 219 | 68 | 61 | 26 | 6 | 130 | 89 | 88 | 15 | 76 | 8 | -- ## Dataset Creation ### Curation Rationale [More information needed](https://github.com/csebuetnlp/CrossSum) ### Source Data [BBC News](https://www.bbc.co.uk/ws/languages) #### Initial Data Collection and Normalization [Detailed in the paper](https://arxiv.org/abs/2112.08804/) #### Who are the source language producers? [Detailed in the paper](https://arxiv.org/abs/2112.08804/) ### Annotations [Detailed in the paper](https://arxiv.org/abs/2112.08804/) #### Annotation process [Detailed in the paper](https://arxiv.org/abs/2112.08804/) #### Who are the annotators? [Detailed in the paper](https://arxiv.org/abs/2112.08804/) ### Personal and Sensitive Information [More information needed](https://github.com/csebuetnlp/CrossSum) ## Considerations for Using the Data ### Social Impact of Dataset [More information needed](https://github.com/csebuetnlp/CrossSum) ### Discussion of Biases [More information needed](https://github.com/csebuetnlp/CrossSum) ### Other Known Limitations [More information needed](https://github.com/csebuetnlp/CrossSum) ## Additional Information ### Dataset Curators [More information needed](https://github.com/csebuetnlp/CrossSum) ### Licensing Information Contents of this repository are restricted to only non-commercial research purposes under the [Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License (CC BY-NC-SA 4.0)](https://creativecommons.org/licenses/by-nc-sa/4.0/). Copyright of the dataset contents belongs to the original copyright holders. ### Citation Information If you use any of the datasets, models or code modules, please cite the following paper: ``` @article{hasan2021crosssum, author = {Tahmid Hasan and Abhik Bhattacharjee and Wasi Uddin Ahmad and Yuan-Fang Li and Yong-bin Kang and Rifat Shahriyar}, title = {CrossSum: Beyond English-Centric Cross-Lingual Abstractive Text Summarization for 1500+ Language Pairs}, journal = {CoRR}, volume = {abs/2112.08804}, year = {2021}, url = {https://arxiv.org/abs/2112.08804}, eprinttype = {arXiv}, eprint = {2112.08804} } ``` ### Contributions Thanks to [@abhik1505040](https://github.com/abhik1505040) and [@Tahmid](https://github.com/Tahmid04) for adding this dataset.
csebuetnlp/CrossSum
[ "task_categories:summarization", "task_ids:news-articles-summarization", "annotations_creators:found", "language_creators:found", "multilinguality:multilingual", "size_categories:1M<n<10M", "source_datasets:original", "language:am", "language:ar", "language:az", "language:bn", "language:my", "language:zh", "language:en", "language:fr", "language:gu", "language:ha", "language:hi", "language:ig", "language:id", "language:ja", "language:rn", "language:ko", "language:ky", "language:mr", "language:ne", "language:om", "language:ps", "language:fa", "language:pcm", "language:pt", "language:pa", "language:ru", "language:gd", "language:sr", "language:si", "language:so", "language:es", "language:sw", "language:ta", "language:te", "language:th", "language:ti", "language:tr", "language:uk", "language:ur", "language:uz", "language:vi", "language:cy", "language:yo", "license:cc-by-nc-sa-4.0", "arxiv:2112.08804", "region:us" ]
2022-04-20T07:27:10+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["am", "ar", "az", "bn", "my", "zh", "en", "fr", "gu", "ha", "hi", "ig", "id", "ja", "rn", "ko", "ky", "mr", "ne", "om", "ps", "fa", "pcm", "pt", "pa", "ru", "gd", "sr", "si", "so", "es", "sw", "ta", "te", "th", "ti", "tr", "uk", "ur", "uz", "vi", "cy", "yo"], "license": ["cc-by-nc-sa-4.0"], "multilinguality": ["multilingual"], "size_categories": ["1M<n<10M"], "source_datasets": ["original"], "task_categories": ["summarization"], "task_ids": ["news-articles-summarization"], "pretty_name": "CrossSum"}
2023-12-16T13:27:47+00:00
1436852d084f57851cb02ae13e16e4fa23c1e977
UmnatHCU/test
[ "license:afl-3.0", "region:us" ]
2022-04-20T07:30:01+00:00
{"license": "afl-3.0"}
2022-04-20T07:30:01+00:00
fdf84275bb8ce4b49c971d02e84dd1abc677a50f
mteb/sts12-sts
[ "language:en", "region:us" ]
2022-04-20T09:47:29+00:00
{"language": ["en"]}
2022-09-27T18:11:50+00:00
1591bfcbe8c69d4bf7fe2a16e2451017832cafb9
mteb/sts13-sts
[ "language:en", "region:us" ]
2022-04-20T09:47:41+00:00
{"language": ["en"]}
2022-09-27T18:12:02+00:00
e2125984e7df8b7871f6ae9949cf6b6795e7c54b
mteb/sts14-sts
[ "language:en", "region:us" ]
2022-04-20T09:47:52+00:00
{"language": ["en"]}
2022-09-27T18:11:37+00:00
1cd7298cac12a96a373b6a2f18738bb3e739a9b6
mteb/sts15-sts
[ "language:en", "region:us" ]
2022-04-20T09:48:04+00:00
{"language": ["en"]}
2022-09-27T18:12:14+00:00
360a0b2dff98700d09e634a01e1cc1624d3e42cd
mteb/sts16-sts
[ "language:en", "region:us" ]
2022-04-20T09:48:15+00:00
{"language": ["en"]}
2022-09-27T18:12:09+00:00
1f93eb3df343353f9b0d0f9bc724ab9473643bfe
### Dataset Summary This dataset is extracted from Climate Text dataset (https://www.sustainablefinance.uzh.ch/en/research/climate-fever/climatext.html), pre-processed and, ready to evaluate. The evaluation objective is a text classification task - given a climate related claim and evidence, predict if claim is related to evidence.
mwong/climatetext-claim-related-evaluation
[ "task_categories:text-classification", "task_ids:fact-checking", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:extended|climate_text", "language:en", "license:cc-by-sa-3.0", "license:gpl-3.0", "region:us" ]
2022-04-20T11:00:50+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["cc-by-sa-3.0", "gpl-3.0"], "multilinguality": ["monolingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["extended|climate_text"], "task_categories": ["text-classification"], "task_ids": ["fact-checking"]}
2022-10-25T09:08:44+00:00