text-classification
bool
2 classes
text
stringlengths
0
664k
true
# Dataset Card for MultiRC_TH_drop ### Dataset Description This dataset is Thai translated version of [multirc](https://huggingface.co/datasets/super_glue/viewer/multirc) using google translate with [Multilingual Universal Sentence Encoder](https://arxiv.org/abs/1907.04307) to calculate score for Thai translation. Th...
false
## DreamEditBench for Subject Replacement task and Subject Addition task. ## Dataset Description - **Homepage:** https://dreameditbenchteam.github.io - **Repository:** https://github.com/DreamEditBenchTeam/DreamEdit <!-- **Paper:** https://arxiv.org/abs/2306.12624 --> The goal of subject replacement is to replace a...
true
# AutoTrain Dataset for project: based-in-fact ## Dataset Description This dataset has been automatically processed by AutoTrain for project based-in-fact. ### Languages The BCP-47 code for the dataset's language is unk. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```jso...
false
# DEplain-web-sent: A corpus for German Sentence Simplification DEplain-web-sent is a subcorpus of DEplain [Stodden et al., 2023]((https://arxiv.org/abs/2305.18939)) for evaluation of sentence simplification. The corpus consists of 1846 sentence pairs of 147 parallel documents crawled from the web in standard German...
true
false
# Dataset Card for "xglm_enth2" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
true
# Dataset Card for "articles_dataset" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
false
# MF Rocket Paraphrase Corpus (MFRPC) - A State of the Art Paraphrasing Solution ## Dataset Description MF Rocket Paraphrase Corpus (MFRPC) ) is a corpus consisting of 10,000 sentence pairs. Each sentence pair contains a source sentence and the paraphrased version of the source sentence. The source sentences are crea...
false
This is a copy of the [Multi-XScience](https://huggingface.co/datasets/multi_x_science_sum) dataset, except the input source documents of its `test` split have been replaced by a __dense__ retriever. The retrieval pipeline used: - __query__: The `related_work` field of each example - __corpus__: The union of all docu...
false
This is a copy of the [Multi-XScience](https://huggingface.co/datasets/multi_x_science_sum) dataset, except the input source documents of its `train`, `validation` and `test` splits have been replaced by a __dense__ retriever. The retrieval pipeline used: - __query__: The `related_work` field of each example - __corp...
false
This is a copy of the [Multi-XScience](https://huggingface.co/datasets/multi_x_science_sum) dataset, except the input source documents of the `train`, `validation`, and `test` splits have been replaced by a __dense__ retriever. The retrieval pipeline used: - __query__: The `related_work` field of each example - __cor...
false
This is a copy of the [Cochrane](https://huggingface.co/datasets/allenai/mslr2022) dataset, except the input source documents of its `validation` split have been replaced by a __dense__ retriever. The retrieval pipeline used: - __query__: The `target` field of each example - __corpus__: The union of all documents in ...
false
This is a copy of the [Cochrane](https://huggingface.co/datasets/allenai/mslr2022) dataset, except the input source documents of the `train`, `validation`, and `test` splits have been replaced by a __dense__ retriever. - __query__: The `target` field of each example - __corpus__: The union of all documents in the `tr...
false
This is a copy of the [MS^2](https://huggingface.co/datasets/allenai/mslr2022) dataset, except the input source documents of its `validation` split have been replaced by a __dense__ retriever. The retrieval pipeline used: - __query__: The `background` field of each example - __corpus__: The union of all documents in ...
false
This is a copy of the [MS^2](https://huggingface.co/datasets/allenai/mslr2022) dataset, except the input source documents of the `train`, `validation`, and `test` splits have been replaced by a __dense__ retriever. - __query__: The `background` field of each example - __corpus__: The union of all documents in the `tr...
false
This is a copy of the [WCEP-10](https://huggingface.co/datasets/ccdv/WCEP-10) dataset, except the input source documents of its `test` split have been replaced by a __dense__ retriever. The retrieval pipeline used: - __query__: The `summary` field of each example - __corpus__: The union of all documents in the `train...
false
This is a copy of the [WCEP-10](https://huggingface.co/datasets/ccdv/WCEP-10) dataset, except the input source documents of the `train`, `validation`, and `test` splits have been replaced by a __dense__ retriever. The retrieval pipeline used: - __query__: The `summary` field of each example - __corpus__: The union of...
false
This is a copy of the [WCEP-10](https://huggingface.co/datasets/ccdv/WCEP-10) dataset, except the input source documents of its `train`, `validation, and `test` splits have been have been replaced by a __dense__ retriever. The retrieval pipeline used: - __query__: The `summary` field of each example - __corpus__: The...
false
This is a copy of the [Multi-News](https://huggingface.co/datasets/multi_news) dataset, except the input source documents of its `test` split have been replaced by a __dense__ retriever. The retrieval pipeline used: - __query__: The `summary` field of each example - __corpus__: The union of all documents in the `trai...
false
This is a copy of the [Multi-News](https://huggingface.co/datasets/multi_news) dataset, except the input source documents of its `train`, `validation` and `test` splits have been replaced by a __dense__ retriever. The retrieval pipeline used: - __query__: The `summary` field of each example - __corpus__: The union of...
false
This is a copy of the [Multi-News](https://huggingface.co/datasets/multi_news) dataset, except the input source documents of the `train`, `validation`, and `test` splits have been replaced by a __dense__ retriever. The retrieval pipeline used: - __query__: The `summary` field of each example - __corpus__: The union o...
false
# Dataset Card for [for-ULPGL-Dissertation] ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#...
false
# Dataset Card for Racó Forums Corpus ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#datase...
false
# About Dataset The dataset consists of data from a bunch of youtube videos ranging from videos from fastai lessons, FSDL lesson to random videos teaching something. In total this dataset contains 600 chapter markers in youtube and contains 25, 000 lesson transcript. This dataset can be used for NLP tasks like summ...
false
# Dataset Card for "CompanyWeb" ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-stru...
false
false
# laion2b_multi_korean_subset_with_image ## Dataset Description - **Download Size** 342 GB img2dataset을 통해 다운로드에 성공한 [Bingsu/laion2B-multi-korean-subset](https://huggingface.co/datasets/Bingsu/laion2B-multi-korean-subset) 이미지를 정리한 데이터셋입니다. 이미지는 9,800,137장입니다. 이미지는 짧은 쪽 길이가 256이 되도록 리사이즈 되었으며, 품질 100인 webp파일로 다운로드 ...
true
# Dataset Card for SloIE ### Dataset Summary SloIE is a manually labelled dataset of Slovene idiomatic expressions. It contains 29399 sentences with 75 different expressions that can occur with either a literal or an idiomatic meaning, with appropriate manual annotations for each token. The idiomatic expressions wer...
true
# AutoTrain Dataset for project: ethos-sentiments ## Dataset Description This dataset has been automatically processed by AutoTrain for project ethos-sentiments. ### Languages The BCP-47 code for the dataset's language is en. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: `...
true
# Dataset Card for `reviews_with_drift` ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [language](#language) - [Dataset Structure](#datase...
true
# Dataset Card for Kill-Me-Please Dataset ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-split...
false
# AutoTrain Dataset for project: swin-muppet ## Dataset Description This dataset has been automatically processed by AutoTrain for project swin-muppet. ### Languages The BCP-47 code for the dataset's language is unk. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json [ ...
false
# Dataset Card for Thesaurus of Modern Slovene 1.0 Also known as "Sopomenke 1.0". Available in application form online: https://viri.cjvt.si/sopomenke/slv/. ### Dataset Summary This is an automatically created Slovene thesaurus from Slovene data available in a comprehensive English–Slovenian dictionary, a monolingu...
false
# Dataset Card for FB15k-237 ## Table of Contents - [Dataset Card for FB15k-237](#dataset-card-for-fb15k-237) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards...
false
# Dataset Card for SloWNet ### Dataset Summary sloWNet is the Slovene WordNet developed in the expand approach: it contains the complete Princeton WordNet 3.0 and over 70 000 Slovene literals. These literals have been added automatically using different types of existing resources, such as bilingual dictionaries, pa...
false
# Dataset Card for CrossWOZ - **Repository:** https://github.com/thu-coai/CrossWOZ - **Paper:** https://aclanthology.org/2020.tacl-1.19/ - **Leaderboard:** None - **Who transforms the dataset:** Qi Zhu(zhuq96 at gmail dot com) To use this dataset, you need to install [ConvLab-3](https://github.com/ConvLab/ConvLab-3)...
false
# Dataset Card for WSDMCup2023 ## Dataset Description - **Homepage:** [Toloka Visual Question Answering Challenge](https://toloka.ai/challenges/wsdm2023) - **Repository:** [WSDM Cup 2023 Starter Pack](https://github.com/Toloka/WSDMCup2023) - **Paper:** - **Leaderboard:** [CodaLab Competition Leaderboard](https://co...
false
true
# Dataset Card for VUA Metaphor Corpus **Important note#1**: This is a slightly simplified but mostly complete parse of the corpus. What is missing are lemmas and some metadata that was not important at the time of writing the parser. See the section `Simplifications` for more information on this. **Important note...
true
# Dataset Card for GuiaCat ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure...
false
# Dataset Card for Quasimodo ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-struct...
true
false
# Dataset Card for Wine Recognition dataset ## Dataset Description - **Homepage:** https://archive.ics.uci.edu/ml/datasets/wine - **Papers:** 1. S. Aeberhard, D. Coomans and O. de Vel, Comparison of Classifiers in High Dimensional Settings, Tech. Rep. no. 92-02, (1992), Dept. of Computer Science ...
false
# Dataset Card for OLM May 2017 Common Crawl Cleaned and deduplicated pretraining dataset, created with the OLM repo [here](https://github.com/huggingface/olm-datasets) from 16% of the May 2017 Common Crawl snapshot. Note: `last_modified_timestamp` was parsed from whatever a website returned in it's `Last-Modified` he...
false
# Dataset Card for V4Design Europeana style dataset ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Stru...
false
# Dataset Card for Early Printed Books Font Detection Dataset ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Da...
false
# AutoTrain Dataset for project: lojban-translation ## Dataset Description This dataset has been automatically processed by AutoTrain for project lojban-translation. ### Languages The BCP-47 code for the dataset's language is en2jb. ## Dataset Structure ### Data Instances A sample from this dataset looks as foll...
false
# Dataset Card for [naacl2022] ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-struc...
false
# AfroLM: A Self-Active Learning-based Multilingual Pretrained Language Model for 23 African Languages - [GitHub Repository of the Paper](https://github.com/bonaventuredossou/MLM_AL) This repository contains the dataset for our paper [`AfroLM: A Self-Active Learning-based Multilingual Pretrained Language Model for 23 ...
false
# Dataset Card for mt_en_it ## Table of Contents - [Dataset Card for mt_en_it](#dataset-card-for-mt-en-it) - [Table of Contents](#table-of-contents) - [Dataset Summary](#dataset-summary) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data ...
true
# Dataset Card for GLUE ## Table of Contents - [Dataset Card for GLUE](#dataset-card-for-glue) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [ax](...
false
false
# Dataset Card for Pokémon BLIP captions with English and Japanese. Dataset used to train Pokémon text to image model, add a Japanese Column of [Pokémon BLIP captions](https://huggingface.co/datasets/lambdalabs/pokemon-blip-captions) BLIP generated captions for Pokémon images from Few Shot Pokémon dataset introduced...
true
false
# AutoTrain Dataset for project: oaoqoqkaksk ## Dataset Description This dataset has been automatically processed by AutoTrain for project oaoqoqkaksk. ### Languages The BCP-47 code for the dataset's language is en2nl. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json ...
true
# Dataset Card for openpi_v2 ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structu...
true
# MNLI Norwegian The Multi-Genre Natural Language Inference (MultiNLI) corpus is a crowd-sourced collection of 433k sentence pairs annotated with textual entailment information. The corpus is modeled on the SNLI corpus, but differs in that it covers a range of genres of spoken and written text, and supports a distinct...
false
# AutoTrain Dataset for project: testtextexists ## Dataset Description This dataset has been automatically processed by AutoTrain for project testtextexists. ### Languages The BCP-47 code for the dataset's language is en. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```js...
false
# Dataset Card for "lmqg/qa_squadshifts" ## Dataset Description - **Repository:** [https://github.com/asahi417/lm-question-generation](https://github.com/asahi417/lm-question-generation) - **Paper:** [https://arxiv.org/abs/2004.14444](https://arxiv.org/abs/2004.14444) - **Point of Contact:** [Asahi Ushio](http://asah...
false
# Dataset Card for "eclassQuery" This Dataset consists of paraphrases of ECLASS-standard pump-properties. It can be used to evaluate models on the task of matching these paraphrases to the actual ECLASS-standard pump-properties based on their semantics.
true
true
# Dataset Card for PubMed ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [...
true
# Dataset Card for Machine Paraphrase Dataset (MPC) ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Stru...
false
# Dataset Card for DBLP Discovery Dataset (D3) ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](...
false
# Dataset Card for laion2B-multi-turkish-subset ## Dataset Description - **Homepage:** [laion-5b](https://laion.ai/blog/laion-5b/) - **Huggingface:** [laion/laion2B-multi](https://huggingface.co/datasets/laion/laion2B-multi) - **Point of Contact:** [mcemilg](mailto:mcg@mcemilg.dev) ### Dataset Summary [LAION-5B](h...
false
# AutoTrain Dataset for project: led-samsum-dialogsum ## Dataset Description This dataset has been automatically processed by AutoTrain for project led-samsum-dialogsum. ### Languages The BCP-47 code for the dataset's language is unk. ## Dataset Structure ### Data Instances A sample from this dataset looks as fo...
false
Over 20,000 256x256 mel spectrograms of 5 second samples of music from my Spotify liked playlist. The code to convert from audio to spectrogram and vice versa can be found in https://github.com/teticio/audio-diffusion along with scripts to train and run inference using De-noising Diffusion Probabilistic Models. ``` x_r...
false
# Dataset Card for "bill_summarization" ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#data...
false
# Dataset Card for "lmqg/qa_harvesting_from_wikipedia_pseudo" ## Dataset Description - **Repository:** [https://github.com/asahi417/lm-question-generation](https://github.com/asahi417/lm-question-generation) - **Paper:** [https://arxiv.org/abs/2210.03992](https://arxiv.org/abs/2210.03992) - **Point of Contact:** [Asa...
false
Dataset contains queries for Problog database of facts about USA geography. Taken from [this source](https://www.cs.utexas.edu/users/ml/nldata/geoquery.html)
false
# Dataset Card for "saf_micro_job_german" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data...
true
# TyDiP: A Dataset for Politeness Classification in Nine Typologically Diverse Languages This repo contains the code and data for the EMNLP 2022 findings paper TyDiP: A Dataset for Politeness Classification in Nine Typologically Diverse Languages which can be found [here](https://aclanthology.org/2022.findings-emnlp.42...
true
# AutoTrain Dataset for project: tweet-es-sent ## Dataset Description This dataset has been automatically processed by AutoTrain for project tweet-es-sent. ### Languages The BCP-47 code for the dataset's language is unk. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```jso...
true
# AutoTrain Dataset for project: compliance ## Dataset Description This dataset has been automatically processed by AutoTrain for project compliance. ### Languages The BCP-47 code for the dataset's language is en. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json [ {...
true
# AutoTrain Dataset for project: demo_compliance ## Dataset Description This dataset has been automatically processed by AutoTrain for project demo_compliance. ### Languages The BCP-47 code for the dataset's language is en. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```...
false
# Dataset Card for Leipzig Corpora Swiss German ## Dataset Description - **Homepage:** https://wortschatz.uni-leipzig.de/en/download/Swiss%20German - **Repository:** https://huggingface.co/datasets/statworx/leipzip-swiss ### Dataset Summary Swiss German Wikipedia corpus based on material from 2021. The corpus g...
false
# Dataset Card for Danish WIT ## Dataset Description - **Repository:** <https://gist.github.com/saattrupdan/bb6c9c52d9f4b35258db2b2456d31224> - **Point of Contact:** [Dan Saattrup Nielsen](mailto:dan.nielsen@alexandra.dk) - **Size of downloaded dataset files:** 7.5 GB - **Size of the generated dataset:** 7.8 GB - **...
false
# Dataset Card for MMCRSC ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure)...
false
# Gold standards and outputs ## Dataset Description - MapReader’s GitHub: https://github.com/Living-with-machines/MapReader - MapReader paper: https://dl.acm.org/doi/10.1145/3557919.3565812 - Zenodo link for gold standards and outputs: https://doi.org/10.5281/zenodo.7147906 - Contacts: Katherine McDonough, The Alan...
false
# Dataset Card for "diana_uribe" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
false
# AutoTrain Dataset for project: mm ## Dataset Description This dataset has been automatically processed by AutoTrain for project mm. ### Languages The BCP-47 code for the dataset's language is unk. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json [ { "text": "E...
false
# Romanian paraphrase dataset This data set was created by me, special for paraphrase [t5-small-paraphrase-ro](https://huggingface.co/BlackKakapo/t5-small-paraphrase-ro) [t5-small-paraphrase-ro-v2](https://huggingface.co/BlackKakapo/t5-small-paraphrase-ro-v2) [t5-base-paraphrase-ro](https://huggingface.co/BlackKakapo/...
false
# Romanian grammar dataset This data set was created by me, special for grammar Here you can find: ~1600k examples of grammar (TRAIN). ~220k examples of grammar (TEST).
false
# Dataset Card for "whisper-transcripts-linustechtips" ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Da...
true
# Dataset Card for [Dataset Name] ## Table of Contents - [Dataset Card for [Dataset Name]](#dataset-card-for-dataset-name) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-l...
false
# Dataset Card for Danish WIT ## Dataset Description - **Repository:** <https://gist.github.com/saattrupdan/bb6c9c52d9f4b35258db2b2456d31224> - **Point of Contact:** [Dan Saattrup Nielsen](mailto:dan.nielsen@alexandra.dk) - **Size of downloaded dataset files:** 7.5 GB - **Size of the generated dataset:** 7.8 GB - **...
false
# Dataset Card for [Dataset Name] ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-st...
false
# Dataset Card for "ChristmasClaymation-blip-captions" All captions end with the suffix ", Christmas claymation style"
false
# IMaSC: ICFOSS Malayalam Speech Corpus **IMaSC** is a Malayalam text and speech corpus made available by [ICFOSS](https://icfoss.in/) for the purpose of developing speech technology for Malayalam, particularly text-to-speech. The corpus contains 34,473 text-audio pairs of Malayalam sentences spoken by 8 speakers, to...
false
false
# Dataset Summary AfriCLIRMatrix is a test collection for cross-lingual information retrieval research in 15 diverse African languages. This resource comprises English queries with query–document relevance judgments in 15 African languages automatically mined from Wikipedia This dataset stores documents of AfriCLIRMa...
false
# Dataset Card for MultiLegalPile_Wikipedia_Filtered: A filtered version of the MultiLegalPile dataset, together with wikipedia articles ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Le...
false
**Homepage:** https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-56 Used lydfiler_16_1.tar.gz and metadata_se_csv.zip
false
# Dataset Card for "Reddit Haiku" This dataset contains haikus from the subreddit [/r/haiku](https://www.reddit.com/r/haiku/) scraped and filtered between October 19th and 10th 2022, combined with a [previous dump](https://zissou.infosci.cornell.edu/convokit/datasets/subreddit-corpus/corpus-zipped/hackintosh_ja~-~ham...
false
# Dataset Card for "Yannic-Kilcher" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
false
# Dataset Card for "azure" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
false
_The Dataset Teaser is now enabled instead! Isn't this better?_ ![preview of all texture sets](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/teaser.webp) # TD 01: Natural Ground Textures This dataset contains multi-photo texture captures in outdoor nature scenes — all...
true
# AutoTrain Dataset for project: autotrain_goodreads_string ## Dataset Description This dataset has been automatically processed by AutoTrain for project autotrain_goodreads_string. ### Languages The BCP-47 code for the dataset's language is en. ## Dataset Structure ### Data Instances A sample from this dataset ...
false
# Dataset Card for "two-minute-papers" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
false
# Dataset Card for "goog-tech-talks" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)