text-classification
bool
2 classes
text
stringlengths
0
664k
false
# Shellcode_IA32 ___Shellcode_IA32___ is a dataset containing _20_ years of shellcodes from a variety of sources is the largest collection of shellcodes in assembly available to date. This dataset consists of 3,200 examples of instructions in assembly language for _IA-32_ (the 32-bit version of the x86 Intel Archite...
false
# Dataset Card for **cointegrated/ru-paraphrase-NMT-Leipzig** ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Da...
false
## Dataset Description - **Homepage:** https://www.caito.de/2019/01/the-m-ailabs-speech-dataset/ ### Dataset Summary The M-AILABS Speech Dataset is the first large dataset that we are providing free-of-charge, freely usable as training data for speech recognition and speech synthesis. Most of the data is based on L...
false
# Dataset Card for librispeech_asr ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-insta...
true
# AutoNLP Dataset for project: vaccinfaq ## Table of content - [Dataset Description](#dataset-description) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) ## Dataset Descritpion This dataset ...
false
## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Dataset Structure](#dataset-structure) - [Dataset Evaluation](#dataset-evaluation) - [Additional Information](#additional-information) ## Dataset Description We prov...
false
# Dataset Card for Monolingual-Quechua-IIC ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#da...
true
true
# References - [alper bayram](https://github.com/alperbayram)
true
# AutoNLP Dataset for project: user-review-classification ## Table of content - [Dataset Description](#dataset-description) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) ## Dataset Descritpi...
false
# Dataset Card for common_language ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-insta...
false
# PIE synthetic dataset Repo: https://github.com/awasthiabhijeet/PIE Paper: https://aclanthology.org/D19-1435.pdf
true
# AutoNLP Dataset for project: antisemitism-2 ## Table of content - [Dataset Description](#dataset-description) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) ## Dataset Descritpion This dat...
false
# AspectEmo ## Description AspectEmo Corpus is an extended version of a publicly available PolEmo 2.0 corpus of Polish customer reviews used in many projects on the use of different methods in sentiment analysis. The AspectEmo corpus consists of four subcorpora, each containing online customer reviews from the follo...
false
# nkjp-pos ## Description NKJP-POS is a part the National Corpus of Polish (*Narodowy Korpus Języka Polskiego*). Its objective is part-of-speech tagging, e.g. nouns, verbs, adjectives, adverbs, etc. During the creation of corpus, texts of were annotated by humans from various sources, covering many domains and genre...
true
The COPA-HR dataset (Choice of plausible alternatives in Croatian) is a translation of the English COPA dataset (https://people.ict.usc.edu/~gordon/copa.html) by following the XCOPA dataset translation methodology (https://arxiv.org/abs/2005.00333). The dataset consists of 1000 premises (My body cast a shadow over t...
false
The dataset contains 6273 training samples, 762 validation samples and 749 test samples. Each sample represents a sentence and includes the following features: sentence ID ('sent\_id'), list of tokens ('tokens'), list of normalised word forms ('norms'), list of lemmas ('lemmas'), list of Multext-East tags ('xpos\_ta...
false
This dataset is based on 3,748 Serbian tweets that were segmented into sentences, tokens, and annotated with normalized forms, lemmas, MULTEXT-East tags (XPOS), UPOS tags and morphological features, and named entities. The dataset contains 5462 training samples (sentences), 711 validation samples and 725 test samples....
false
# Dataset Card for TV3Parla ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structur...
false
## Dataset Summary We introduce a Romanian IT Dataset (RoITD) resembling SQuAD 1.1. RoITD consists of 9575 Romanian QA pairs formulated by crowd workers. QA pairs are based on 5043 articles from Romanian Wikipedia articles describing IT and household products. Of the total number of questions, 5103 are possible...
true
# Dataset Card for "quora_swe" The dataset quora_swe is a subset of the automatically translated (MNT) Swedish Semantic Textual Similarity dataset: quora-deduplicates .
true
### KR3: Korean Restaurant Reviews with Ratings Korean sentiment classification dataset - Size: 460K(+180K) - Language: Korean-centric ### ⚠️ Caution with `Rating` Column 0 stands for negative review, 1 stands for positive review, and 2 stands for ambiguous review. **Note that rating 2 is not intended to be use...
false
# MERLIN corpus Project URL: https://merlin-platform.eu/C_mcorpus.php Dataset URL: https://clarin.eurac.edu/repository/xmlui/handle/20.500.12124/6 The MERLIN corpus is a written learner corpus for Czech, German, and Italian that has been designed to illustrate the Common European Framework of Reference for Language...
false
This dataset is based on 3,871 Croatian tweets that were segmented into sentences, tokens, and annotated with normalized forms, lemmas, MULTEXT-East tags (XPOS), UPOS tags and morphological features, and named entities. The dataset contains 6339 training samples (sentences), 815 validation samples and 785 test samples...
false
# SAE-door-abstracts This dataset includes ~1,550 texts of abstracts of technical papers and journal articles from the SAE Mobilus database that cover the topics of automotive or aerospace doors, noise, acoustics, and vibrations.
true
# Dataset Card for WANLI ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) ...
false
# Dataset Card for MASC: MASSIVE ARABIC SPEECH CORPUS ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset St...
false
# Dataset Card for SAT ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) ...
false
## Dataset Description A small subset of [the-stack](https://huggingface.co/datasets/bigcode/the-stack) dataset, with 87 programming languages, each has 10,000 random samples from the original dataset. ## Languages The dataset contains 87 programming languages: ```` 'ada', 'agda', 'alloy', 'antlr', 'applescript'...
false
# Open Instruct V1 - A dataset for having LLMs follow instructions. Open Instruct V1 is an amalgamation of different datasets which are cleaned and then collated into a singular format for training. ## Dataset Breakdown | Dataset | Amount of Samples | |----------------|-------------------| | [Alpaca](https:/...
true
# Dataset Card for Vietnamese Students’ Feedback Corpus ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset ...
false
# Dataset Card for Catalan General Crawling ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#...
false
# Dataset Card for Greek Legal Named Entity Recognition ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset ...
false
# Dataset Card for "tner/mit_movie_trivia" ## Dataset Description - **Repository:** [T-NER](https://github.com/asahi417/tner) - **Dataset:** MIT Movie - **Domain:** Movie - **Number of Entity:** 12 ### Dataset Summary MIT Movie NER dataset formatted in a part of [TNER](https://github.com/asahi417/tner) project. - ...
false
# Dataset Card for Dataset Name ## Dataset Description - **Homepage:** - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary The dataset.json file contains ~1.7 million synthetic data for arithmetic tasks, generated by dataset.ipynb. ### Supported Tasks and Leaderboards...
false
# The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions - Original Paper and Dataset [here](https://dataverse.harvard.edu/dataset.xhtml?persistentId=doi:10.7910/DVN/DBW86T) - Kaggle dataset [here](https://www.kaggle.com/datasets/kmader/skin-cancer-mnist-ham10000...
true
# Dataset Card for ProsocialDialog Dataset ## Dataset Description - **Repository:** [Dataset and Model](https://github.com/skywalker023/prosocial-dialog) - **Paper:** [ProsocialDialog: A Prosocial Backbone for Conversational Agents](https://aclanthology.org/2022.emnlp-main.267/) - **Point of Contact:** [Hyunwoo Kim](...
false
Source of data: https://github.com/FudanVI/benchmarking-chinese-text-recognition
false
# **CAMEL: Communicative Agents for “Mind” Exploration of Large Scale Language Model Society** - **Github:** https://github.com/lightaime/camel - **Website:** https://www.camel-ai.org/ - **Arxiv Paper:** https://arxiv.org/abs/2303.17760 ## Dataset Summary Math dataset is composed of 50K problem-solution pairs obtain...
false
false
# Titanic The [Titanic dataset](https://www.kaggle.com/datasets/vinicius150987/titanic3) from [Kaggle](https://www.kaggle.com/). # Configurations and tasks | **Configuration** | **Task** | **Description** | |-------------------|---------------------------|----------------------------| | sur...
false
# Dataset Card for TED descriptions [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
false
# Dataset Card for Dataset Name ## Dataset Description - **Homepage:** - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary This dataset card aims to be a base template for new datasets. It has been generated using [this raw template](https://github.com/huggingface/hugg...
false
# Dataset Card for BEIR Benchmark ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instan...
false
# Dataset Card for 🥤SODA Synthetic Dialogue ## Dataset Description - **Repository:** [Code](https://github.com/LAION-AI/Open-Assistant/openassistant/datasets/soda_synthetic_dialogue/) ## Dataset Summary 🥤SODA Synthetic Dialogue is a set of synthetic dialogues between Assistant and User. In each conversation, Us...
false
# Dataset Card for ontonotes_english ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset...
false
# Dataset Card for BEIR Benchmark ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instan...
true
# Dataset Card for BIOSSES ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) ...
true
# Dataset Card for "kbp37" ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) ...
true
# ChatGPT-Research-Abstracts This is a dataset created in relation to a bachelor thesis written by Nicolai Thorer Sivesind and Andreas Bentzen Winje. It contains human-produced and machine-generated text samples of scientific research abstracts. A reformatted version for text-classification is available in the dataset...
true
# Dataset Card for "cardiffnlp/tweet_topic_single" ## Dataset Description - **Paper:** [https://arxiv.org/abs/2209.09824](https://arxiv.org/abs/2209.09824) - **Dataset:** Tweet Topic Dataset - **Domain:** Twitter - **Number of Class:** 6 ### Dataset Summary This is the official repository of TweetTopic (["Twitter ...
false
# Dataset Card for `tldr_news` ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - ...
true
# Datasheet for the dataset: multilingual-NLI-26lang-2mil7 ## Dataset Summary This dataset contains 2 730 000 NLI text pairs in 26 languages spoken by more than 4 billion people. The dataset can be used to train models for multilingual NLI (Natural Language Inference) or zero-shot classification. The dataset is base...
false
# Dataset Card for DocLayNet base ## About this card (01/27/2023) ### Property and license All information from this page but the content of this paragraph "About this card (01/27/2023)" has been copied/pasted from [Dataset Card for DocLayNet](https://huggingface.co/datasets/ds4sd/DocLayNet). DocLayNet is a datase...
true
# Dataset accompanying the "Probing neural language models for understanding of words of estimative probability" article This dataset tests the capabilities of language models to correctly capture the meaning of words denoting probabilities (WEP, also called verbal probabilities), e.g. words like "probably", "maybe",...
false
# Dataset Card for [GPT4All Prompt Generations] ## Dataset Description Dataset used to train [GPT4All](https://huggingface.co/nomic-ai/gpt4all-lora) - **Homepage:** - **Repository:** [gpt4all](https://github.com/nomic-ai/gpt4all) - **Paper:** [Technical Report](https://s3.amazonaws.com/static.nomic.ai/gpt4all/2023_...
false
# Dataset Card for "fake-w2-us-tax-form-dataset" This is a dataset of synthetically generated US Tax Return W2 Forms, with generated fake data such as names, ids, dates and addresses. Only real city, state and zipcodes have been used. This dataset is created from the existing public [Fake W-2 (US Tax Form) Dataset](h...
true
# Dataset Card for the_pile_openwebtext2 ## Table of Contents - [Dataset Card for the_pile_openwebtext2](#dataset-card-for-the_pile_openwebtext2) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboard...
true
# The Touché23-ValueEval Dataset ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Dataset Usage](#dataset-usage) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languag...
false
# Dataset Card for "ml-arxiv-papers" This is a dataset containing ML ArXiv papers. The dataset is a version of the original one from [CShorten](https://huggingface.co/datasets/CShorten/ML-ArXiv-Papers), which is a part of the ArXiv papers dataset from [Kaggle](https://www.kaggle.com/datasets/Cornell-University/arxiv)....
false
# Dataset Card for medical 中文医疗数据集 - LLM Supervised Finetuning repository: https://github.com/shibing624/textgen - MeidcalGPT repository: https://github.com/shibing624/MedicalGPT ## Dataset Description medical is a Chinese Medical dataset. 医疗数据集,可用于医疗领域大模型训练。 ``` tree medical |-- finetune # 监督微调数据集,可用于SFT和RLHF ...
false
# **CAMEL: Communicative Agents for “Mind” Exploration of Large Scale Language Model Society** - **Github:** https://github.com/lightaime/camel - **Website:** https://www.camel-ai.org/ - **Arxiv Paper:** https://arxiv.org/abs/2303.17760 ## Dataset Summary Physics dataset is composed of 20K problem-solution pairs obt...
true
# Dataset Card for Dataset Name ## Dataset Description - **Homepage:https://github.com/kaistAI/CoT-Collection** - **Repository:https://github.com/kaistAI/CoT-Collection** - **Paper:https://arxiv.org/abs/2305.14045** - **Point of Contact:sejune@lklab.io** ### Dataset Summary This dataset card aims to be a base ...
false
# scientific_lay_summarisation - PLOS - normalized This dataset is a modified version of [tomasg25/scientific_lay_summarization](https://huggingface.co/datasets/tomasg25/scientific_lay_summarisation) and contains scientific lay summaries that have been preprocessed [with this code](https://gist.github.com/pszemraj/bd...
false
# Dataset Card for Swiss Citation Extraction ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](...
false
# Dataset Card for "dev_pretrain" [Tigerbot模型](https://github.com/TigerResearch/TigerBot#%E6%A8%A1%E5%9E%8B%E4%B8%8B%E8%BD%BD)develop pretrain数据。 在[train_clm.py](https://github.com/TigerResearch/TigerBot/blob/main/train/train_clm.py)中被使用。 ## Usage ```python import datasets ds_sft = datasets.load_dataset('TigerResea...
false
# Dataset Card for "tner/wikineural" ## Dataset Description - **Repository:** [T-NER](https://github.com/asahi417/tner) - **Paper:** [https://aclanthology.org/2021.findings-emnlp.215/](https://aclanthology.org/2021.findings-emnlp.215/) - **Dataset:** WikiNeural - **Domain:** Wikipedia - **Number of Entity:** 16 ##...
false
# CORD 19 ## Dataset Description - **Homepage:** https://www.kaggle.com/datasets/allen-institute-for-ai/CORD-19-research-challenge ### Dataset Summary In response to the COVID-19 pandemic, the White House and a coalition of leading research groups have prepared the COVID-19 Open Research Dataset (CORD-19). CORD-19...
true
# Dataset Card for TSATC: Twitter Sentiment Analysis Training Corpus ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) ...
false
# Dataset Card for LogiQA ## Dataset Description - **Homepage:** - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary LogiQA is constructed from the logical comprehension problems from publically available questions of the National Civil Servants Examination of China, w...
false
# Dataset Card for "bert_dataset_202203" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
true
# Dataset Card for Moral Stories ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fie...
false
# Dataset Card for "cmu_wiki_qa" A filtered / cleaned version of the http://www.cs.cmu.edu/~ark/QA-data/ Q&A dataset, which provides manually-generated factoid questions from Wikipedia articles. **Acknowledgments** These data were collected by Noah Smith, Michael Heilman, Rebecca Hwa, Shay Cohen, Kevin Gimpel, and ma...
false
# ImageRewardDB ## Dataset Description - **Homepage: https://huggingface.co/datasets/wuyuchen/ImageRewardDB** - **Repository: https://github.com/THUDM/ImageReward** - **Paper: https://arxiv.org/abs/2304.05977** ### Dataset Summary ImageRewardDB is a comprehensive text-to-image comparison dataset, focusing on te...
true
# Leyzer: A Dataset for Multilingual Virtual Assistants Leyzer is a multilingual text corpus designed to study multilingual and cross-lingual natural language understanding (NLU) models and the strategies of localization of virtual assistants. It consists of 20 domains across three languages: English, Spanish and Poli...
false
false
# 20,000+ chinese sentences with translations and pinyin - Source: https://mnemosyne-proj.org/cards/20000-chinese-sentences-translations-and-pinyin - Contributed by: Brian Vaughan http://brianvaughan.net/ # Dataset Structure Each sample consists of: 1. English sentence 2. HSK level 3. Chinese translation 4. Pinyin 5...
false
# Dataset Card for [Dataset Name] ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-st...
false
# Dataset Card for xP3 ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) ...
true
# Dataset Card for The Pile GitHub ## Table of Contents - [Dataset Card for Smart Contracts](#dataset-card-for-the-pile-github) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-task...
false
# GPT Role-play Realm Dataset: The AI-generated character compendium This is a dataset of GPT-generated characters made to increase the ability of open-source language models to role-play. <img src="https://cdn.midjourney.com/9c17407c-9ce8-435f-99ab-e349b900a6ed/0_3.png" > * 219 characters in the Russian part, and 2...
false
# Dataset Card for BSARD ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) ...
false
# Dataset Card for xP3 ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) ...
false
# Dataset Card Creation Guide ## Table of Contents - [Dataset Card Creation Guide](#dataset-card-creation-guide) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboa...
true
# Dataset Card for [Dataset Name] ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#da...
false
# Dataset Card for PersiNLU (Textual Entailment) ## Table of Contents - [Dataset Card for PersiNLU (Textual Entailment)](#dataset-card-for-persi_nlu_entailment) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks ...
false
# Dataset Card for "Medical_reports_Splits" Orignal Source [openi.nlm.nih.gov](https://openi.nlm.nih.gov/) Kaggle Source [Chest X-rays (Indiana University)](https://www.kaggle.com/datasets/raddar/chest-xrays-indiana-university) [For more information](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#...
false
# RED<sup>FM</sup>: a Filtered and Multilingual Relation Extraction Dataset This is the automatically-filtered dataset from the 2023 ACL paper [RED^{FM}: a Filtered and Multilingual Relation Extraction Dataset](https://arxiv.org/abs/2306.09802). If you use the model, please reference this work in your paper: @inp...
false
Preprocessed version of Super-Natural-Instructions from https://github.com/allenai/natural-instructions/tree/master/splits. The same inputs may appear with different outputs, thus to avoid duplicate inputs, you can deduplicate by the `id` or the `inputs` field. This is modified from https://huggingface.co/datasets/Mu...
true
# Dataset Card for AfriSenti Dataset <p align="center"> <img src="https://raw.githubusercontent.com/afrisenti-semeval/afrisent-semeval-2023/main/images/afrisenti-twitter.png", width="700" height="500"> -------------------------------------------------------------------------------- ## Dataset Description - **Home...
false
# Dataset Card for "dolly_wizard_codepy" This is a mix of datasets from: - Abzu/dolly_hhrlhf - Abzu/Wizard - Abzu/CodeAlpacaPython Which are versions of: - mosaicml/dolly_hhrlhf - ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered - HuggingFaceH4/CodeAlpaca_20K [More Information needed](https://github.com/hug...
false
# Dataset Card for "EuroSAT2" ## Dataset Description - **Paper** [Eurosat: A novel dataset and deep learning benchmark for land use and land cover classification](https://ieeexplore.ieee.org/iel7/4609443/8789745/08736785.pdf) - **Paper** [Introducing EuroSAT: A Novel Dataset and Deep Learning Benchmark for Land Use a...
false
This dataset was taken from the creators [GitHub repository](https://github.com/salute-developers/golos/tree/master/dusha) and converted for my own studying needs. # Dusha dataset Dusha is a bi-modal corpus suitable for speech emotion recognition (SER) tasks. The dataset consists of about 300 000 audio recordings wit...
false
<p align="center"><img src="https://huggingface.co/datasets/cfilt/HiNER-collapsed/raw/main/cfilt-dark-vec.png" alt="Computation for Indian Language Technology Logo" width="150" height="150"/></p> # Dataset Card for HiNER-original [![Twitter Follow](https://img.shields.io/twitter/follow/cfiltnlp?color=1DA1F2&logo=twi...
false
# Model-Written Evaluation Datasets This repository includes datasets written by language models, used in the paper "Discovering Language Model Behaviors with Model-Written Evaluations." The evaluations in this dataset were designed for dialogue agents, such as models fine-tuned to respond to user utterances or pret...
false
# Dataset Card for "ko_alpaca_data" ## Dataset Description - **Repository:** [Beomi/KoAlpaca](https://github.com/Beomi/KoAlpaca) - **Huggingface:** [beomi/KoAlpaca](https://huggingface.co/beomi/KoAlpaca) - **Size of downloaded dataset files:** 8.10 MB - **Size of the generated dataset:** 13.15 MB ### Dataset Summar...
false
# Dataset Card for "databricks-dolly-15k-curated-multilingual" A curated and multilingual version of the Databricks Dolly instructions dataset. It includes a programmatically and manually corrected version of the original `en` dataset. See below. **STATUS**: Currently, the original Dolly v2 English version has been ...
true
# Turkish Texts for Toxic Language Detection ## Dataset Description ### Dataset Summary This text dataset is a collection of Turkish texts that have been merged from various existing offensive language datasets found online. The dataset contains a total of 77,800 instances, each labeled as either offensive or not ...
false
# Dataset Card for CIFAR-100-Enriched (Enhanced by Renumics) ## Dataset Description - **Homepage:** [Renumics Homepage](https://renumics.com/?hf-dataset-card=cifar100-enriched) - **GitHub** [Spotlight](https://github.com/Renumics/spotlight) - **Dataset Homepage** [CS Toronto Homepage](https://www.cs.toronto.edu/~kriz...
false
LIMA dataset in Vicuna ShareGPT format. License under LIMA's License. Original Repo: https://huggingface.co/datasets/GAIR/lima