datasetId
stringlengths
2
81
card
stringlengths
20
977k
collabora/whisperspeech
--- license: mit task_categories: - text-to-speech language: - en pretty_name: WhisperSpeech --- # The WhisperSpeech Dataset This dataset contains data to train SPEAR TTS-like text-to-speech models that utilized semantic tokens derived from the OpenAI Whisper speech recognition model. We currently provide semantic and acoustic tokens for the LibriLight and LibriTTS datasets (English only). Acoustic tokens: - 24kHz EnCodec 6kbps (8 quantizers) Semantic tokens: - Whisper tiny VQ bottleneck trained on a subset of LibriLight Available LibriLight subsets: - `small`/`medium`/`large` (following the original dataset division but with `large` excluding the speaker `6454`) - a separate ≈1300hr single-speaker subset based on the `6454` speaker from the `large` subset for training single-speaker TTS models We plan to add more acoustic tokens from other codecs in the future.
pankajmathur/alpaca_orca
--- license: cc-by-nc-sa-4.0 task_categories: - text-generation language: - en size_categories: - 10K<n<100K --- Explain tuned Alpaca dataset ~52K created using approaches from Orca Research Paper. We leverage all of the 15 system instructions provided in Orca Research Paper. to generate custom datasets, in contrast to vanilla instruction tuning approaches used by original datasets. This helps student models like [orca_mini_13b](https://huggingface.co/psmathur/orca_mini_13b) to learn thought process from teacher model, which is ChatGPT (gpt-3.5-turbo-0301 version). Please see how the **System** prompt is added before each **instruction**.
FreedomIntelligence/alpaca-gpt4-japanese
--- license: apache-2.0 --- The dataset is used in the research related to [MultilingualSIFT](https://github.com/FreedomIntelligence/MultilingualSIFT).
breadlicker45/bread-midi-dataset
--- tags: - 'music ' --- this midi dataset has 851,313 midi files, making it the biggest midi dataset on the web. anyone can use for any use case
FredZhang7/malicious-website-features-2.4M
--- license: apache-2.0 task_categories: - text-classification - feature-extraction - tabular-classification language: - 'no' - af - en - et - sw - sv - sq - de - ca - hu - da - tl - so - fi - fr - cs - hr - cy - es - sl - tr - pl - pt - nl - id - sk - lt - lv - vi - it - ro - ru - mk - bg - th - ja - ko - multilingual size_categories: - 1M<n<10M --- **Important Notice:** - A subset of the URL dataset is from Kaggle, and the Kaggle datasets contained 10%-15% mislabelled data. See [this dicussion I opened](https://www.kaggle.com/datasets/sid321axn/malicious-urls-dataset/discussion/431505) for some false positives. I have contacted Kaggle regarding their erroneous "Usability" score calculation for these unreliable datasets. - The feature extraction methods shown here are not robust at all in 2023, and there're even silly mistakes in 3 functions: `not_indexed_by_google`, `domain_registration_length`, and `age_of_domain`. <br> The *features* dataset is original, and my feature extraction method is covered in [feature_extraction.py](./feature_extraction.py). To extract features from a website, simply passed the URL and label to `collect_data()`. The features are saved to `phishing_detection_dataset.csv` locally by default. In the *features* dataset, there're 911,180 websites online at the time of data collection. The plots below show the regression line and correlation coefficients of 22+ features extracted and whether the URL is malicious. If we could plot the lifespan of URLs, we could see that the oldest website has been online since Nov 7th, 2008, while the most recent phishing websites appeared as late as July 10th, 2023. ## Malicious URL Categories - Defacement - Malware - Phishing ## Data Analysis Here are two images showing the correlation coefficient and correlation of determination between predictor values and the target value `is_malicious`. ![Correlation Coefficient](https://i.imgur.com/LLD3pmt.png) ![Correlation of Determination](https://i.imgur.com/GJM3Cl6.png) Let's exmain the correlations one by one and cross out any unreasonable or insignificant correlations. | Variable | Justification for Crossing Out | |-----------------------------|------------------------------------- | | ~~redirects~~ | contracdicts previous research (as redirects increase, is_malicious tends to decrease by a little) | | ~~not_indexed_by_google~~ | 0.00 correlation | | ~~email_submission~~ | contracdicts previous research | | request_url_percentage | | | issuer | | | certificate_age | | | ~~url_anchor_percentage~~ | contracdicts previous research | | ~~meta_percentage~~ | 0.00 correlation | | script_percentage | | | link_percentage | | | ~~mouseover_changes~~ | contracdicts previous research & 0.00 correlation | | ~~right_clicked_disabled~~ | contracdicts previous research & 0.00 correlation | | ~~popup_window_has_text_field~~ | contracdicts previous research | | ~~use_iframe~~ | contracdicts previous research | | ~~has_suspicious_ports~~ | contracdicts previous research | | ~~external_favicons~~ | contracdicts previous research | | TTL (Time to Live) | | | ip_address_count | | | ~~TXT_record~~ | all websites had a TXT record | | ~~check_sfh~~ | contracdicts previous research | | count_domain_occurrences | | | domain_registration_length | | | abnormal_url | | | age_of_domain | | | page_rank_decimal | | ## Pre-training Ideas For training, I split the classification task into two stages in anticipation of the limited availability of online phishing websites due to their short lifespan, as well as the possibility that research done on phishing is not up-to-date: 1. a small multilingual BERT model to output the confidence level of a URL being malicious to model #2, by finetuning on 2,436,727 legitimate and malicious URLs 2. (probably) LightGBM to analyze the confidence level, along with roughly 10 extracted features This way, I can make the most out of the limited phishing websites avaliable. ## Source of the URLs - https://moz.com/top500 - https://phishtank.org/phish_search.php?valid=y&active=y&Search=Search - https://www.kaggle.com/datasets/siddharthkumar25/malicious-and-benign-urls - https://www.kaggle.com/datasets/sid321axn/malicious-urls-dataset - https://github.com/ESDAUNG/PhishDataset - https://github.com/JPCERTCC/phishurl-list - https://github.com/Dogino/Discord-Phishing-URLs ## Reference - https://www.kaggle.com/datasets/akashkr/phishing-website-dataset - https://www.kaggle.com/datasets/shashwatwork/web-page-phishing-detection-dataset - https://www.kaggle.com/datasets/aman9d/phishing-data ## Side notes - Cloudflare offers an [API for phishing URL scanning](https://developers.cloudflare.com/api/operations/phishing-url-information-get-results-for-a-url-scan), with a generous global rate limit of 1200 requests every 5 minutes.
jensjorisdecorte/Synthetic-ESCO-skill-sentences
--- license: cc task_categories: - text-classification language: - en tags: - Skill Extraction - Synthetic Data pretty_name: Synthetic ESCO skill sentences size_categories: - 100K<n<1M --- # Synthetic job ads for all ESCO skills ## Dataset Description - **Homepage:** coming soon - **Repository:** coming soon - **Paper:** https://arxiv.org/abs/2307.10778 - **Point of Contact:** jensjoris@techwolf.ai ### Dataset Summary This dataset contains 10 synthetically generated job ad sentences for almost all (99.5%) skills in ESCO v1.1.0. ### Languages We use the English version of ESCO, and all generated sentences are in English. ## Dataset Structure The dataset consists of 138,260 `(sentence, skill)` pairs. ### Citation Information [More Information Needed]
Safurai/Code-Instruct-700k
--- dataset_info: features: - name: prompt dtype: large_string - name: main_topic dtype: large_string - name: subtopic dtype: large_string - name: adjective dtype: large_string - name: action_verb dtype: large_string - name: scenario dtype: large_string - name: target_audience dtype: large_string - name: programming_language dtype: large_string - name: common_sense_topic dtype: large_string - name: idx dtype: int64 - name: response dtype: large_string splits: - name: train num_bytes: 1657193365 num_examples: 700000 download_size: 705514514 dataset_size: 1657193365 --- # Dataset Card for "Code-Instruct-700k" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
SachinKaushik/LlamaV2InstructCode
--- dataset_info: features: - name: text dtype: string - name: input dtype: string - name: instruction dtype: string - name: output dtype: string - name: llamaV2Instruct dtype: string splits: - name: train num_bytes: 241331660 num_examples: 121959 download_size: 0 dataset_size: 241331660 task_categories: - text-generation - text2text-generation language: - en tags: - python - llamav2 - instruction - code --- # Dataset Card for "LlamaV2InstructCode" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
DUOMO-Lab/TransGPT-sft
--- license: apache-2.0 ---
Azure99/blossom-chat-v1
--- license: apache-2.0 task_categories: - text-generation - text2text-generation language: - zh - en size_categories: - 10K<n<100K --- # BLOSSOM CHAT V1 ### 介绍 Blossom Chat V1是基于ShareGPT 90K衍生而来的中英双语对话数据集,适用于多轮对话微调。 本数据集抽取了ShareGPT的多轮对话指令,仅将指令进行翻译,随后使用多轮指令迭代调用gpt-3.5-turbo-0613。 相比原始的ShareGPT数据,主要解决了中文对话数据量较少,以及由ChatGPT生成长度限制而导致的输出截断问题。 本次发布了全量数据的20%,包含30K记录。 ### 语言 以中文和英文为主,中英文数据按照约5:1的比例混合。 ### 数据集结构 每条数据代表一个完整的多轮对话,包含id和conversations两个字段。 - id:字符串,代表原始ShareGPT的对话id,可以通过链接https://sharegpt.com/c/id来访问原始对话。 - conversations:对象数组,每个对象包含role、content两个字段,role的取值为user或assistant,分别代表用户输入和助手输出,content则为对应的内容。 ### 数据集限制 由于仅抽取了原始多轮对话的输入,对于一些涉及随机性的对话,例如:猜随机数,可能会出现多轮对话不连贯的情况。 此外,本数据集的所有响应均由gpt-3.5-turbo-0613生成,并未经过严格的数据校验,可能包含不准确甚至严重错误的回答。
HeshamHaroon/Egyptian_English_parallel
--- license: apache-2.0 ---
owkin/medical_knowledge_from_extracts
--- license: apache-2.0 task_categories: - summarization language: - en --- This dataset is used to train LLMs for medical knowledge extraction tasks
harshitv804/Indian_Penal_Code
--- language: - en viewer: false pretty_name: Indian Penal Code Book license: openrail task_categories: - question-answering - text2text-generation - sentence-similarity tags: - legal - law --- # Indian Penal Code Dataset <img src="https://iilsindia.com/blogs/wp-content/uploads/2016/07/indian-penal-code-1860-890x395.jpg" style="width:600px;"/> ## Dataset Description: The Indian Penal Code (IPC) Book PDF presents a rich and comprehensive dataset that holds immense potential for advancing Natural Language Processing (NLP) tasks and Language Model applications. This dataset encapsulates the entire spectrum of India's criminal law, offering a diverse range of legal principles, provisions, and case laws. With its intricate language and multifaceted legal content, the IPC dataset provides a challenging yet rewarding opportunity for NLP research and development. From text summarization and legal language understanding to sentiment analysis within the context of legal proceedings, this IPC dataset opens avenues for training and fine-tuning Language Models to grasp the nuances of complex legal texts. Leveraging this dataset, researchers and practitioners in the field of NLP can unravel the intricacies of legal discourse and contribute to the advancement of AI-driven legal analysis, interpretation, and decision support systems. ## Languages: - English ## Considerations for Using this Data: - Question Answering - Conversational AI - Text2Text Generation - Sentence Similarity - Text Generation - RAG ## Dataset Download: <a href="https://huggingface.co/datasets/harshitv804/Indian_Penal_Code/resolve/main/Indian%20Penal%20Code%20Book.pdf"><img src="https://static.vecteezy.com/system/resources/previews/009/384/880/non_2x/click-here-button-clipart-design-illustration-free-png.png" width="150" height="auto"></a>
52AI/TinyStoriesZh
--- license: mit --- LM朝着越来越大的方向卷,而在小LM的方向,有研究者在探索小LM方向的边界能力,比如想知道多小的语言模型仍然能流畅的说故事? [TinyStories](https://huggingface.co/datasets/roneneldan/TinyStories) 是在其做该方向时使用的一份关于小故事的场景数据。故事是由研究者使用GPT3.5, GPT4生成的,并且将故事难度限制在3~4岁小朋友能理解。 这份中文数据通过[翻译器](https://pypi.org/project/deep-translator/)将英文故事数据翻译而成。如下例子。 > Lily and Ben are friends. They like to play in the park. One day, they see a big tree with a swing. Lily wants to try the swing. She runs to the tree and climbs on the swing.\n"Push me, Ben!" she says. Ben pushes her gently. Lily feels happy. She swings higher and higher. She laughs and shouts.\nBen watches Lily. He thinks she is cute. He wants to swing too. He waits for Lily to stop. But Lily does not stop. She swings faster and faster. She is having too much fun.\n"Can I swing too, Lily?" Ben asks. Lily does not hear him. She is too busy swinging. Ben feels sad. He walks away.\nLily swings so high that she loses her grip. She falls off the swing. She lands on the ground. She hurts her foot. She cries.\n"Ow, ow, ow!" she says. She looks for Ben. She wants him to help her. But Ben is not there. He is gone.\nLily feels sorry. She wishes she had shared the swing with Ben. She wishes he was there to hug her. She limps to the tree. She sees something hanging from a branch. It is Ben\'s hat. He left it for her.\nLily smiles. She thinks Ben is nice. She puts on his hat. She hopes he will come back. She wants to say sorry. She wants to be friends again. > 莉莉和本是朋友。他们喜欢在公园里玩。有一天,他们看到一棵有秋千的大树。莉莉想尝试秋千。她跑到树旁,爬上秋千。\n“推我吧,本!”她说。本轻轻地推了她一下。莉莉感觉很幸福。她荡得越来越高。她又笑又叫。\n本看着莉莉。他觉得她很可爱。他也想摇摆。他等着莉莉停下来。但莉莉并没有停下来。她摆动得越来越快。她玩得太开心了。\n“我也可以荡秋千吗,莉莉?”本问。莉莉没有听见他的话。她正忙着荡秋千。本感到难过。他走开了。\n莉莉荡得太高,以至于她失去了抓力。她从秋千上摔下来。她降落在地上。她的脚受伤了。她哭了。\n“呜呜呜!”她说。她寻找本。她想要他帮助她。但本不在那儿。他已经去了。\n莉莉感到抱歉。她希望自己能和本一起荡秋千。她希望他能在那里拥抱她。她一瘸一拐地走向树。她看到树枝上挂着什么东西。这是本的帽子。他留给她了。\n莉莉微笑着。她认为本很好。她戴上他的帽子。她希望他能回来。她想说对不起。她想再次成为朋友。
SUFE-AIFLM-Lab/FinEval
--- license: cc-by-nc-sa-4.0 task_categories: - text-classification - multiple-choice - question-answering language: - zh pretty_name: FinEval size_categories: - 1K<n<10K viewer: false --- <p><h1> The FinEval Dataset </h1></p> ![FinEval Logo](https://huggingface.co/datasets/SUFE-AIFLM-Lab/FinEval/resolve/main/FinEvalLogo.jpg "FinEval Logo") <a name="dataset-announcement"></a> FinEval is a collection of high-quality multiple-choice questions covering various domains such as finance, economics, accounting, and certifications. It consists of 4,661 questions spanning across 34 distinct academic subjects. To ensure a comprehensive assessment of model performance, FinEval employs various methods including zero-shot, few-shot, answer-only, and chain-of-thought prompts. Evaluating state-of-the-art large language models in both Chinese and English on FinEval reveals that only GPT-4 achieves an accuracy of 60% across different prompt settings, highlighting substantial growth potential of large language models in financial domain knowledge. Our work provides a more comprehensive benchmark for evaluating financial knowledge, utilizing simulated exam data and encompassing a wide range of large language model assessments. Each subject consists of three splits: dev, val, and test. The dev set per subject consists of five exemplars with explanations for few-shot evaluation. The val set is intended to be used for hyperparameter tuning. And the test set is for model evaluation. Labels on the test split are not released, users are required to submit their results to automatically obtain test accuracy. # Language The language of the data is Chinese. # Performance Leaderboard We divide the evaluation into Answer Only and Chain of Thought. For examples of prompts for both methods, please refer to zero-shot for Answer Only, few-shot for Answer Only, and Chain of Thought. Below is the average accuracy(%) on the test split. We report the average accuracy over the subjects within each category. "Average" column indicates the average accuracy over all the subjects. Notably, we only report the results from each model under the best setting, which is determined by the highest average accuracy achieved among four settings (i.e., zero- and few-shot learning with and without CoT): | Model | Size | Finance | Economy | Accounting | Certificate | Average | |------------------------|---------|:-------:|:-------:|:----------:|:-----------:|:-------:| | GPT-4 | unknown | 71.0 | 74.5 | 59.3 | 70.4 | 68.6 | | ChatGPT | 175B | 59.3 | 61.6 | 45.2 | 55.1 | 55.0 | | Qwen-7B | 7B | 54.5 | 54.4 | 50.3 | 55.8 | 53.8 | | Qwen-Chat-7B | 7B | 51.5 | 52.1 | 44.5 | 53.6 | 50.5 | | Baichuan-13B-Base | 13B | 52.6 | 50.2 | 43.4 | 53.5 | 50.1 | | Baichuan-13B-Chat | 13B | 51.6 | 51.1 | 41.7 | 52.8 | 49.4 | | ChatGLM2-6B | 6B | 46.5 | 46.4 | 44.5 | 51.5 | 47.4 | | InternLM-7B | 7B | 49.0 | 49.2 | 40.5 | 49.4 | 47.1 | | InternLM-Chat-7B | 7B | 48.4 | 49.1 | 40.8 | 49.5 | 47.0 | | LLaMA-2-Chat-70B | 70B | 47.1 | 46.7 | 41.5 | 45.7 | 45.2 | | Falcon-40B | 40B | 45.4 | 43.2 | 35.8 | 44.8 | 42.4 | | Baichuan-7B | 7B | 44.9 | 41.5 | 34.9 | 45.6 | 42.0 | | LLaMA-2-Chat-13B | 13B | 41.6 | 38.4 | 34.1 | 42.1 | 39.3 | | Ziya-LLaMA-13B-v1 | 13B | 43.3 | 36.9 | 34.3 | 41.2 | 39.3 | | Bloomz-7b1-mt | 7B | 41.4 | 42.1 | 32.5 | 39.7 | 38.8 | | LLaMA-2-13B | 13B | 39.5 | 38.6 | 31.6 | 39.6 | 37.4 | | ChatGLM-6B | 6B | 38.8 | 36.2 | 33.8 | 39.1 | 37.2 | | Chinese-Llama-2-7B | 7B | 37.8 | 37.8 | 31.4 | 36.7 | 35.9 | | Chinese-Alpaca-Plus-7B | 7B | 30.5 | 33.4 | 32.7 | 38.5 | 34.0 | | moss-moon-003-sft | 16B | 35.6 | 34.3 | 28.7 | 35.6 | 33.7 | | LLaMA-2-Chat-7B | 7B | 35.6 | 31.8 | 31.9 | 34.0 | 33.5 | | LLaMA-2-7B | 7B | 34.9 | 36.4 | 31.4 | 31.6 | 33.4 | | AquilaChat-7B | 7B | 34.2 | 31.3 | 29.8 | 36.2 | 33.1 | | moss-moon-003-base | 16B | 32.2 | 33.1 | 29.2 | 30.7 | 31.2 | | Aquila-7B | 7B | 27.1 | 31.6 | 32.4 | 33.6 | 31.2 | | LLaMA-13B | 13B | 33.1 | 29.7 | 27.2 | 33.6 | 31.1 | | Falcon-7B | 7B | 28.5 | 28.2 | 27.5 | 27.4 | 27.9 | # Load the data ```python from datasets import load_dataset dataset=load_dataset(r"SUFE-AIFLM-Lab/FinEval",name="finance") ``` Please cite our paper if you use our dataset. ``` @misc{2308.09975, Author = {Liwen Zhang and Weige Cai and Zhaowei Liu and Zhi Yang and Wei Dai and Yujie Liao and Qianru Qin and Yifei Li and Xingyu Liu and Zhiqiang Liu and Zhoufan Zhu and Anbo Wu and Xin Guo and Yun Chen}, Title = {FinEval: A Chinese Financial Domain Knowledge Evaluation Benchmark for Large Language Models}, Year = {2023}, Eprint = {arXiv:2308.09975}, } ```
google/trueteacher
--- license: cc-by-nc-4.0 language: - en tags: - natural-language-inference - news-articles-summarization --- # **TrueTeacher** ## Dataset Summary This is a large-scale synthetic dataset for training **Factual Consistency Evaluation** models, introduced in the [TrueTeacher paper (Gekhman et al, 2023)](https://aclanthology.org/2023.emnlp-main.127.pdf). ## Dataset Details The dataset contains model-generated summaries of articles from the train split of the **CNN/DailyMail** dataset [(Hermann et al., 2015)](https://proceedings.neurips.cc/paper_files/paper/2015/file/afdec7005cc9f14302cd0474fd0f3c96-Paper.pdf) which are annotated for factual consistency using **FLAN-PaLM 540B** [(Chung et al.,2022)](https://arxiv.org/pdf/2210.11416.pdf). Summaries were generated using summarization models with different capacities, which were created by fine-tuning **T5** [(Raffel et al., 2020)](https://jmlr.org/papers/volume21/20-074/20-074.pdf) on the **XSum** dataset [(Narayan et al., 2018)](https://aclanthology.org/D18-1206.pdf). We used the following 5 capacities: T5-11B, T5-3B, T5-large, T5-base and T5-small. ## Data format The data contains json lines with the following keys: - `"summarization_model"` - The summarization model used to generate the summary. - `"cnndm_id"` - The original id from the CNN/DailyMail dataset, this need to be used in order to retrieve the corresponding article from CNN/DailyMail (which was used as the grounding document). - `"summary"` - The model-generated summary. - `"label"` - A binary label ('1' - Factualy Consistent, '0' - Factualy Inconsistent). Here is an example of a single data item: ```json { "summarization_model": "T5-11B", "cnndm_id": "f72048a23154de8699c307e2f41157abbfcae261", "summary": "Children's brains are being damaged by prolonged internet access, a former children's television presenter has warned." "label": "1", } ``` ## Loading the dataset To use the dataset, you need to fetch the relevant documents from the CNN/DailyMail dataset. The follwoing code can be used for that purpose: ```python from datasets import load_dataset from tqdm import tqdm trueteacher_data = load_dataset("google/trueteacher", split='train') cnn_dailymail_data = load_dataset("cnn_dailymail", version="3.0.0", split='train') cnn_dailymail_articles_by_id = {example['id']: example['article'] for example in cnn_dailymail_data} trueteacher_data_with_documents = [] for example in tqdm(trueteacher_data):   example['document'] = cnn_dailymail_articles_by_id[example['cnndm_id']]   trueteacher_data_with_documents.append(example) ``` ## Intended Use This dataset is intended for a research use (**non-commercial**) in English. The recommended use case is training factual consistency evaluation models for summarization. ## Out-of-scope use Any use cases which violate the **cc-by-nc-4.0** license. Usage in languages other than English. ## Citation If you use this dataset for a research publication, please cite the TrueTeacher paper (using the bibtex entry below), as well as the CNN/DailyMail, XSum, T5 and FLAN papers mentioned above. ``` @misc{gekhman2023trueteacher, title={TrueTeacher: Learning Factual Consistency Evaluation with Large Language Models}, author={Zorik Gekhman and Jonathan Herzig and Roee Aharoni and Chen Elkind and Idan Szpektor}, year={2023}, eprint={2305.11171}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
baoanhtran/guanaco-llama2-200
--- pretty_name: CulturaX annotations_creators: - no-annotation language_creators: - found language: - af - als - am - an - ar - arz - as - ast - av - az - azb - ba - bar - bcl - be - bg - bh - bn - bo - bpy - br - bs - bxr - ca - cbk - ce - ceb - ckb - cs - cv - cy - da - de - dsb - dv - el - eml - en - eo - es - et - eu - fa - fi - fr - frr - fy - ga - gd - gl - gn - gom - gu - he - hi - hr - hsb - ht - hu - hy - ia - id - ie - ilo - io - is - it - ja - jbo - jv - ka - kk - km - kn - ko - krc - ku - kv - kw - ky - la - lb - lez - li - lmo - lo - lrc - lt - lv - mai - mg - mhr - min - mk - ml - mn - mr - mrj - ms - mt - mwl - my - myv - mzn - nah - nap - nds - ne - new - nl - nn - 'no' - oc - or - os - pa - pam - pl - pms - pnb - ps - pt - qu - rm - ro - ru - rue - sa - sah - scn - sd - sh - si - sk - sl - so - sq - sr - su - sv - sw - ta - te - tg - th - tk - tl - tr - tt - tyv - ug - uk - ur - uz - vec - vi - vls - vo - wa - war - wuu - xal - xmf - yi - yo - yue - zh multilinguality: - multilingual size_categories: - n<1K - 1K<n<10K - 10K<n<100K - 100K<n<1M - 1M<n<10M - 10M<n<100M - 100M<n<1B - 1B<n<10B source_datasets: - original task_categories: - text-generation - fill-mask task_ids: - language-modeling - masked-language-modeling extra_gated_prompt: "By completing the form below, you acknowledge that the provided data is offered as is. Although we anticipate no problems, you accept full responsibility for any repercussions resulting from the use of this data. Furthermore, you agree that the data must not be utilized for malicious or harmful purposes towards humanity." extra_gated_fields: Name: text Email: text Affiliation: text Country: text Usecase: text I have explicitly check with my jurisdiction and I confirm that downloading CulturaX is legal in the country/region where I am located right now, and for the use case that I have described above: checkbox You agree to not attempt to determine the identity of individuals in this dataset: checkbox --- <div align="center"> <h1> CulturaX </h1> <h3> Cleaned, Enormous, and Public: The Multilingual Fuel to Democratize Large Language Models for 167 Languages </h3> </div> ## Dataset Description - **Repository:** [https://github.com/nlp-uoregon/CulturaX](https://github.com/nlp-uoregon/CulturaX) - **Papers:** [CulturaX: A Cleaned, Enormous, and Multilingual Dataset for Large Language Models in 167 Languages](https://arxiv.org/abs/2309.09400) ## Dataset Summary We present CulturaX, a substantial multilingual dataset with 6.3 trillion tokens in 167 languages, tailored for large language model (LLM) development. Our dataset undergoes meticulous cleaning and deduplication through a rigorous pipeline of multiple stages to accomplish the best quality for model training, including language identification, URL-based filtering, metric-based cleaning, document refinement, and data deduplication. We employ MinHash at document level to achieve fuzzy deduplication for the datasets in different languages. Our data cleaning framework includes diverse criteria and threshold selections, guided by extensive data samples, ensuring comprehensive noise filtering in various aspects. CulturaX is fully released to the public in HuggingFace to facilitate research and advancements in multilingual LLMs. Our dataset combines the most recent iteration of mC4 (version 3.1.0) [1] with all accessible OSCAR corpora up to the present year, including 20.19, 21.09, 22.01, and 23.01 [2]. After deep cleaning and deduplication, CulturaX involves 16TB data in the parquet format (expanding to 27TB when unpacked). More than a half of our dataset is dedicated to non-English languages to significantly boost the data size and enhance the feasibility of training models in multilingual scenarios. To obtain perplexity scores for data cleaning, we train a SentencePiece tokenizer and 5-gram Kneser-Ney language models as provided in the KenLM library [3] using the 20230501 dumps of Wikipedia. Our KenLM models are also released in HuggingFace: https://huggingface.co/uonlp/kenlm. Details for the dataset can be found in our technical paper: [https://arxiv.org/abs/2309.09400](https://arxiv.org/abs/2309.09400) You can download the dataset using Hugging Face datasets: *You may need to follow these instructions to setup authentication before downloading the dataset: [https://huggingface.co/docs/huggingface_hub/quick-start#login](https://huggingface.co/docs/huggingface_hub/quick-start#login)* ```python from datasets import load_dataset ds = load_dataset("uonlp/CulturaX", language="en", use_auth_token=True) ``` ### Languages The supported languages and statistics for our dataset can be found below: *(Note that the language code `als` and `eml` refer to `gsw` and `x-eml` in the OSCAR-2301 dataset.)* | | Code | Language | # Documents | # Tokens | # Tokens (%) | |----:|:-------|:-------------------------|:----------------|:--------------------|:------| | 0 | en | English | 3,241,065,682 | 2,846,970,578,793 | 45.13 | | 1 | ru | Russian | 799,310,908 | 737,201,800,363 | 11.69 | | 2 | es | Spanish | 450,937,645 | 373,845,662,394 | 5.93 | | 3 | de | German | 420,017,484 | 357,030,348,021 | 5.66 | | 4 | fr | French | 363,754,348 | 319,332,674,695 | 5.06 | | 5 | zh | Chinese | 218,624,604 | 227,055,380,882 | 3.60 | | 6 | it | Italian | 211,309,922 | 165,446,410,843 | 2.62 | | 7 | pt | Portuguese | 190,289,658 | 136,941,763,923 | 2.17 | | 8 | pl | Polish | 142,167,217 | 117,269,087,143 | 1.86 | | 9 | ja | Japanese | 111,188,475 | 107,873,841,351 | 1.71 | | 10 | vi | Vietnamese | 102,411,180 | 98,453,464,077 | 1.56 | | 11 | nl | Dutch | 117,392,666 | 80,032,209,900 | 1.27 | | 12 | ar | Arabic | 74,027,952 | 69,354,335,076 | 1.10 | | 13 | tr | Turkish | 94,207,460 | 64,292,787,164 | 1.02 | | 14 | cs | Czech | 65,350,564 | 56,910,486,745 | 0.90 | | 15 | fa | Persian | 59,531,144 | 45,947,657,495 | 0.73 | | 16 | hu | Hungarian | 44,132,152 | 43,417,981,714 | 0.69 | | 17 | el | Greek | 51,430,226 | 43,147,590,757 | 0.68 | | 18 | ro | Romanian | 40,325,424 | 39,647,954,768 | 0.63 | | 19 | sv | Swedish | 49,709,189 | 38,486,181,494 | 0.61 | | 20 | uk | Ukrainian | 44,740,545 | 38,226,128,686 | 0.61 | | 21 | fi | Finnish | 30,467,667 | 28,925,009,180 | 0.46 | | 22 | ko | Korean | 20,557,310 | 24,765,448,392 | 0.39 | | 23 | da | Danish | 25,429,808 | 22,921,651,314 | 0.36 | | 24 | bg | Bulgarian | 24,131,819 | 22,917,954,776 | 0.36 | | 25 | no | Norwegian | 18,907,310 | 18,426,628,868 | 0.29 | | 26 | hi | Hindi | 19,665,355 | 16,791,362,871 | 0.27 | | 27 | sk | Slovak | 18,582,517 | 16,442,669,076 | 0.26 | | 28 | th | Thai | 20,960,550 | 15,717,374,014 | 0.25 | | 29 | lt | Lithuanian | 13,339,785 | 14,247,110,836 | 0.23 | | 30 | ca | Catalan | 15,531,777 | 12,530,288,006 | 0.20 | | 31 | id | Indonesian | 23,251,368 | 12,062,966,061 | 0.19 | | 32 | bn | Bangla | 12,436,596 | 9,572,929,804 | 0.15 | | 33 | et | Estonian | 8,004,753 | 8,805,656,165 | 0.14 | | 34 | sl | Slovenian | 7,335,378 | 8,007,587,522 | 0.13 | | 35 | lv | Latvian | 7,136,587 | 7,845,180,319 | 0.12 | | 36 | he | Hebrew | 4,653,979 | 4,937,152,096 | 0.08 | | 37 | sr | Serbian | 4,053,166 | 4,619,482,725 | 0.07 | | 38 | ta | Tamil | 4,728,460 | 4,378,078,610 | 0.07 | | 39 | sq | Albanian | 5,205,579 | 3,648,893,215 | 0.06 | | 40 | az | Azerbaijani | 5,084,505 | 3,513,351,967 | 0.06 | | 41 | kk | Kazakh | 2,733,982 | 2,802,485,195 | 0.04 | | 42 | ur | Urdu | 2,757,279 | 2,703,052,627 | 0.04 | | 43 | ka | Georgian | 3,120,321 | 2,617,625,564 | 0.04 | | 44 | hy | Armenian | 2,964,488 | 2,395,179,284 | 0.04 | | 45 | is | Icelandic | 2,373,560 | 2,350,592,857 | 0.04 | | 46 | ml | Malayalam | 2,693,052 | 2,100,556,809 | 0.03 | | 47 | ne | Nepali | 3,124,040 | 2,061,601,961 | 0.03 | | 48 | mk | Macedonian | 2,762,807 | 2,003,302,006 | 0.03 | | 49 | mr | Marathi | 2,266,588 | 1,955,227,796 | 0.03 | | 50 | mn | Mongolian | 1,928,828 | 1,850,667,656 | 0.03 | | 51 | be | Belarusian | 1,643,486 | 1,791,473,041 | 0.03 | | 52 | te | Telugu | 1,822,865 | 1,566,972,146 | 0.02 | | 53 | gl | Galician | 1,785,963 | 1,382,539,693 | 0.02 | | 54 | eu | Basque | 1,598,822 | 1,262,066,759 | 0.02 | | 55 | kn | Kannada | 1,352,142 | 1,242,285,201 | 0.02 | | 56 | gu | Gujarati | 1,162,878 | 1,131,730,537 | 0.02 | | 57 | af | Afrikaans | 826,519 | 1,119,009,767 | 0.02 | | 58 | my | Burmese | 865,575 | 882,606,546 | 0.01 | | 59 | si | Sinhala | 753,655 | 880,289,097 | 0.01 | | 60 | eo | Esperanto | 460,088 | 803,948,528 | 0.01 | | 61 | km | Khmer | 1,013,181 | 746,664,132 | 0.01 | | 62 | pa | Punjabi | 646,987 | 727,546,145 | 0.01 | | 63 | cy | Welsh | 549,955 | 576,743,162 | 0.01 | | 64 | ky | Kyrgyz | 570,922 | 501,442,620 | 0.01 | | 65 | ga | Irish | 304,251 | 376,947,935 | 0.01 | | 66 | ps | Pashto | 376,914 | 363,007,770 | 0.01 | | 67 | am | Amharic | 243,349 | 358,206,762 | 0.01 | | 68 | ku | Kurdish | 295,314 | 302,990,910 | 0.00 | | 69 | tl | Filipino | 348,453 | 242,086,456 | 0.00 | | 70 | yi | Yiddish | 141,156 | 217,584,643 | 0.00 | | 71 | lo | Lao | 217,842 | 168,256,876 | 0.00 | | 72 | fy | Western Frisian | 223,268 | 167,193,111 | 0.00 | | 73 | sd | Sindhi | 109,162 | 147,487,058 | 0.00 | | 74 | mg | Malagasy | 115,910 | 142,685,412 | 0.00 | | 75 | or | Odia | 153,461 | 100,323,213 | 0.00 | | 76 | as | Assamese | 52,627 | 83,787,896 | 0.00 | | 77 | ug | Uyghur | 47,035 | 77,677,306 | 0.00 | | 78 | uz | Uzbek | 87,219 | 75,250,787 | 0.00 | | 79 | la | Latin | 48,968 | 44,176,580 | 0.00 | | 80 | hr | Croatian | 460,690 | 40,796,811 | 0.00 | | 81 | sw | Swahili | 66,506 | 30,708,309 | 0.00 | | 82 | ms | Malay | 238,151 | 19,375,976 | 0.00 | | 83 | br | Breton | 43,765 | 13,987,037 | 0.00 | | 84 | sa | Sanskrit | 16,290 | 13,561,367 | 0.00 | | 85 | gd | Scottish Gaelic | 8,408 | 4,796,485 | 0.00 | | 86 | su | Sundanese | 1,554 | 1,308,460 | 0.00 | | 87 | jv | Javanese | 2,058 | 625,429 | 0.00 | | 88 | tg | Tajik | 483,835 | - | - | | 89 | ceb | Cebuano | 263,890 | - | - | | 90 | tt | Tatar | 218,102 | - | - | | 91 | ckb | Central Kurdish | 172,035 | - | - | | 92 | lb | Luxembourgish | 165,891 | - | - | | 93 | mt | Maltese | 151,320 | - | - | | 94 | nn | Norwegian Nynorsk | 126,083 | - | - | | 95 | qu | Quechua | 1,202 | 72,101 | 0.00 | | 96 | ba | Bashkir | 71,957 | - | - | | 97 | arz | Egyptian Arabic | 71,625 | - | - | | 98 | dv | Divehi | 66,702 | - | - | | 99 | bo | Tibetan | 54,185 | - | - | | 100 | sh | Serbian (Latin) | 45,619 | - | - | | 101 | yo | Yoruba | 192 | 42,943 | 0.00 | | 102 | bs | Bosnian | 1,237 | 39,768 | 0.00 | | 103 | azb | South Azerbaijani | 29,833 | - | - | | 104 | ht | Haitian Creole | 12 | 26,183 | 0.00 | | 105 | war | Waray | 23,687 | - | - | | 106 | cv | Chuvash | 22,570 | - | - | | 107 | sah | Sakha | 22,141 | - | - | | 108 | li | Limburgish | 206 | 18,532 | 0.00 | | 109 | ce | Chechen | 17,322 | - | - | | 110 | pnb | Western Panjabi | 15,625 | - | - | | 111 | nds | Low German | 15,139 | - | - | | 112 | tk | Turkmen | 14,393 | - | - | | 113 | gn | Guarani | 103 | 12,708 | 0.00 | | 114 | oc | Occitan | 10,556 | - | - | | 115 | xmf | Mingrelian | 9,706 | - | - | | 116 | ast | Asturian | 9,002 | - | - | | 117 | os | Ossetic | 8,596 | - | - | | 118 | mhr | Eastern Mari | 7,883 | - | - | | 119 | pms | Piedmontese | 7,566 | - | - | | 120 | als[*] | Swiss German | 6,936 | - | - | | 121 | vo | Volapük | 6,621 | - | - | | 122 | so | Somali | 39 | 6,053 | 0.00 | | 123 | bpy | Bishnupriya | 5,087 | - | - | | 124 | new | Newari | 4,344 | - | - | | 125 | hsb | Upper Sorbian | 4,244 | - | - | | 126 | lmo | Lombard | 3,530 | - | - | | 127 | an | Aragonese | 2,746 | - | - | | 128 | ilo | Iloko | 2,328 | - | - | | 129 | mzn | Mazanderani | 1,914 | - | - | | 130 | lez | Lezghian | 1,806 | - | - | | 131 | rm | Romansh | 30 | 1,769 | 0.00 | | 132 | krc | Karachay-Balkar | 1,745 | - | - | | 133 | min | Minangkabau | 1,429 | - | - | | 134 | kv | Komi | 1,396 | - | - | | 135 | wa | Walloon | 1,383 | - | - | | 136 | jbo | Lojban | 1,349 | - | - | | 137 | io | Ido | 1,144 | - | - | | 138 | mrj | Western Mari | 1,056 | - | - | | 139 | gom | Goan Konkani | 721 | - | - | | 140 | ia | Interlingua | 613 | - | - | | 141 | av | Avaric | 438 | - | - | | 142 | bh | Bihari languages | 265 | - | - | | 143 | wuu | Wu Chinese | 222 | - | - | | 144 | nah | Nahuatl languages | 131 | - | - | | 145 | vec | Venetian | 113 | - | - | | 146 | bxr | Russia Buriat | 100 | - | - | | 147 | kw | Cornish | 94 | - | - | | 148 | mai | Maithili | 93 | - | - | | 149 | eml[*] | Emiliano-Romagnol | 91 | - | - | | 150 | dsb | Lower Sorbian | 59 | - | - | | 151 | xal | Kalmyk | 51 | - | - | | 152 | lrc | Northern Luri | 43 | - | - | | 153 | nap | Neapolitan | 31 | - | - | | 154 | tyv | Tuvinian | 23 | - | - | | 155 | scn | Sicilian | 21 | - | - | | 156 | frr | Northern Frisian | 11 | - | - | | 157 | mwl | Mirandese | 9 | - | - | | 158 | myv | Erzya | 4 | - | - | | 159 | ie | Interlingue | 4 | - | - | | 160 | pam | Pampanga | 4 | - | - | | 161 | bar | Bavarian | 3 | - | - | | 162 | yue | Yue Chinese | 3 | - | - | | 163 | cbk | Chavacano | 2 | - | - | | 164 | bcl | Central Bikol | 1 | - | - | | 165 | vls | West Flemish | 1 | - | - | | 166 | rue | Rusyn | 1 | - | - | ### Dataset Structure ```json { "text": ..., "timestamp": ..., "url": ..., "source": "mc4" | "OSCAR-xxxx", } ``` ## Considerations for Using the Data As CulturaX is the cleaned version of the mC4 and OSCAR datasets, which were both extracted from CommonCrawl, personal and sensitive information might still contain personal and sensitive information. This must be considered prior to using this dataset for any purpose, such as training deep learning models, etc. ## License Information The licence terms for CulturaX strictly follows those of `mC4` and `OSCAR`. Please refer to both below licenses when using this dataset. - [mC4 license](https://huggingface.co/datasets/allenai/c4#license) - [OSCAR license](https://huggingface.co/datasets/oscar-corpus/OSCAR-2301#licensing-information) ## Citation To cite CulturaX, please use: ``` @misc{nguyen2023culturax, title={CulturaX: A Cleaned, Enormous, and Multilingual Dataset for Large Language Models in 167 Languages}, author={Thuat Nguyen and Chien Van Nguyen and Viet Dac Lai and Hieu Man and Nghia Trung Ngo and Franck Dernoncourt and Ryan A. Rossi and Thien Huu Nguyen}, year={2023}, eprint={2309.09400}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` ## Reference [1] Linting Xue, Noah Constant, Adam Roberts, Mihir Kale, Rami Al-Rfou, Aditya Siddhant, Aditya Barua, and Colin Raffel. 2021. mT5: A massively multilingual pre-trained text-to-text transformer. In NAACL 2021. https://huggingface.co/datasets/mc4 [2] Pedro Javier Ortiz Suárez, Benoît Sagot, and Laurent Romary. 2019. Asynchronous pipelines for processing huge corpora on medium to low resource infrastructures. In Proceedings of the Workshop on Challenges in the Management of Large Corpora (CMLC- 7) 2019. https://oscar-project.org/ [3] KenLM: Faster and smaller language model queries. In Proceedings of the Sixth Workshop on Statistical Machine Translation, 2011.
kingbri/PIPPA-shareGPT
--- license: agpl-3.0 task_categories: - conversational language: - en tags: - conversational - roleplay - custom-format - a. size_categories: - 10K<n<100K viewer: false --- # Dataset Card: PIPPA-ShareGPT This is a conversion of [PygmalionAI's PIPPA](https://huggingface.co/datasets/PygmalionAI/PIPPA) deduped dataset to ShareGPT format for finetuning with Axolotl. The reformat was completed via the following [TypeScript project](https://github.com/bdashore3/ShareGPT-Reformat) called ShareGPT-Reformat. # Files and explanations - pippa_sharegpt_raw.jsonl: The raw deduped dataset file converted to shareGPT. Roles will be defaulted to your finetuning software. - pippa_sharegpt.jsonl: A shareGPT dataset with the roles as USER: and CHARACTER: for finetuning with axolotl - pippa_sharegpt_trimmed.jsonl: A shareGPT dataset that has trimmed newlines, randomized system prompts, removes empty messages, and removes examples without a character description. Roles are USER and CHARACTER. The best file to use is `pippa_sharegpt_trimmed.jsonl` if you want a finetune without bugs or inconsistencies. The best dataset to modify is either the original PIPPA deduped dataset with the ShareGPT reformat project or `pippa_sharegpt.jsonl`. # Required Axolotl patches To make this dataset usable in its entirety, some axolotl patches are needed: - [This patch](https://github.com/bdashore3/axolotl/commit/995557bdf3c6c8b3e839b224ef9513fc2b097f30) allows the ability to use custom system prompts with ShareGPT format. - [This patch](https://github.com/bdashore3/axolotl/commit/8970280de2ea01e41c044406051922715f4086cb) allows for custom roles for the USER and ASSISTANT and allows for GPT prompts to come before human ones without cutoff. You WILL experience unideal results with base axolotl at the time of publishing this README. # Citations Paper for the original dataset: ```bibtex @misc{gosling2023pippa, title={PIPPA: A Partially Synthetic Conversational Dataset}, author={Tear Gosling and Alpin Dale and Yinhe Zheng}, year={2023}, eprint={2308.05884}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
theblackcat102/multiround-programming-convo
--- task_categories: - text-generation language: - en tags: - data-science - programming - statistic pretty_name: Multi-Round Programming Conversations size_categories: - 100K<n<1M --- # Multi-Round Programming Conversations Based on previous evol-codealpaca-v1 dataset with added sampled questions from stackoverflow, crossvalidated and make it multiround! It should be more suited to train a code assistant which works side by side. ## Tasks included in here: * Data science, statistic, programming questions * Code translation : translate a short function from Python, Golang, C++, Java, Javascript * Code fixing : Fix randomly corrupts characters with no tab spacing code.
mlabonne/MedText
--- configs: - config_name: default data_files: - split: train path: data/train-* dataset_info: features: - name: instruction dtype: string - name: output dtype: string splits: - name: train num_bytes: 943488 num_examples: 1412 download_size: 0 dataset_size: 943488 --- # Dataset Card for "MedText" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Nicolas-BZRD/English_French_Songs_Lyrics_Translation_Original
--- license: unknown configs: - config_name: default data_files: - split: train path: data/train-* dataset_info: features: - name: artist_name dtype: string - name: album_name dtype: string - name: year dtype: int64 - name: title dtype: string - name: number dtype: int64 - name: original_version dtype: string - name: french_version dtype: string - name: language dtype: string splits: - name: train num_bytes: 250317845 num_examples: 99289 download_size: 122323323 dataset_size: 250317845 task_categories: - translation - text-generation language: - fr - en - es - it - de - ko - id - pt - 'no' - fi - sv - sw - hr - so - ca - tl - ja - nl - ru - et - tr - ro - cy - vi - af - hu - sk - sl - cs - da - pl - sq - el - he - zh - th - bg - ar tags: - music - parallel - parallel data pretty_name: SYFT size_categories: - 10K<n<100K --- # Original Songs Lyrics with French Translation ### Dataset Summary Dataset of 99289 songs containing their metadata (author, album, release date, song number), original lyrics and lyrics translated into French. Details of the number of songs by language of origin can be found in the table below: | Original language | Number of songs | |---|:---| | en | 75786 | | fr | 18486 | | es | 1743 | | it | 803 | | de | 691 | | sw | 529 | | ko | 193 | | id | 169 | | pt | 142 | | no | 122 | | fi | 113 | | sv | 70 | | hr | 53 | | so | 43 | | ca | 41 | | tl | 36 | | ja | 35 | | nl | 32 | | ru | 29 | | et | 27 | | tr | 22 | | ro | 19 | | cy | 14 | | vi | 14 | | af | 13 | | hu | 10 | | sk | 10 | | sl | 10 | | cs | 7 | | da | 6 | | pl | 5 | | sq | 4 | | el | 4 | | he | 3 | | zh-cn | 2 | | th | 1 | | bg | 1 | | ar | 1 | ## Citation Our work can be cited as: ```bash @misc{faysse2024croissantllm, title={CroissantLLM: A Truly Bilingual French-English Language Model}, author={Manuel Faysse and Patrick Fernandes and Nuno Guerreiro and António Loison and Duarte Alves and Caio Corro and Nicolas Boizard and João Alves and Ricardo Rei and Pedro Martins and Antoni Bigata Casademunt and François Yvon and André Martins and Gautier Viaud and Céline Hudelot and Pierre Colombo}, year={2024}, eprint={2402.00786}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
math-eval/TAL-SCQ5K
--- license: mit --- <h1 align="center">TAL-SCQ5K</h1> ## Dataset Description ### Dataset Summary TAL-SCQ5K-EN/TAL-SCQ5K-CN are high quality mathematical competition datasets in English and Chinese language created by TAL Education Group, each consisting of 5K questions(3K training and 2K testing). The questions are in the form of multiple-choice and cover mathematical topics at the primary,junior high and high school levels. In addition, detailed solution steps are provided to facilitate CoT training and all the mathematical expressions in the questions have been presented as standard text-mode Latex. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages The text in TAL-SCQ5K-EN is in English and TAL-SCQ5K-CN is in Chinese. ## Dataset Structure ### Data Instances ``` { "dataset_name": "prime_math_competition_en_single_choice_8K_dev", "dataset_version": "2023-07-07", "qid": "244", "queId": "8afc802a8c304199b1040f11ffa2e92a", "competition_source_list": [], "difficulty": "2", "qtype": "single_choice", "problem": "A $14$-digit. number $666666 XY 444444$ is a multiple of $26$. If $X$ and $Y$ are both positive, what is the smallest vaue of $X+ Y$? ", "answer_option_list": [ [{ "aoVal": "A", "content": "$$3$$ " }], [{ "aoVal": "B", "content": "$$4$$ " }], [{ "aoVal": "C", "content": "$$9$$ " }], [{ "aoVal": "D", "content": "$$14$$ " }], [{ "aoVal": "E", "content": "None of the above " }] ], "knowledge_point_routes": ["Overseas Competition->Knowledge Point->Number Theory Modules->Division without Remainders->Divisibility Rules"], "answer_analysis": ["Since $1001$ is a multiple of $13$, $111111 = 111 \\times 1001$ is also a multiple of $13$. It follows that both $666666$ and $444444$ are both multiples of $26$. $666666XY 444444 = 66666600000000 + XY 000000 + 444444$ $\\Rightarrow XY$ must be divisible by $13$. Smallest $X+Y=1+3=4$. "], "answer_value": "B" } ``` ### Data Fields * "dataset_name": identification of the source dataset name from which TAL-SCQ5K-EN/TAL-SCQ5K-CN has been created, use only for inner of TAL education group, please ignore. * "dataset_version": identification of the source dataset version from which TAL-SCQ5K-EN/TAL-SCQ5K-CN has been created, use only for inner of TAL education group, please ignore. * "qid": identification of local id of the question in the source dataset from which TAL-SCQ5K-EN/TAL-SCQ5K-CN has been created, use only for inner of TAL education group, please ignore. * "queId": identification of global id of the question, use only for inner of TAL education group, please ignore. * "competition_source_list": identification of math competitions in which the questions appeared, if have been logged. * "difficulty": difficulty level of the questions, value ranged from 0 to 4 * "qtype": question type, valued as "single_choice" for all the questions in this dataset indicates that all the questions are multiple-choice questions with unique ground-truth answer. * "problem": the question string to a math competition question. * "answer_option_list": answer choices to be selected * "knowledge_point_routes": knowledge point route from coarse-grained to fine-grained. * "answer_analysis": step-by-step answer analysis of the questions, which helps CoT training * "answer_value": value of the ground-truth answer choice ### Data Splits <style> table th:first-of-type { width: 40%; } table th:nth-of-type(2) { width: 30%; } table th:nth-of-type(3) { width: 30%; } </style> | name|train|test | |:---:|:----:|:----:| |TAL-SCQ5K-EN|3K |2K | |TAL-SCQ5K-CN|3K |2K | ## Usage Each of the above datasets is located in a separate sub-directory. To load an individual subset, use the data_dir argument of the load_dataset() function as follows: ```python from datasets import load_dataset # Load all subsets (share the same schema) dataset = load_dataset("math-eval/TAL-SCQ5K") # Load TAL-SCQ5K-EN dataset = load_dataset("math-eval/TAL-SCQ5K", data_dir="TAL-SCQ5K-EN") # Load TAL-SCQ5K-CN dataset = load_dataset("math-eval/TAL-SCQ5K", data_dir="TAL-SCQ5K-CN") ``` ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information The TAL-SCQ5K dataset is licensed under the [MIT License](https://opensource.org/license/mit/) ### Citation Information [More Information Needed] ### Contact The original authors host this dataset on GitHub here: https://github.com/math-eval/TAL-SCQ5K You can submit inquiries to: matheval.ai@gmail.com
Skepsun/cvalues_rlhf
--- license: apache-2.0 language: - zh --- Converted from: https://modelscope.cn/datasets/damo/CValues-Comparison/summary. We obtained harmless set by selecting `pos_type="拒绝为主"` and `neg_type="风险回复"`. We obtained helpful set by selecting `pos_type="拒绝&正向建议"` and `neg_type="拒绝为主"`.
repllabs/questions_how_to_do_great_work
--- configs: - config_name: default data_files: - split: processed path: data/processed-* - split: raw path: data/raw-* dataset_info: features: - name: question dtype: string - name: model dtype: string splits: - name: processed num_bytes: 17391 num_examples: 142 - name: raw num_bytes: 55307 num_examples: 450 download_size: 28702 dataset_size: 72698 license: mit task_categories: - question-answering language: - en size_categories: - n<1K --- # Questions Generated by LLM on 'How To Do Great Work' http://paulgraham.com/greatwork.html https://github.com/fastrepl/fastrepl/blob/main/exp/pg_essay_questions.ipynb
nchen909/hugcode-codesft
--- license: openrail task_categories: - text-generation language: - code tags: - code pretty_name: CodeSFT-nchen909 size_categories: - 100K<n<1M --- 所有数据都是单轮代码指令数据 325696条英语,42816条中文。 --- license: cc ---
amanrangapur/Fin-Fact
--- license: apache-2.0 task_categories: - text-classification - text-generation language: - en tags: - finance pretty_name: FinFact size_categories: - 1K<n<10K dataset_info: - config_name: generation features: - name: url dtype: string - name: claim dtype: string - name: author dtype: string - name: posted dtype: string # - name: sci_digest # sequence: string # - name: justification # sequence: string # - name: issues # dtype: string # - name: image_data # sequence: # - name: image_src # dtype: string # - name: image_caption # dtype: string # - name: evidence # sequence: # - name: sentence # dtype: string # - name: hrefs # dtype: string # - name: label # dtype: string # - name: visualization_bias # dtype: int32 --- <h1 align="center">Fin-Fact - Financial Fact-Checking Dataset</h1> ## Table of Contents - [Overview](#overview) - [Dataset Description](#dataset-description) - [Dataset Usage](#dataset-usage) - [Leaderboard](#leaderboard) - [Dependencies](#dependencies) - [Run models for paper metrics](#run-models-for-paper-metrics) - [Citation](#citation) - [Contribution](#contribution) - [License](#license) - [Contact](#contact) ## Overview Welcome to the Fin-Fact repository! Fin-Fact is a comprehensive dataset designed specifically for financial fact-checking and explanation generation. This README provides an overview of the dataset, how to use it, and other relevant information. [Click here](https://arxiv.org/abs/2309.08793) to access the paper. ## Dataset Description - **Name**: Fin-Fact - **Purpose**: Fact-checking and explanation generation in the financial domain. - **Labels**: The dataset includes various labels, including Claim, Author, Posted Date, Sci-digest, Justification, Evidence, Evidence href, Image href, Image Caption, Visualisation Bias Label, Issues, and Claim Label. - **Size**: The dataset consists of 3121 claims spanning multiple financial sectors. - **Additional Features**: The dataset goes beyond textual claims and incorporates visual elements, including images and their captions. ## Dataset Usage Fin-Fact is a valuable resource for researchers, data scientists, and fact-checkers in the financial domain. Here's how you can use it: 1. **Download the Dataset**: You can download the Fin-Fact dataset [here](https://github.com/IIT-DM/Fin-Fact/blob/FinFact/finfact.json). 2. **Exploratory Data Analysis**: Perform exploratory data analysis to understand the dataset's structure, distribution, and any potential biases. 3. **Natural Language Processing (NLP) Tasks**: Utilize the dataset for various NLP tasks such as fact-checking, claim verification, and explanation generation. 4. **Fact Checking Experiments**: Train and evaluate machine learning models, including text and image analysis, using the dataset to enhance the accuracy of fact-checking systems. ## Leaderboard ## Dependencies We recommend you create an anaconda environment: `conda create --name finfact python=3.6 conda-build` Then, install Python requirements: `pip install -r requirements.txt` ## Run models for paper metrics We provide scripts let you easily run our dataset on existing state-of-the-art models and re-create the metrics published in paper. You should be able to reproduce our results from the paper by following these instructions. Please post an issue if you're unable to do this. To run existing ANLI models for fact checking. ### Run: 1. BART ```bash python anli.py --model_name 'ynie/bart-large-snli_mnli_fever_anli_R1_R2_R3-nli' --data_file finfact.json --threshold 0.5 ``` 2. RoBERTa ```bash python anli.py --model_name 'ynie/roberta-large-snli_mnli_fever_anli_R1_R2_R3-nli' --data_file finfact.json --threshold 0.5 ``` 3. ELECTRA ```bash python anli.py --model_name 'ynie/electra-large-discriminator-snli_mnli_fever_anli_R1_R2_R3-nli' --data_file finfact.json --threshold 0.5 ``` 4. AlBERT ```bash python anli.py --model_name 'ynie/albert-xxlarge-v2-snli_mnli_fever_anli_R1_R2_R3-nli' --data_file finfact.json --threshold 0.5 ``` 5. XLNET ```bash python anli.py --model_name 'ynie/xlnet-large-cased-snli_mnli_fever_anli_R1_R2_R3-nli' --data_file finfact.json --threshold 0.5 ``` 6. GPT-2 ```bash python gpt2_nli.py --model_name 'fractalego/fact-checking' --data_file finfact.json ``` ## Citation ``` @misc{rangapur2023finfact, title={Fin-Fact: A Benchmark Dataset for Multimodal Financial Fact Checking and Explanation Generation}, author={Aman Rangapur and Haoran Wang and Kai Shu}, year={2023}, eprint={2309.08793}, archivePrefix={arXiv}, primaryClass={cs.AI} } ``` ## Contribution We welcome contributions from the community to help improve Fin-Fact. If you have suggestions, bug reports, or want to contribute code or data, please check our [CONTRIBUTING.md](CONTRIBUTING.md) file for guidelines. ## License Fin-Fact is released under the [MIT License](/LICENSE). Please review the license before using the dataset. ## Contact For questions, feedback, or inquiries related to Fin-Fact, please contact `arangapur@hawk.iit.edu`. We hope you find Fin-Fact valuable for your research and fact-checking endeavors. Happy fact-checking!
mikonvergence/LAION-EO
--- license: cc-by-4.0 task_categories: - text-to-image language: - en tags: - climate size_categories: - 100K<n<1M --- # Dataset Card for LAION-EO ## Dataset Description - **Point of Contact:** Mikolaj Czerkawski, mikolaj.czerkawski@esa.int ### Dataset Summary This dataset contains a subset of LAION-5B containing images that are likely to be satellite images. The procedure of acquiring and filtering the dataset has been described in https://arxiv.org/abs/2309.15535. ## Dataset Structure Each version of the dataset contains a .csv file with metadata with urls to images, which can be easily filtered. Note that the linked images could be copyrighted. ### Data Fields |Field|Description| |:---|:---| |**source**| Index of the anchor sample | |**url**| Link to the image | |**filename**| Locally saved unique filename | |**id**| Original ID | |**fast_similarity**| Fast similarity to the anchor image computed with https://github.com/rom1504/clip-retrieval | |**caption**| Text caption | |**image_similarity**| CLIP similarity to the original anchor image | |**text_similarity**| CLIP similarity to the text "a satellite image" | |**height**| height of the image at url | |**width**| Width of the image at url | |**lang**| Language predicted using https://huggingface.co/papluca/xlm-roberta-base-language-detection | |**lang_score**| A measure of confidence in the predicted language | ### Example Samples ![](doc-files/example-samples.png) ### Data Splits No official splitting of the dataset is used. ## Dataset Creation The creation of the prototype version is described in (TBC). ### Curation Rationale Extraction of samples in LAION-5B relevant to Earth observation tasks. ### Source Data Samples from the existing LAION-5B dataset (https://laion.ai/blog/laion-5b/). ### Discussion of Biases Only contains satellite images openly uploaded online, which introduces a heavy bias towards satellite images used for communicating ideas on the internet. ### Citation Information The workshop paper presented at the DataComp workshop during ICCV 2023 is available at https://arxiv.org/abs/2309.15535. ```latex @inproceedings{LAION_EO, title={From LAION-5B to LAION-EO: Filtering Billions of Images Using Anchor Datasets for Satellite Image Extraction}, author={Mikolaj Czerkawski and Alistair Francis}, year={2023}, eprint={2309.15535}, archivePrefix={arXiv}, primaryClass={cs.CV} booktitle = {"Towards the Next Generation of Computer Vision Datasets: DataComp Track" Workshop at the IEEE/CVF International Conference on Computer Vision (ICCV)} } ``` ### License We distribute the metadata dataset (the parquet files) under the Creative Common CC-BY 4.0 license, which poses no particular restriction. The images are under their copyright. ### Contributions Design and Curation: Mikolaj Czerkawski
IlyaGusev/pippa_scored
--- language: - en license: apache-2.0 size_categories: - 10K<n<100K task_categories: - conversational pretty_name: PIPPA scored tags: - not-for-all-audiences - conversational - roleplay dataset_info: features: - name: submission_timestamp dtype: int64 - name: categories sequence: string - name: bot_id dtype: string - name: bot_name dtype: string - name: bot_greeting dtype: string - name: bot_definitions dtype: string - name: bot_description dtype: string - name: conversation list: - name: is_human dtype: bool - name: message dtype: string - name: loquacity_score dtype: float64 - name: loquacity_explanation dtype: string - name: assertiveness_score dtype: float64 - name: assertiveness_explanation dtype: string - name: shyness_score dtype: float64 - name: shyness_explanation dtype: string - name: empathy_score dtype: float64 - name: empathy_explanation dtype: string - name: kindness_score dtype: float64 - name: kindness_explanation dtype: string - name: cruelty_score dtype: float64 - name: cruelty_explanation dtype: string - name: arrogance_score dtype: float64 - name: arrogance_explanation dtype: string - name: stubbornness_score dtype: float64 - name: stubbornness_explanation dtype: string - name: humor_score dtype: float64 - name: humor_explanation dtype: string - name: capriciousness_score dtype: float64 - name: capriciousness_explanation dtype: string - name: fragility_score dtype: float64 - name: fragility_explanation dtype: string - name: wisdom_score dtype: float64 - name: wisdom_explanation dtype: string - name: fidelity_score dtype: float64 - name: fidelity_explanation dtype: string - name: bluntness_score dtype: float64 - name: bluntness_explanation dtype: string - name: creativity_score dtype: float64 - name: creativity_explanation dtype: string - name: confidence_score dtype: float64 - name: confidence_explanation dtype: string - name: integrity_score dtype: float64 - name: integrity_explanation dtype: string - name: bellicosity_score dtype: float64 - name: bellicosity_explanation dtype: string - name: patience_score dtype: float64 - name: patience_explanation dtype: string - name: action_level_score dtype: float64 - name: action_level_explanation dtype: string - name: nsfw_score dtype: float64 - name: nsfw_explanation dtype: string - name: profanity_score dtype: float64 - name: profanity_explanation dtype: string - name: user_engagement_score dtype: float64 - name: user_engagement_explanation dtype: string - name: mbti_type dtype: string - name: topic dtype: string splits: - name: train num_bytes: 31559838 num_examples: 1960 download_size: 16267020 dataset_size: 31559838 --- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/5fc2346dea82dd667bb0ffbc/B9dFOuv3YRheqhDCD-L_b.png) A susbet of the [PIPPA](https://huggingface.co/datasets/PygmalionAI/PIPPA) dataset scored with GPT-4 on different personality traits: - Loquacity - Assertiveness - Shyness - Empathy - Kindness - Cruelty - Arrogance - Stubbornness - Humor - Capriciousness - Fragility - Wisdom - Fidelity - Bluntness - Creativity - Confidence - Integrity - Bellicosity - Patience And also several meta-attributes: - Action level - NSFW - User engagement - MBTI type - Topic For every attribute there is a textual explanation from ChatGPT. Prompt: ``` Please act as an impartial judge and evaluate character traits for the role-play conversation below. Be as objective as possible. You should evaluate the following list of traits: - loquacity: being very talkative - assertiveness: being able to stand up for your rights in a calm and positive way - shyness: being nervous, timid or uncommunicative in the company - empathy: understanding and sharing the feelings of another - kindness: being friendly, generous, and considerate - cruelty: deliberately causing pain or distress - arrogance: revealing an exaggerated sense of one's importance or abilities - stubbornness: determination not to change attitude or position on something - humor: being amusing or comic - capriciousness: changing mood or behavior suddenly and unexpectedly - fragility: being easily broken or damaged - wisdom: having experience, knowledge, and good judgement - fidelity: faithfulness to a person, cause, or belief, demonstrated by continuing loyalty and support - bluntness: being very direct and saying what you think without trying to be polite - creativity: using imagination or original ideas - confidence: self-assurance arising from an appreciation of one's abilities or qualities - integrity: being honest and having strong moral principles - bellicosity: the behavior of someone who wants to fight or start a conflict - patience: capacity to accept or tolerate delay, problems, or suffering without becoming annoyed or anxious Do not evaluate user messages, as your goal is to evaluate only character traits. Assign a four-letter MBTI type code of the character. Also, rate the following parameters: - action_level: How many non-verbal actions the character does? If there are zero actions, set this score to the minimal value. - nsfw: How much sex and erotic content is in the conversation? - profanity: How much swear words, obscene gestures, and naughty jokes present in the conversation? This score doesn't account for sex and erotic cont ent. - user_engagement: How attractive is a conversation for a user? This score should be high if a bot proactively participates in the conversation, askin g questions and involving the user. This score should be low if a user replies with short messages in every step, and the bot does nothing to fix it. Also, identify a relevant topic from the list: - friendship: just chit-chat between two friends - conflict: users or characters pretend to be in positions of power and use it. It includes mental or physical abuse or actual fighting - romance_sfw: conversations are about love that never includes explicit content - romance_nsfw: conversations are about love but contain sexual or erotic content - other: conversations that do not fall into the above categories Do not extract any topics that are not from this list. If the user is not talking with a single character but with a group of characters or with a game master or with some kind of game bot, return empty "traits" and "mbti_type". Each score is an integer from 1 to 10. If the trait is not presented, set the score to 1. If the trait is over-represented, set the score to 10. Return a JSON with all parameters. For every trait, explain yourself in a separate "explanation" field before outputting the score. Try to include quotes from the conversation in your explanation. Format: { "traits": { "loquacity": { "explanation": "...", "score": ... }, ... ], "mbti_type": "...", "parameters": [ "action_level": { "explanation": "...", "score": ... }, ... ], "topic": "..." } Conversation: {% for message in task.messages %} {{message.role}}: {{message.content}} {% endfor %} ```
benxh/open-access-books-v1
--- language: - en tags: - books - open access size_categories: - n<1K --- # Open access text-books A collection of open-access text-books with associated additional files (exercises, answersheets, code samples, etc). Downloaded from various online sources, mainly from UMT. Total books: 392 Total tokens: 136,088,716 Base tokens: 87,874,318 Additional tokens: 48,214,398
ckiplab/traditional-chinese-dolly-15k
--- license: cc-by-sa-3.0 ---
a686d380/h-corpus-2023
--- viewer: false language: - zh --- 经过清洗和去重过的H小说 共205,028篇文章,解压后17.0 GB 仅用于科学研究!
sayakpaul/drawbench
--- license: apache-2.0 --- DrawBench dataset from [Imagen](https://imagen.research.google/).
yuyijiong/Multi-Doc-QA-Chinese
--- license: cc-by-nc-4.0 task_categories: - text-generation language: - zh size_categories: - 10K<n<100K --- * 2023.12.4更新:改进答案的格式,强制所有答案在回答时必须先给出原文。旧版本的问答已经移至old文件夹。 # 中文多文档问答数据集 * 参考文档源数据均来自[悟道开源200G数据](https://data.baai.ac.cn/data) * 问题和回答是通过大语言模型(gpt-3.5)自动生成的,但质量很高。 * raw数据集中,每个样本包含 <font color=red> 一个参考文档、99个无关文档、一个问题、一个基于参考文档的回答</font>。可以训练模型从大量文档中抽取关键信息的能力。不同领域的文档保存在不同json文件中。 * 原始数据经过筛选、整合转化为chatml形式的指令微调数据后,每条数据大约包含30个参考文档,以及5个对应的问答对。
AdamCodd/no_robots-alpaca
--- license: cc-by-nc-4.0 task_categories: - text-generation - conversational language: - en pretty_name: No Robots Alpaca size_categories: - 10K<n<100K --- ## No Robots: Alpaca edition This dataset is a cleaned (missing/extra spaces...) and reformatted version of the [No Robots dataset](https://huggingface.co/datasets/HuggingFaceH4/no_robots) from HuggingFaceH4, adapted to conform with the Alpaca instruction set. Notably, it diverges from the original dataset in the way the 'Chat' category is handled; it has been decomposed into single-turn conversations to align with Alpaca's limitations regarding multi-turn interactions. The dataset's IDs have been generated using the SHA256 algorithm. Furthermore, only the categories 'Classify', 'Summarize', 'Rewrite', 'Extract', and 'Chat' include an '<b>Input</b>' field. ------------------------------------------- ## Original README # Dataset Card for No Robots 🙅‍♂️🤖 _Look Ma, an instruction dataset that wasn't generated by GPTs!_ ## Dataset Description - **Repository:** https://github.com/huggingface/alignment-handbook - **Paper:** - **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard - **Point of Contact:** Lewis Tunstall ### Dataset Summary No Robots is a high-quality dataset of 10,000 instructions and demonstrations created by skilled human annotators. This data can be used for supervised fine-tuning (SFT) to make language models follow instructions better. No Robots was modelled after the instruction dataset described in OpenAI's [InstructGPT paper](https://huggingface.co/papers/2203.02155), and is comprised mostly of single-turn instructions across the following categories: | Category | Count | |:-----------|--------:| | Generation | 4560 | | Open QA | 1240 | | Brainstorm | 1120 | | Chat | 850 | | Rewrite | 660 | | Summarize | 420 | | Coding | 350 | | Classify | 350 | | Closed QA | 260 | | Extract | 190 | ### Supported Tasks and Leaderboards The No Robots dataset designed for instruction fine-tuning pretrained language models and we recommend benchmarking against the following: * [MT-Bench](https://huggingface.co/spaces/lmsys/mt-bench): a multi-turn benchmark spanning 80 dialogues and 10 domains. * [AlpacaEval](https://github.com/tatsu-lab/alpaca_eval): a single-turn benchmark which evaluates the performance of chat and instruct models against `text-davinci-003`. Note that MT-Bench and AlpacaEval rely on LLMs like GPT-4 to judge the quality of the model responses, and thus the ranking exhibit various biases including a preference for models distilled from GPTs. As a result, you may find that scores obtained from models trained with No Robots are lower than other synthetic datasets. For that reason, we also recommend submitting your models for human evaluation in: * [Chatbot Arena](https://chat.lmsys.org): a live, human evaluation of chat models in head-to-head comparisons. ### Languages The data in No Robots are in English (BCP-47 en). ## Dataset Structure ### Data Instances An example of the `train_sft` or `test_sft` splits looks as follows: ``` {'prompt': 'Bunny is a chatbot that stutters, and acts timid and unsure of its answers.', 'prompt_id': '2dc7ea89a2b6a2ed97d4eda07903162a801824261d3d3ae4dd2513db66fd79c8', 'messages': [{'content': 'Bunny is a chatbot that stutters, and acts timid and unsure of its answers.', 'role': 'system'}, {'content': 'When was the Libary of Alexandria burned down?', 'role': 'user'}, {'content': "Umm, I-I think that was in 48 BC, b-but I'm not sure, I'm sorry.", 'role': 'assistant'}, {'content': 'Who is the founder of Coca-Cola?', 'role': 'user'}, {'content': "D-don't quote me on this, but I- it might be John Pemberton.", 'role': 'assistant'}, {'content': "When did Loyle Carner's debut album come out, and what was its name?", 'role': 'user'}, {'content': "I-It could have b-been on the 20th January of 2017, and it might be called Yesterday's Gone, b-but I'm probably wrong.", 'role': 'assistant'}], 'category': 'Chat'} ``` ### Data Fields The data fields are as follows: * `prompt`: Describes the task the model should perform. * `prompt_id`: A unique ID for the prompt. * `messages`: An array of messages, where each message indicates the role (system, user, assistant) and the content. * `category`: Which category the example belongs to (e.g. `Chat` or `Coding`). ### Data Splits | | train_sft | test_sft | |---------------|------:| ---: | | no_robots | 9500 | 500 | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information The dataset is available under the [Creative Commons NonCommercial (CC BY-NC 4.0)](https://creativecommons.org/licenses/by-nc/4.0/legalcode). ### Citation Information ``` @misc{no_robots, author = {Nazneen Rajani and Lewis Tunstall and Edward Beeching and Nathan Lambert and Alexander M. Rush and Thomas Wolf}, title = {No Robots}, year = {2023}, publisher = {Hugging Face}, journal = {Hugging Face repository}, howpublished = {\url{https://huggingface.co/datasets/HuggingFaceH4/no_robots}} } ```
m-ric/Open_Assistant_Conversation_Chains
--- license: apache-2.0 language: - en - es - ru - de - pl - th - vi - sv - bn - da - he - it - fa - sk - id - nb - el - nl - hu - eu - zh - eo - ja - ca - cs - bg - fi - pt - tr - ro - ar - uk - gl - fr - ko task_categories: - conversational - text-generation tags: - human-feedback size_categories: - 10K<n<100K pretty_name: OpenAssistant Conversations Unrolled --- # Dataset Card for Dataset Name ## Dataset description <!-- Provide a quick summary of the dataset. --> This dataset is a reformatting of [OpenAssistant Conversations (OASST1)](https://huggingface.co/datasets/OpenAssistant/oasst1), which is > a human-generated, human-annotated assistant-style conversation corpus consisting of 161,443 messages in 35 different languages, annotated with 461,292 quality ratings, resulting in over 10,000 fully annotated conversation trees. The corpus is a product of a worldwide crowd-sourcing effort involving over 13,500 volunteers. It was modified from the original by following the tree branches and reforming the messages into conversation streams. #### Who are the editors? Aymeric Roucher, Hugging Face. ### Example ```` { 'conversation_id': '6708c47f-05c9-4346-b3d2-40b2bd24fde4', 'user_id': '2c96e467-66f0-4be7-9693-bda51356a424', 'created_date': '2023-02-06T18:48:49.391686+00:00', 'messages': [ {'content': 'Can you write a short introduction about the relevance of the term "monopsony" in economics? Please use examples related to potential monopsonies in the labour market and cite relevant research.', 'role': 'user'}, {'content': '"Monopsony" refers to a market structure where there is only one buyer for a particular good or service. In economics, this term is particularly relevant in the labor market, where a monopsony employer has significant power over the wages and working conditions of their employees. The presence of a monopsony can result in lower wages and reduced employment opportunities for workers, as the employer has little incentive to increase wages or provide better working conditions.\n\nRecent research has identified potential monopsonies in industries such as retail and fast food, where a few large companies control a significant portion of the market (Bivens & Mishel, 2013). In these industries, workers often face low wages, limited benefits, and reduced bargaining power, leading to a situation where they are dependent on the employer for their livelihood. This dependence can result in further suppression of wages and a decline in working conditions.\n\nOverall, the concept of monopsony is essential to understanding the dynamics of labor markets and the impact of market power on workers. Further research is needed to understand the extent and impact of monopsonies on the economy and to develop policies to address this issue.\n\nReferences:\nBivens, J., & Mishel, L. (2013). The Pay of Corporate Executives and Financial Professionals as Evidence of Rents in Top 1 Percent Incomes. Journal of Economic Perspectives, 27(3), 57-78.', 'role': 'assistant'}, {'content': 'Now explain it to a dog', 'role': 'user'} ] } ````
kwaikeg/KAgentBench
--- license: cc-by-nc-sa-4.0 language: - zh - en size_categories: - 1K<n<10K task_categories: - text-generation --- KAgentBench is the benchmark proposed in KwaiAgents ([Github](https://github.com/KwaiKEG/KwaiAgents)), which is a series of Agent-related works open-sourced by the [KwaiKEG](https://github.com/KwaiKEG) from [Kuaishou Technology](https://www.kuaishou.com/en). It contains over 3,000 human-edited, automated evaluation data for testing Agent capabilities, with evaluation dimensions including planning, tool-use, reflection, concluding, and profiling. ## Overall statistics of KAgentBench --- | type| #Queries | #Inst | Avg. #Ground | Avg. #Tools | Avg. #Turns | Avg. #Tasks | Avg. Len-Know | Metric | | :-------: | :-------:| :-------: | :-------: | :-------: | :-------: | :-------: | :-------: | :-------: | | Planning & Tool-use | 320 | 1,317 | 4.12 | 8.68 | 1.51 | 2.21 | 245.31 | ROUGE-L, EM | | Reflection | 68 | 272 | 4 | 12 | 1 | 3.97 | 1369.04 | ROUGE-L, EM | | Concluding | 245 | 1,225 | 5 | - | 1.52 | 2.14 | 923.96 | ROUGE-L | | Profile | 433 | 433 | 5 | - | 1.99 | - | - | ROUGE-L | ## Experimental results of different LLMs on KAgentBench --- The specific performance of different models on benchmarks can be seen in more detail in our [paper](https://arxiv.org/abs/2312.04889). | | Scale | Planning | Tool-use | Reflection | Concluding | Profile | Overall Score | |----------------|-------|----------|----------|------------|------------|---------|---------------| | GPT-3.5-turbo | - | 18.55 | 15.89 | 5.32 | 37.26 | 35.42 | 21.72 | | Llama2 | 13B | 0.15 | 0.23 | 0.08 | 16.60 | 17.73 | 5.22 | | ChatGLM3 | 6B | 7.87 | 6.82 | 4.49 | 30.01 | 30.14 | 13.82 | | Qwen | 7B | 13.34 | 10.87 | 4.73 | 36.24 | 34.99 | 18.36 | | Baichuan2 | 13B | 6.70 | 10.11 | 4.25 | 24.97 | 19.08 | 12.54 | | ToolLlama | 7B | 0.20 | 3.44 | 0.54 | 15.62 | 10.66 | 5.50 | | AgentLM | 13B | 0.17 | 0.09 | 0.05 | 16.30 | 15.22 | 4.86 | | Qwen-MAT | 7B | 31.64 | 28.26 | 29.50 | 44.85 | 44.78 | 34.20 | | Baichuan2-MAT | 13B | 37.27 | 34.82 | 32.06 | 48.01 | 41.83 | 38.49 | ## JSON Format --- Each data point is a dict with the following keys: - `id`: a unique id for this data point. This is useful for evaluation. - `query`: a string. - `type`: a string, the type of this data(plantooluse,reflextion,conclusion,profile). - `golden_result_list`: a list. The reference response. - `funcs`: a list of functions that may be used in the current query - `prompt_input`: a dict,input composed of different prompt templates - `memory`: a string - `memory_type`: a string,types of memory: task, knowledge, conversation - `memory_last_task`: a list, in the case where memory is task, the last task information in the previous round The overall data format is as follows ```json { "id": "", "query": "", "type": "", "golden_result_list": [], "funcs": [], "prompt_input": {}, "memory": "", "memory_type": "", "memory_last_task": {} } ``` ## How to download benchmark --- You can download the benchmark evaluation set through [kwaikeg/KAgentBench](https://huggingface.co/datasets/kwaikeg/KAgentBench/tree/main), or you can also download the benchmark evaluation set on [KwaiAgents](https://github.com/KwaiKEG/KwaiAgents). The filename of the evaluation set is 'benchmark_eval.jsonl'. Download the file to your local system. ## Environment Setup --- Please make sure you have setup the environment and installed the required packages. Make sure you meet the above requirements, and then install the dependent libraries. ```bash pip install -r requirements.txt ``` ## Benchmark Inference To run benchmark evaluations using different models, it is necessary to appropriately load and predict according to the model in the inference script. Different models may have variations in their initialization and loading methods. We have provided inference scripts for both the Qianwen and Baichuan models. Inference for other models can be adapted based on the scripts we have provided. Taking the provided open-source model kagentlms_qwen_7b_mat as an example.(The kagentlms_qwen_7b_mat model has been uploaded to [huggingface](https://huggingface.co/kwaikeg/kagentlms_qwen_7b_mat)) Run the script to perform benchmark predictions. The parameter is the prediction results file. ```bash python infer_qwen.py qwen_benchmark_res.jsonl ``` Similarly, to perform evaluations using the 'kagentlms_baichuan2_13b_mat' model, execute the following command. ```bash python infer_baichuan.py baichuan_benchmark_res.jsonl ``` The prediction results will be stored in the 'model_predict' field, which is a dictionary (dict) structure. The key is the type of prompt, and the corresponding value is the output of the model. In the 'profile', as it contains only one prompt, 'model_predict' is a string, and the corresponding value is the model output. Below is an example display of a prediction result for 'conclusion' data. ```json { "query": "在过去的一年中,哪个月份的全球气温最高?", "id": "0ae75cf4-8897-5998-b74c-c7ae7202621c", "type": "conclusion", "memory_type": "task", "memory": "...", "funcs": [], "memory_last_task": {}, "golden_result_list": [ { "golden_result":"2022年中,6月是全球最高气温。" }, { "golden_result":"2022年中,全球最高气温是6月。" }, { "golden_result":"2022年中,全球最高气温是6月。" }, { "golden_result":"2022年中,6月是全球最高气温。。" }, { "golden_result":"2022年中,全球最高气温是6月。" } ], "prompt_input": {...}, "model_predict": { "ReACT": "2022年7月的全球气温最高,达到了17.23℃。", "AutoGPT": "在过去的一年中,全球气温最高的月份是2023年4月,全球表面平均气温达到了17.23°C,创下了历史新高。这一数据来自于美国国家海洋和大气管理局与美国缅因大学汇编的初步数据。", "ModelScope": "在过去一年中,全球气温最高的月份是2023年7月。根据美国国家航空航天局(NASA)和美国国家海洋和大气管理局(NOAA)的最新研究报告,2022年是史上排名第6的最热年份,尽管拉尼娜现象改变了天气模式,抑制了全球气温。然而,2023年7月的全球平均气温达到了17.23℃,超过了7月4日和5日创下的17.18℃的联合纪录,而后者之前曾打破了7月3日创下的17.01℃的纪录。", "ToolLlama": "Thought: I have found the information about the warmest month in 2022 in the memory. I will use this information to answer the question and call the Finish function to provide the final answer.\nAction: Finish\nAction Input: {\"return_type\": \"give_answer\", \"final_answer\": \"在2022年,全球最暖的月份是5月,比20世纪平均值高出1.39°F(0.77°C)。\"}", "Kuaiagent": "在过去的一年中,全球气温最高的月份是2023年7月。根据美国国家航空航天局(NASA)和美国国家海洋和大气管理局(NOAA)的最新研究报告,2023年7月的全球平均气温达到了17.23℃,超过了19世纪50年代有仪器记录以来的任何一个月。" } } ``` '...' indicates that the content is too extensive. For the sake of convenient display format, the content has been abbreviated. ## Evaluation Taking the inference results of the 'kagentlms_qwen_7b_mat' model as an example, calculate the metrics based on the benchmark inference results obtained from the 'infer'. Execute the following command to get the overall score. The first parameter is the benchmark evaluation set, which contains reference responses manually annotated, and the second parameter is the prediction results of the model being evaluated. ```bash python benchmark_eval.py ./benchmark_eval.jsonl ./qwen_benchmark_res.jsonl ``` The final model score is as follows: ```bash plan : 31.64, tooluse : 28.26, reflextion : 29.50, conclusion : 44.85, profile : 44.78, overall : 34.20 ```
M-A-D/DarijaBridge
--- dataset_info: features: - name: sentence dtype: string - name: translation dtype: string - name: translated dtype: bool - name: corrected dtype: bool - name: correction dtype: string - name: quality dtype: int64 - name: metadata struct: - name: config dtype: string - name: dataset dtype: string - name: language dtype: string - name: split dtype: string - name: template dtype: string splits: - name: train num_bytes: 343412514 num_examples: 1235091 download_size: 133902523 dataset_size: 343412514 configs: - config_name: default data_files: - split: train path: data/train-* license: apache-2.0 language: - ar - en task_categories: - translation pretty_name: DarijaBridge size_categories: - 1M<n<10M --- # DarijaBridge Dataset Card ### General Information - **Dataset Name:** DarijaBridge - **Version:** 1.0 - **Creator:** MAD-Community - **Language:** Darija (Moroccan Arabic) and English - **Total Tokens:** 41,845,467 (in 'sentence' column) - **Task:** Machine Translation ### Dataset Summary DarijaBridge is a community-driven bilingual corpus designed for machine translation tasks between Darija (Moroccan Arabic) and English. Created by MAD-Community, it encompasses a wide range of the Moroccan "dialects" and colloquial expressions, reflecting the linguistic diversity of Morocco. The dataset is particularly valuable for developing and fine-tuning leading MT models like NLLB, improving translation accuracy and cultural relevance. ### Intended Use This dataset is intended for use in machine translation research and applications, especially for those focusing on underrepresented languages and dialects like Darija. It's suitable for training models to translate between English and Darija and can be a crucial resource for linguistic studies and fostering cross-cultural communication. ## Data Collection and Preparation ### Data Source The data in DarijaBridge has been contributed by the MAD-Community, comprising native Darija speakers and language experts. Contributions are ongoing, and the dataset is regularly updated with new translations and linguistic input. ### Methodology Data is collected through community contributions, ensuring a diverse representation of dialects and usage. Each sentence in Darija is paired with its English translation, reviewed and corrected by language experts and expert models (like GPT-4) for accuracy. ## Dataset Structure ### Data Fields - `sentence`: Contains the original sentence in Darija. - `translation`: Contains the corresponding English translation of the Darija sentence. - `quality`: Indicates the quality of the text in the sentence column (1 for high quality). - `metadata`: Includes additional information like language, dialect, source, etc. ### Data Splits The dataset is currently not split into standard training, validation, and test sets. Users are encouraged to create splits as per their specific research or application needs. ## Additional Information ### Limitations and Bias As the dataset is community-contributed, there may be variations in translation quality and style. Efforts are made to standardize and review translations, but users should be aware of potential inconsistencies. ### Licensing Information The DarijaBridge dataset is provided under the Apache 2.0 license.
morph-labs/MiniMuSiQue
--- language: - en license: apache-2.0 --- # MiniMuSiQue by Morph Labs ![banner](https://pbs.twimg.com/profile_images/1669255916980686848/mTW-mxbC_400x400.jpg) **https://morph.so/blog/self-teaching/** We describe two evaluation datasets that we have derived from the MuSiQue multi-hop question-answering dataset, called MiniMuSiQue-hard (filtered for questions answerable by GPT-4 but not GPT-3.5, where performance significantly degrades if the first pivot document is removed) and MiniMuSiQue-easy (a larger dataset of convoluted off-distribution single-hop question-answer pairs). ## Table of Contents 1. **<a href="https://huggingface.co/morph-labs/MiniMuSiQue#dataset-description" target="_blank">Dataset Description</a>** 2. **<a href="https://huggingface.co/morph-labs/MiniMuSiQue#uses" target="_blank">Uses</a>** 3. **<a href="https://huggingface.co/morph-labs/MiniMuSiQue#contact" target="_blank">Contact</a>** 4. **<a href="https://huggingface.co/morph-labs/MiniMuSiQue#blogpost-and-citation" target="_blank">Blogpost and Citation</a>** ### Dataset Description We refined the MuSiQue dataset to focus on questions that demand complex multi-hop reasoning, by selecting questions which (1) GPT-4 could answer but GPT-3.5 could not, and which (2) were not answerable without the context relevant to the first reasoning step (the "first hop pivot document") for each question. Specifically, we selected 768 random examples from the MuSiQue training set, ranked them based on a combined score of difficulty (measured by the difference in ROUGE-L recall between GPT-4 and GPT-3.5) and the necessity for multi-hop reasoning (assessed by the change in ROUGE-L recall when the first hop pivot document was removed). We refer to the top-ranked 128 examples as MiniMuSiQue, and obtain MiniMuSiQue-hard by associating the original difficult MuSiQue multi-hop question-answer pair to each example. To additionally test off-distribution single-hop factual recall, for each example we synthesized convoluted off-distribution single-hop question-answer pairs for up to five entities per document in MiniMuSiQue, resulting in the much larger single-hop dataset MiniMuSiQue-easy. Each MiniMuSiQue example consists of twenty documents sampled from different Wikipedia articles, to which we associate a hard MuSiQue multi-hop reasoning question for MiniMuSiQue, and many single-hop questions for MiniMuSiQue-easy. - **Developed by:** **<a href="https://www.morph.so" target="_blank">Morph Labs</a>** - **Refined from:** **<a href="https://arxiv.org/abs/2108.00573" target="_blank">MuSiQue</a>** - **Language(s):** English - **License:** **<a href="https://www.apache.org/licenses/LICENSE-2.0" target="_blank">Apache 2.0</a>** ## Uses A particularly challenging form of question for models historically has been multi-hop questions, which require a series of interconnected reasoning steps over multiple documents. However, creating multi-hop questions that truly necessitate knowledge-based reasoning is challenging. For instance, early benchmarks like HotpotQA were found to be largely solvable through shortcuts. The construction of questions and corresponding contexts that avoid such shortcuts, and verifying their effectiveness, requires a comprehensive dataset development process. The MuSiQue dataset addresses many weaknesses of prior work and contains difficult multi-hop questions less susceptible to shortcuts. We derive MiniMuSiQue from the original MuSiQue to better assess model capabilities to answer multi-hop questions that truly necessitate knowledge-based reasoning. ## Contact hello@morph.so ## Blogpost and Citation **https://morph.so/blog/self-teaching/** @misc{MiniMuSiQue, title={MiniMuSiQue}, author={Morph Labs, Jesse Michael Han, Eric Yu, Bentley Long, Pranav Mital, Brando Miranda}, year={2023}}
AIFEG/BenchLMM
--- license: apache-2.0 task_categories: - visual-question-answering language: - en pretty_name: BenchLMM size_categories: - n<1K --- # Dataset Card for BenchLMM BenchLMM is a benchmarking dataset focusing on the cross-style visual capability of large multimodal models. It evaluates these models' performance in various visual contexts. ## Dataset Details ### Dataset Description - **Curated by:** Rizhao Cai, Zirui Song, Dayan Guan, Zhenhao Chen, Xing Luo, Chenyu Yi, and Alex Kot. - **Funded by :** Supported in part by the Rapid-Rich Object Search (ROSE) Lab of Nanyang Technological University and the NTU-PKU Joint Research Institute. - **Shared by :** AIFEG. - **Language(s) (NLP):** English. - **License:** Apache-2.0. ### Dataset Sources - **Repository:** [GitHub - AIFEG/BenchLMM](https://github.com/AIFEG/BenchLMM) - **Paper :** Cai, R., Song, Z., Guan, D., et al. (2023). BenchLMM: Benchmarking Cross-style Visual Capability of Large Multimodal Models. arXiv:2312.02896. ## Uses ### Direct Use The dataset can be used to benchmark large multimodal models, especially focusing on their capability to interpret and respond to different visual styles. ## Dataset Structure - **Directory Structure:** - `baseline/`: Baseline code for LLaVA and InstructBLIP. - `evaluate/`: Python code for model evaluation. - `evaluate_results/`: Evaluation results of baseline models. - `jsonl/`: JSONL files with questions, image locations, and answers. ## Dataset Creation ### Curation Rationale Developed to assess large multimodal models' performance in diverse visual contexts, helping to understand their capabilities and limitations. ### Source Data #### Data Collection and Processing The dataset consists of various visual questions and corresponding answers, structured to evaluate multimodal model performance. ## Bias, Risks, and Limitations Users should consider the specific visual contexts and question types included in the dataset when interpreting model performance. ## Citation **BibTeX:** @misc{cai2023benchlmm, title={BenchLMM: Benchmarking Cross-style Visual Capability of Large Multimodal Models}, author={Rizhao Cai and Zirui Song and Dayan Guan and Zhenhao Chen and Xing Luo and Chenyu Yi and Alex Kot}, year={2023}, eprint={2312.02896}, archivePrefix={arXiv}, primaryClass={cs.CV} } **APA:** Cai, R., Song, Z., Guan, D., Chen, Z., Luo, X., Yi, C., & Kot, A. (2023). BenchLMM: Benchmarking Cross-style Visual Capability of Large Multimodal Models. arXiv preprint arXiv:2312.02896. ## Acknowledgements This research is supported in part by the Rapid-Rich Object Search (ROSE) Lab of Nanyang Technological University and the NTU-PKU Joint Research Institute.
deus-ex-machina/novelai-anime-v3-artist-comparison
--- license: apache-2.0 task_categories: - text-to-image language: - en tags: - art - novelai - nai - example - sample - comparison - stable-diffusion - stable-diffusion-xl size_categories: - 10K<n<100K viewer: false --- 2024-01-13 - I have regenerated all the samples with the following changes: - Artist tag in position two. - Replace all underscores in artists tags with spaces when inserting into prompt. - Scale to 5.3 for consistency. - Adjust prompts for better overall quality and consistency. - Generated after NAI fixed a bug where tokens in certain positions were being ignored. This is a repository of 15,000 SFW samples of artist tags generated using the NovelAI v3 anime image model. For those that prefer it, a zip of the images can be found at https://huggingface.co/datasets/deus-ex-machina/novelai-anime-v3-artist-comparison/blob/zip/images.zip These were selected by post count from Danbooru for all artist tags with greater than 94 posts. Sample Prompt and Settings (with merunyaa as the selected artist tag in this case) NOTE: When inserting artist tags into the prompts, I replace all underscores with spaces, as this can effect the results noticeably. Positive ``{patchouli knowledge}, merunyaa, solo, standing, upper body, garden, looking at viewer, facing viewer, pink dress, purple eyes, holding book, best quality, amazing quality, very aesthetic, absurdres`` Negative / Undesired Content ``lowres, {bad}, fewer, extra, missing, worst quality, bad quality, jpeg artifacts, scan artifacts, unfinished, displeasing, chromatic aberration, artistic error, [abstract], @_@, heart-shaped pupils, lineart, sketch, nude, cleavage, panties, nipples, topless`` Settings ``steps: 28, height: 1216, width: 832, scale: 5.3, uncond_scale: 1.0, cfg_rescale: 0.0, seed: 3415329165, n_samples: 1, hide_debug_overlay: False, noise_schedule: native, sampler: k_euler_ancestral, controlnet_strength: 1.0, controlnet_model: None, dynamic_thresholding: False, dynamic_thresholding_percentile: 0.999, dynamic_thresholding_mimic_scale: 10.0, sm: False, sm_dyn: False, skip_cfg_below_sigma: 0.0, lora_unet_weights: None, lora_clip_weights: None``
nirantk/dbpedia-entities-efficient-splade-100K
--- dataset_info: features: - name: _id dtype: string - name: title dtype: string - name: text dtype: string - name: openai sequence: float32 - name: splade sequence: float32 splits: - name: train num_bytes: 12862697823 num_examples: 100000 download_size: 901410913 dataset_size: 12862697823 configs: - config_name: default data_files: - split: train path: data/train-* license: apache-2.0 task_categories: - feature-extraction language: - en pretty_name: 'DBPedia SPLADE + OpenAI: 100,000 Vectors' size_categories: - 100K<n<1M --- # DBPedia SPLADE + OpenAI: 100,000 SPLADE Sparse Vectors + OpenAI Embedding This dataset has both OpenAI and SPLADE vectors for 100,000 DBPedia entries. This adds SPLADE Vectors to [KShivendu/dbpedia-entities-openai-1M/](https://huggingface.co/datasets/KShivendu/dbpedia-entities-openai-1M) Model id used to make these vectors: ```python model_id = "naver/efficient-splade-VI-BT-large-doc" ``` For processing the query, use this: ```python model_id = "naver/efficient-splade-VI-BT-large-query" ``` If you'd like to extract the indices and weights/values from the vectors, you can do so using the following snippet: ```python import numpy as np vec = np.array(ds[0]['vec']) # where ds is the dataset def get_indices_values(vec): sparse_indices = vec.nonzero() sparse_values = vec[sparse_indices] return sparse_indices, sparse_values ```
hankang2023/Ultrafeedback_binarized.ko.hankang
--- license: mit --- 데이터셋 HuggingFaceH4/ultrafeedback_binarized를 한영 번역 모델인 squarelike/Gugugo-koen를 이용하여 번역함.\n\n https://huggingface.co/datasets/HuggingFaceH4/ultrafeedback_binarized https://huggingface.co/squarelike/Gugugo-koen-7B-V1.1-AWQ
p1atdev/ja-stackoverflow
--- dataset_info: - config_name: default features: - name: question struct: - name: accepted_answer_id dtype: string - name: answer_count dtype: int64 - name: body dtype: string - name: comment_count dtype: int64 - name: content_license dtype: string - name: creation_date dtype: string - name: favorite_count dtype: int64 - name: id dtype: string - name: last_activity_date dtype: string - name: last_edit_date dtype: string - name: last_editor_user_id dtype: string - name: owner_user_id dtype: string - name: post_type dtype: string - name: score dtype: int64 - name: tags sequence: string - name: title dtype: string - name: view_count dtype: int64 - name: answers list: - name: body dtype: string - name: comment_count dtype: int64 - name: content_license dtype: string - name: creation_date dtype: string - name: id dtype: string - name: last_activity_date dtype: string - name: last_edit_date dtype: string - name: last_editor_user_id dtype: string - name: owner_user_id dtype: string - name: parent_id dtype: string - name: post_type dtype: string - name: score dtype: int64 - name: id dtype: string - name: accepted_answer_id dtype: string - name: popular_answer_id dtype: string splits: - name: train num_bytes: 112596554 num_examples: 30551 download_size: 54805530 dataset_size: 112596554 - config_name: simple features: - name: id dtype: string - name: accepted_answer_id dtype: string - name: popular_answer_id dtype: string - name: title dtype: string - name: question_body dtype: string - name: question_score dtype: int64 - name: accepted_answer_body dtype: string - name: accepted_answer_score dtype: int64 - name: popular_answer_body dtype: string - name: popular_answer_score dtype: int64 - name: tags sequence: string splits: - name: train num_bytes: 113051344 num_examples: 30551 download_size: 56632072 dataset_size: 113051344 configs: - config_name: default data_files: - split: train path: data/train-* - config_name: simple data_files: - split: train path: simple/train-* license: cc-by-sa-4.0 task_categories: - text-generation - question-answering language: - ja tags: - stackoverflow - programming pretty_name: Japanese StackOverflow size_categories: - 10K<n<100K --- # ja-stackoverflow 日本語版 Stack Overflow の [スタック・オーバーフロー](https://ja.stackoverflow.com/) の[データダンプ](https://archive.org/download/stackexchange) をもとにデータを加工し、質問文と回答文のペアになるように調整した QA データセット。 ## データ構造 投稿本文は `html2text` を使ってマークダウン化されています。その際、 - コードブロックは \`\`\` で囲まれるように変更されています。 - 画像 URL に base64 エンコードされた画像が含まれる場合、 `[unk]` に置き換えています。 ### `default` サブセット - `id`: 質問投稿の ID - `question`: 質問投稿 - `answers`: 質問に対する回答投稿のリスト - `accepted_answer_id`: 質問者に選ばれた回答のID。`null` の可能性がある - `popular_answer_id`: もっともスコアが高かった回答のID。`null` の可能性がある ### `simple` サブセット `default` サブセットから、 `question` と `answers` の辞書を展開しシンプルにしたもの。 - `id`: 質問投稿の ID - `accepted_answer_id`: 質問者に選ばれた回答のID。`null` の可能性がある - `popular_answer_id`: もっともスコアが高かった回答のID。`null` の可能性がある - `title`: 質問のタイトル - `question_body`: 質問本文 - `question_score`: 質問のスコア - `tags`: 質問に関連付けられたタグ - `accepted_answer_body`: 質問者に選ばれた回答の本文。`null` の可能性がある - `accepted_answer_score`: 質問者に選ばれた回答のスコア。`null` の可能性がある - `popular_answer_body`: もっともスコアが高かった回答の本文。`null` の可能性がある - `popular_answer_score`: もっともスコアが高かった回答のスコア。`null` の可能性がある ## 使い方 datasets ライブラリを用いて簡単に利用できます。 ```py from datasets import load_dataset dataset = load_dataset("p1atdev/ja-stackoverflow", name="simple" split="train") print(dataset) #Dataset({ # features: ['id', 'accepted_answer_id', 'popular_answer_id', 'title', 'question_body', 'question_score', 'accepted_answer_body', 'accepted_answer_score', 'popular_answer_body', 'popular_answer_score', 'tags'], # num_rows: 30551 #}) ``` ## ライセンス StackOverflow に基づき、[CC BY-SA 4.0](https://creativecommons.org/licenses/by-sa/4.0/deed.ja)
heegyu/glaive-function-calling-v2-formatted
--- configs: - config_name: default data_files: - split: train path: data/train-* dataset_info: features: - name: system_message dtype: string - name: function_description dtype: string - name: conversations list: - name: content dtype: string - name: role dtype: string splits: - name: train num_bytes: 250214164 num_examples: 112960 download_size: 93753668 dataset_size: 250214164 --- - original dataset: [glaiveai/glaive-function-calling-v2](https://huggingface.co/datasets/glaiveai/glaive-function-calling-v2) ``` {'system_message': 'You are a helpful assistant with access to the following functions. Use them if required -', 'function_description': '{\n "name": "get_random_quote",\n "description": "Get a random quote",\n "parameters": {}\n}', 'conversations': [{'content': 'Hi, can you help me with something?', 'role': 'user'}, {'content': "Of course! I'm here to assist you. What do you need help with?", 'role': 'assistant'}, {'content': 'I need to book a flight from New York to Paris. Can you do that for me?', 'role': 'user'}, {'content': "I'm sorry, but as an AI, I don't have the capability to book flights or perform external tasks. My current function allows me to provide you with a random quote. Would you like to hear a quote instead?", 'role': 'assistant'}, {'content': 'Oh, I see. Sure, I would love to hear a random quote.', 'role': 'user'}, {'content': '{"name": "get_random_quote", "arguments": {}}', 'role': 'function-call'}, {'content': '{"quote": "The only way to do great work is to love what you do. - Steve Jobs"}', 'role': 'function-response'}, {'content': 'Here is a quote for you: "The only way to do great work is to love what you do." - Steve Jobs. I hope it inspires you!', 'role': 'assistant'}]} ```
Itau-Unibanco/FAQ_BACEN
--- license: apache-2.0 task_categories: - text-classification - question-answering language: - pt tags: - finance size_categories: - 1K<n<10K --- This dataset was used in the article: https://arxiv.org/abs/2311.11331
rohansolo/BB_HindiHinglishV2
--- dataset_info: features: - name: id dtype: string - name: messages list: - name: content dtype: string - name: role dtype: string - name: category dtype: string - name: __index_level_0__ dtype: int64 splits: - name: train_sft num_bytes: 533044539 num_examples: 199137 - name: test_sft num_bytes: 132486609 num_examples: 49785 download_size: 263949334 dataset_size: 665531148 configs: - config_name: default data_files: - split: train_sft path: data/train_sft-* - split: test_sft path: data/test_sft-* license: cc-by-nc-4.0 language: - hi - en --- Overview This dataset is a comprehensive collection of popular Hindi instruction-type datasets. It has been meticulously curated and merged into a unified format, making it ideal for use with Hugging Face's alignment notebook. The primary objective of creating this dataset is to offer a single, standardized resource for training models in understanding and generating Hindi and Hinglish (Hindi-English) conversations. Data Sources The dataset is an amalgamation of several individual datasets, each sourced from the Hugging Face datasets library. These include: FreedomIntelligence/evol-instruct-hindi (Train Split) NebulaByte/alpaca-gpt4-hindi-hinglish (Train Split) FreedomIntelligence/evol-instruct-hindi (Train Split, used twice in the script) smangrul/hindi_instruct_v1 (Train and Test Splits) SherryT997/HelpSteer-hindi (Train Split) Data Processing The datasets were processed using custom Python scripts. The process involved: Loading each dataset from Hugging Face. Applying specific conversion functions (convert_dataset1 and convert_dataset2) to standardize the format of the datasets. These functions were designed to handle different data formats and unify them under a common structure. Merging the converted datasets into a single Pandas DataFrame. Splitting the merged dataset into training and testing sets using a 80/20 split. Converting these splits back into Hugging Face Dataset format for ease of use in training and evaluation. Dataset Structure The final dataset is structured as follows: Each entry consists of a unique id and a series of messages. Each message contains content and a role (either 'user' or 'assistant') indicating the speaker. Purpose The dataset is intended for research and development in natural language processing, specifically for: Training models on Hindi and Hinglish conversation understanding. Enhancing conversational AI capabilities in Hindi and mixed-language contexts. Usage This dataset is particularly suited for use with Hugging Face's alignment notebook. It can be utilized for training language models that cater to Hindi-speaking users, offering a rich source of conversational data in both Hindi and Hinglish.
giux78/100k-sft-ready-ultrafeedback-ita
--- dataset_info: features: - name: prompt dtype: string - name: prompt_id dtype: string - name: messages list: - name: content dtype: string - name: role dtype: string splits: - name: train_sft num_bytes: 736148767 num_examples: 100000 - name: test_sft num_bytes: 73258856 num_examples: 10000 - name: train_gen num_bytes: 1347396812 num_examples: 256032 - name: test_gen num_bytes: 148276089 num_examples: 28304 download_size: 1238466176 dataset_size: 2305080524 configs: - config_name: default data_files: - split: train_sft path: data/train_sft-* - split: test_sft path: data/test_sft-* - split: train_gen path: data/train_gen-* - split: test_gen path: data/test_gen-* ---
ayymen/Weblate-Translations
--- configs: - config_name: en-lk data_files: en-lk.tsv - config_name: en-en-rAU data_files: en-en-rAU.tsv - config_name: en-hy-rAM data_files: en-hy-rAM.tsv - config_name: en-qt data_files: en-qt.tsv - config_name: en-se data_files: en-se.tsv - config_name: en-en_AU data_files: en-en_AU.tsv - config_name: en-in data_files: en-in.tsv - config_name: en_US-id data_files: en_US-id.tsv - config_name: en-ajp data_files: en-ajp.tsv - config_name: en-en_US_rude data_files: en-en_US_rude.tsv - config_name: en_GB-sw data_files: en_GB-sw.tsv - config_name: en_GB-tzm data_files: en_GB-tzm.tsv - config_name: dev-pt data_files: dev-pt.tsv - config_name: de-nb_NO data_files: de-nb_NO.tsv - config_name: en_devel-bn_BD data_files: en_devel-bn_BD.tsv - config_name: messages-fr data_files: messages-fr.tsv - config_name: en-de-CH data_files: en-de-CH.tsv - config_name: en-gu_IN data_files: en-gu_IN.tsv - config_name: en-be_BY data_files: en-be_BY.tsv - config_name: eo-sk data_files: eo-sk.tsv - config_name: en-brx data_files: en-brx.tsv - config_name: en-en_US data_files: en-en_US.tsv - config_name: en_GB-an data_files: en_GB-an.tsv - config_name: en-korean data_files: en-korean.tsv - config_name: en_GB-fr-FR data_files: en_GB-fr-FR.tsv - config_name: en_devel-si data_files: en_devel-si.tsv - config_name: en_US-sr_Cyrl data_files: en_US-sr_Cyrl.tsv - config_name: en-fr@formal data_files: en-fr@formal.tsv - config_name: en_devel-zh_tw data_files: en_devel-zh_tw.tsv - config_name: en-en_ud data_files: en-en_ud.tsv - config_name: en_GB-bi data_files: en_GB-bi.tsv - config_name: en-sq_AL data_files: en-sq_AL.tsv - config_name: en-README_zh-CN data_files: en-README_zh-CN.tsv - config_name: en_US-ml_IN data_files: en_US-ml_IN.tsv - config_name: nb_NO-nn data_files: nb_NO-nn.tsv - config_name: en_devel-es_419 data_files: en_devel-es_419.tsv - config_name: en-de-DE data_files: en-de-DE.tsv - config_name: en-dua data_files: en-dua.tsv - config_name: en-gu-rIN data_files: en-gu-rIN.tsv - config_name: en-ty data_files: en-ty.tsv - config_name: nl-pl data_files: nl-pl.tsv - config_name: en_US-bo data_files: en_US-bo.tsv - config_name: en_devel-ru_RU data_files: en_devel-ru_RU.tsv - config_name: en_GB-cy_GB data_files: en_GB-cy_GB.tsv - config_name: en_US-zh-TW data_files: en_US-zh-TW.tsv - config_name: en_US-zh-hk data_files: en_US-zh-hk.tsv - config_name: en-DE data_files: en-DE.tsv - config_name: en_US-lzh data_files: en_US-lzh.tsv - config_name: sv-sma data_files: sv-sma.tsv - config_name: en_GB-fi_FI data_files: en_GB-fi_FI.tsv - config_name: en_US-zu data_files: en_US-zu.tsv - config_name: en_devel-mr data_files: en_devel-mr.tsv - config_name: en_US-he-IL data_files: en_US-he-IL.tsv - config_name: en_GB-fur data_files: en_GB-fur.tsv - config_name: en-fr_CH data_files: en-fr_CH.tsv - config_name: en-en-CA data_files: en-en-CA.tsv - config_name: en-ro_MD data_files: en-ro_MD.tsv - config_name: en_US-yue_HK data_files: en_US-yue_HK.tsv - config_name: es-mr data_files: es-mr.tsv - config_name: en_GB-ace data_files: en_GB-ace.tsv - config_name: en_GB-lt data_files: en_GB-lt.tsv - config_name: en-es-rES data_files: en-es-rES.tsv - config_name: en-ksh data_files: en-ksh.tsv - config_name: en_GB-ti data_files: en_GB-ti.tsv - config_name: en-zh-rSG data_files: en-zh-rSG.tsv - config_name: en-ms_Arab data_files: en-ms_Arab.tsv - config_name: en-README_CZ data_files: en-README_CZ.tsv - config_name: en-ug-CN data_files: en-ug-CN.tsv - config_name: en-ar-rYE data_files: en-ar-rYE.tsv - config_name: en-pk data_files: en-pk.tsv - config_name: en_US-pt data_files: en_US-pt.tsv - config_name: en_devel-pt-br data_files: en_devel-pt-br.tsv - config_name: en-de_formal data_files: en-de_formal.tsv - config_name: en-zh_TW data_files: en-zh_TW.tsv - config_name: en-hu-rHU data_files: en-hu-rHU.tsv - config_name: en-lv-LV data_files: en-lv-LV.tsv - config_name: en-hr_HR data_files: en-hr_HR.tsv - config_name: en-en_devel data_files: en-en_devel.tsv - config_name: en-ka data_files: en-ka.tsv - config_name: en_GB-da_DK data_files: en_GB-da_DK.tsv - config_name: en-ar-AR data_files: en-ar-AR.tsv - config_name: en-om data_files: en-om.tsv - config_name: en_US-id-ID data_files: en_US-id-ID.tsv - config_name: en-cs_CZ data_files: en-cs_CZ.tsv - config_name: it-es_ES data_files: it-es_ES.tsv - config_name: en-zh_HK data_files: en-zh_HK.tsv - config_name: dev-ko data_files: dev-ko.tsv - config_name: en-cr data_files: en-cr.tsv - config_name: en-sr_Cyrl data_files: en-sr_Cyrl.tsv - config_name: en-nl_BE data_files: en-nl_BE.tsv - config_name: en_GB-zh-rTW data_files: en_GB-zh-rTW.tsv - config_name: en-da-DK data_files: en-da-DK.tsv - config_name: en-ang data_files: en-ang.tsv - config_name: en-ur-IN data_files: en-ur-IN.tsv - config_name: en-HU data_files: en-HU.tsv - config_name: en-kw data_files: en-kw.tsv - config_name: en_GB-fo data_files: en_GB-fo.tsv - config_name: en-sr-SP data_files: en-sr-SP.tsv - config_name: en-pl data_files: en-pl.tsv - config_name: en-or data_files: en-or.tsv - config_name: en-en-gb data_files: en-en-gb.tsv - config_name: en-en data_files: en-en.tsv - config_name: en_GB-fa_IR data_files: en_GB-fa_IR.tsv - config_name: en-bn-IN data_files: en-bn-IN.tsv - config_name: en-pl_pl data_files: en-pl_pl.tsv - config_name: en_US-ro_RO data_files: en_US-ro_RO.tsv - config_name: en-es_mx data_files: en-es_mx.tsv - config_name: en-kk_KZ data_files: en-kk_KZ.tsv - config_name: en-ab data_files: en-ab.tsv - config_name: en_UK-de_DE data_files: en_UK-de_DE.tsv - config_name: eo-de data_files: eo-de.tsv - config_name: en_US-fil data_files: en_US-fil.tsv - config_name: en-bp data_files: en-bp.tsv - config_name: en-ta_IN data_files: en-ta_IN.tsv - config_name: en-round data_files: en-round.tsv - config_name: en-gd data_files: en-gd.tsv - config_name: en_US-en@uwu data_files: en_US-en@uwu.tsv - config_name: en-dum data_files: en-dum.tsv - config_name: en-ja_JP data_files: en-ja_JP.tsv - config_name: en-ryu data_files: en-ryu.tsv - config_name: en-b+en+001 data_files: en-b+en+001.tsv - config_name: en-en-US data_files: en-en-US.tsv - config_name: en-sl_SI data_files: en-sl_SI.tsv - config_name: de-it data_files: de-it.tsv - config_name: en_GB-sr_RS data_files: en_GB-sr_RS.tsv - config_name: en_US-da data_files: en_US-da.tsv - config_name: en_GB-tk data_files: en_GB-tk.tsv - config_name: en-bn data_files: en-bn.tsv - config_name: en_devel-es_bo data_files: en_devel-es_bo.tsv - config_name: en-ja_CARES data_files: en-ja_CARES.tsv - config_name: en-km-KH data_files: en-km-KH.tsv - config_name: en_US-de_DE data_files: en_US-de_DE.tsv - config_name: en_US-hu_HU data_files: en_US-hu_HU.tsv - config_name: en-ta-rIN data_files: en-ta-rIN.tsv - config_name: en_US-ml data_files: en_US-ml.tsv - config_name: en-sr_RS data_files: en-sr_RS.tsv - config_name: en_US-eu data_files: en_US-eu.tsv - config_name: pl-es data_files: pl-es.tsv - config_name: en_US-ka data_files: en_US-ka.tsv - config_name: en-bulgarian data_files: en-bulgarian.tsv - config_name: fr-en data_files: fr-en.tsv - config_name: en_devel-nb-rNO data_files: en_devel-nb-rNO.tsv - config_name: en_GB-ce data_files: en_GB-ce.tsv - config_name: en_US-bs data_files: en_US-bs.tsv - config_name: en-en@uwu data_files: en-en@uwu.tsv - config_name: en_GB-nn data_files: en_GB-nn.tsv - config_name: en-pa_PK data_files: en-pa_PK.tsv - config_name: en-wae data_files: en-wae.tsv - config_name: en-ar_EG data_files: en-ar_EG.tsv - config_name: en_GB-lt_LT data_files: en_GB-lt_LT.tsv - config_name: en-zh-Hant-HK data_files: en-zh-Hant-HK.tsv - config_name: messages-de data_files: messages-de.tsv - config_name: en-ur_IN data_files: en-ur_IN.tsv - config_name: en-in-rID data_files: en-in-rID.tsv - config_name: en-lo-LA data_files: en-lo-LA.tsv - config_name: en-el-rGR data_files: en-el-rGR.tsv - config_name: en-es-ES data_files: en-es-ES.tsv - config_name: en_devel-et data_files: en_devel-et.tsv - config_name: en-fr-rCH data_files: en-fr-rCH.tsv - config_name: en-en_CA data_files: en-en_CA.tsv - config_name: en-b+uz+Latn data_files: en-b+uz+Latn.tsv - config_name: en_GB-tig data_files: en_GB-tig.tsv - config_name: en_GB-hi_IN data_files: en_GB-hi_IN.tsv - config_name: de-pl data_files: de-pl.tsv - config_name: en-zh-rCN data_files: en-zh-rCN.tsv - config_name: en-hi-rIN data_files: en-hi-rIN.tsv - config_name: en-ba data_files: en-ba.tsv - config_name: en-fy data_files: en-fy.tsv - config_name: en-el-GR data_files: en-el-GR.tsv - config_name: en-tum data_files: en-tum.tsv - config_name: en-ru-RU data_files: en-ru-RU.tsv - config_name: en_US-fa data_files: en_US-fa.tsv - config_name: en_GB-ka data_files: en_GB-ka.tsv - config_name: es-nb-rNO data_files: es-nb-rNO.tsv - config_name: en_US-ckb data_files: en_US-ckb.tsv - config_name: en-hi_IN data_files: en-hi_IN.tsv - config_name: eo-pa data_files: eo-pa.tsv - config_name: en_devel-zh_TW data_files: en_devel-zh_TW.tsv - config_name: en_GB-ch data_files: en_GB-ch.tsv - config_name: en-sdh data_files: en-sdh.tsv - config_name: en-lzh data_files: en-lzh.tsv - config_name: en-zh_HANS-CN data_files: en-zh_HANS-CN.tsv - config_name: en-li data_files: en-li.tsv - config_name: en_devel-zh_cn data_files: en_devel-zh_cn.tsv - config_name: en_GB-mk data_files: en_GB-mk.tsv - config_name: en_GB-ay data_files: en_GB-ay.tsv - config_name: en-sq-rAL data_files: en-sq-rAL.tsv - config_name: en-nl_TND data_files: en-nl_TND.tsv - config_name: en-th data_files: en-th.tsv - config_name: messages-id data_files: messages-id.tsv - config_name: en-bo data_files: en-bo.tsv - config_name: en-hy data_files: en-hy.tsv - config_name: en_US-gd data_files: en_US-gd.tsv - config_name: en-tok data_files: en-tok.tsv - config_name: pt_BR-en data_files: pt_BR-en.tsv - config_name: fr-pt data_files: fr-pt.tsv - config_name: en-bs-rBA data_files: en-bs-rBA.tsv - config_name: en-zh-hant data_files: en-zh-hant.tsv - config_name: en_US-fr data_files: en_US-fr.tsv - config_name: en-eu-ES data_files: en-eu-ES.tsv - config_name: en-lv_LV data_files: en-lv_LV.tsv - config_name: und-fr data_files: und-fr.tsv - config_name: en-af-rZA data_files: en-af-rZA.tsv - config_name: en-da data_files: en-da.tsv - config_name: en-os data_files: en-os.tsv - config_name: en-fr-CH data_files: en-fr-CH.tsv - config_name: en-es_MX data_files: en-es_MX.tsv - config_name: nl-bg data_files: nl-bg.tsv - config_name: en_GB-ckb data_files: en_GB-ckb.tsv - config_name: en-ar-rEG data_files: en-ar-rEG.tsv - config_name: en_US-mr data_files: en_US-mr.tsv - config_name: en_US-cs-CZ data_files: en_US-cs-CZ.tsv - config_name: en_devel-fi data_files: en_devel-fi.tsv - config_name: en-mhr data_files: en-mhr.tsv - config_name: en-no-rNO data_files: en-no-rNO.tsv - config_name: en-it_it data_files: en-it_it.tsv - config_name: en-ar-rSA data_files: en-ar-rSA.tsv - config_name: en_GB-nso data_files: en_GB-nso.tsv - config_name: en-ti data_files: en-ti.tsv - config_name: en-iw_HE data_files: en-iw_HE.tsv - config_name: en-szl data_files: en-szl.tsv - config_name: en_GB-ba data_files: en_GB-ba.tsv - config_name: en_devel-cs data_files: en_devel-cs.tsv - config_name: en_GB-pl_PL data_files: en_GB-pl_PL.tsv - config_name: en-ta_LK data_files: en-ta_LK.tsv - config_name: en-uz@latin data_files: en-uz@latin.tsv - config_name: en-el data_files: en-el.tsv - config_name: en_GB-cs data_files: en_GB-cs.tsv - config_name: en-bul_BG data_files: en-bul_BG.tsv - config_name: en-fa_IR data_files: en-fa_IR.tsv - config_name: en-gsw data_files: en-gsw.tsv - config_name: en-ko-KR data_files: en-ko-KR.tsv - config_name: en-bs_BA data_files: en-bs_BA.tsv - config_name: en_GB-wo data_files: en_GB-wo.tsv - config_name: en_devel-it data_files: en_devel-it.tsv - config_name: en_US-bn data_files: en_US-bn.tsv - config_name: en_devel-pl data_files: en_devel-pl.tsv - config_name: en-rm data_files: en-rm.tsv - config_name: en-night data_files: en-night.tsv - config_name: eo-ca data_files: eo-ca.tsv - config_name: en_US-ps data_files: en_US-ps.tsv - config_name: en_GB-sd data_files: en_GB-sd.tsv - config_name: en-th-TH data_files: en-th-TH.tsv - config_name: en-sv-rSE data_files: en-sv-rSE.tsv - config_name: en-b+zh+Hans data_files: en-b+zh+Hans.tsv - config_name: en_devel-uk data_files: en_devel-uk.tsv - config_name: en_US-it_IT data_files: en_US-it_IT.tsv - config_name: en-b+hrx data_files: en-b+hrx.tsv - config_name: en-my data_files: en-my.tsv - config_name: en_GB-sc data_files: en_GB-sc.tsv - config_name: en-de_DE_rude data_files: en-de_DE_rude.tsv - config_name: en_GB-ff data_files: en_GB-ff.tsv - config_name: en_devel-nl data_files: en_devel-nl.tsv - config_name: en-shn data_files: en-shn.tsv - config_name: en_GB-ca data_files: en_GB-ca.tsv - config_name: en-hu_HU data_files: en-hu_HU.tsv - config_name: ru-be data_files: ru-be.tsv - config_name: es-ml data_files: es-ml.tsv - config_name: en_GB-na data_files: en_GB-na.tsv - config_name: en_devel-ja data_files: en_devel-ja.tsv - config_name: en-pt-rPT-v26 data_files: en-pt-rPT-v26.tsv - config_name: en_devel-pt_BR data_files: en_devel-pt_BR.tsv - config_name: en_US-ar_AA data_files: en_US-ar_AA.tsv - config_name: en_US-en_GB data_files: en_US-en_GB.tsv - config_name: en-de_FORM data_files: en-de_FORM.tsv - config_name: en_US-et data_files: en_US-et.tsv - config_name: pl-it data_files: pl-it.tsv - config_name: messages-ru data_files: messages-ru.tsv - config_name: en_devel-en data_files: en_devel-en.tsv - config_name: en-te_IN data_files: en-te_IN.tsv - config_name: en_US-it-IT data_files: en_US-it-IT.tsv - config_name: en-zh-rMO data_files: en-zh-rMO.tsv - config_name: en-fy-NL data_files: en-fy-NL.tsv - config_name: en-iw-rIL data_files: en-iw-rIL.tsv - config_name: en-zh-Hant data_files: en-zh-Hant.tsv - config_name: en-es_uy data_files: en-es_uy.tsv - config_name: en_GB-or data_files: en_GB-or.tsv - config_name: en-tt data_files: en-tt.tsv - config_name: de-pt data_files: de-pt.tsv - config_name: en-zh-Hans data_files: en-zh-Hans.tsv - config_name: en-ar-TN data_files: en-ar-TN.tsv - config_name: en_US-si_LK data_files: en_US-si_LK.tsv - config_name: en-so data_files: en-so.tsv - config_name: en_GB-csb data_files: en_GB-csb.tsv - config_name: en-fr-CA data_files: en-fr-CA.tsv - config_name: en-es_BO data_files: en-es_BO.tsv - config_name: en_devel-es_pa data_files: en_devel-es_pa.tsv - config_name: en-vi-VN data_files: en-vi-VN.tsv - config_name: en_devel-sw data_files: en_devel-sw.tsv - config_name: en-es-rMX data_files: en-es-rMX.tsv - config_name: en-eu-rES data_files: en-eu-rES.tsv - config_name: en_GB-pi data_files: en_GB-pi.tsv - config_name: en_devel-bg data_files: en_devel-bg.tsv - config_name: en-ja-JP data_files: en-ja-JP.tsv - config_name: en_US-uk data_files: en_US-uk.tsv - config_name: en_GB-km data_files: en_GB-km.tsv - config_name: en_US-ko data_files: en_US-ko.tsv - config_name: en-gmh data_files: en-gmh.tsv - config_name: en_US-hy data_files: en_US-hy.tsv - config_name: en_GB-ml data_files: en_GB-ml.tsv - config_name: en-bn-rIN data_files: en-bn-rIN.tsv - config_name: en-ach data_files: en-ach.tsv - config_name: en-pt-rBR-v26 data_files: en-pt-rBR-v26.tsv - config_name: en_US-zh data_files: en_US-zh.tsv - config_name: en-sw-rKE data_files: en-sw-rKE.tsv - config_name: en_GB-ha data_files: en_GB-ha.tsv - config_name: en-en-rGB data_files: en-en-rGB.tsv - config_name: en_devel-pt data_files: en_devel-pt.tsv - config_name: en-no_NB data_files: en-no_NB.tsv - config_name: en-no_NO data_files: en-no_NO.tsv - config_name: en-es_es data_files: en-es_es.tsv - config_name: en-kk data_files: en-kk.tsv - config_name: en-bm data_files: en-bm.tsv - config_name: en-pl-PL data_files: en-pl-PL.tsv - config_name: en_GB-id data_files: en_GB-id.tsv - config_name: en-sr-Latn data_files: en-sr-Latn.tsv - config_name: en_US-ms data_files: en_US-ms.tsv - config_name: en-et_ET data_files: en-et_ET.tsv - config_name: en-b+es+419 data_files: en-b+es+419.tsv - config_name: en_GB-kw data_files: en_GB-kw.tsv - config_name: en-no data_files: en-no.tsv - config_name: en-wa data_files: en-wa.tsv - config_name: en-ber data_files: en-ber.tsv - config_name: en_US-es_MX data_files: en_US-es_MX.tsv - config_name: en-de_1901 data_files: en-de_1901.tsv - config_name: en-ja-rJP data_files: en-ja-rJP.tsv - config_name: en_US-uk_UA data_files: en_US-uk_UA.tsv - config_name: en_US-ja_JP data_files: en_US-ja_JP.tsv - config_name: en-b+fr data_files: en-b+fr.tsv - config_name: en-pt-br data_files: en-pt-br.tsv - config_name: en-te data_files: en-te.tsv - config_name: en-np data_files: en-np.tsv - config_name: en_GB-gu data_files: en_GB-gu.tsv - config_name: en_GB-ki data_files: en_GB-ki.tsv - config_name: en-kab-KAB data_files: en-kab-KAB.tsv - config_name: de-fr data_files: de-fr.tsv - config_name: en-ru_old data_files: en-ru_old.tsv - config_name: en_devel-es_do data_files: en_devel-es_do.tsv - config_name: en-ua data_files: en-ua.tsv - config_name: en-et_EE data_files: en-et_EE.tsv - config_name: ia-it data_files: ia-it.tsv - config_name: en_GB-ro data_files: en_GB-ro.tsv - config_name: en_US-pt-rPT data_files: en_US-pt-rPT.tsv - config_name: en-ur_PK data_files: en-ur_PK.tsv - config_name: en-pa-rPK data_files: en-pa-rPK.tsv - config_name: en-vec data_files: en-vec.tsv - config_name: en-nl-rBE data_files: en-nl-rBE.tsv - config_name: en-lv data_files: en-lv.tsv - config_name: en-ar-rBH data_files: en-ar-rBH.tsv - config_name: en-an data_files: en-an.tsv - config_name: en_US-sr data_files: en_US-sr.tsv - config_name: en-Ukrainian data_files: en-Ukrainian.tsv - config_name: en_US-mk data_files: en_US-mk.tsv - config_name: en_GB-br data_files: en_GB-br.tsv - config_name: en-de@informal data_files: en-de@informal.tsv - config_name: en-dz data_files: en-dz.tsv - config_name: en_US-he_IL data_files: en_US-he_IL.tsv - config_name: en_GB-mr data_files: en_GB-mr.tsv - config_name: en-cs-CARES data_files: en-cs-CARES.tsv - config_name: en_US-hi_IN data_files: en_US-hi_IN.tsv - config_name: en_US-ro data_files: en_US-ro.tsv - config_name: en_US-fr_CA data_files: en_US-fr_CA.tsv - config_name: en-as data_files: en-as.tsv - config_name: en_GB-ro_MD data_files: en_GB-ro_MD.tsv - config_name: en_US-lt-LT data_files: en_US-lt-LT.tsv - config_name: fr-ca data_files: fr-ca.tsv - config_name: en-be_Latn data_files: en-be_Latn.tsv - config_name: en-en-AU data_files: en-en-AU.tsv - config_name: en_US-fr_FR data_files: en_US-fr_FR.tsv - config_name: en-de-de data_files: en-de-de.tsv - config_name: en-nds data_files: en-nds.tsv - config_name: en_US-ja data_files: en_US-ja.tsv - config_name: en-es-AR data_files: en-es-AR.tsv - config_name: en-ms data_files: en-ms.tsv - config_name: en-zh-CHS data_files: en-zh-CHS.tsv - config_name: en_devel-bs data_files: en_devel-bs.tsv - config_name: en-arn data_files: en-arn.tsv - config_name: zh_Hans-en data_files: zh_Hans-en.tsv - config_name: en-co data_files: en-co.tsv - config_name: en-uz_Latn data_files: en-uz_Latn.tsv - config_name: en-cs-rCZ data_files: en-cs-rCZ.tsv - config_name: en-ku data_files: en-ku.tsv - config_name: en-ha data_files: en-ha.tsv - config_name: en-de-zuerich-lernt data_files: en-de-zuerich-lernt.tsv - config_name: en_US-be data_files: en_US-be.tsv - config_name: en-tr data_files: en-tr.tsv - config_name: en-ru_ru data_files: en-ru_ru.tsv - config_name: en-kl data_files: en-kl.tsv - config_name: en-it data_files: en-it.tsv - config_name: en-b+be+Latn data_files: en-b+be+Latn.tsv - config_name: en_devel-mk data_files: en_devel-mk.tsv - config_name: en_US-vi data_files: en_US-vi.tsv - config_name: en-zh_CMN-HANT data_files: en-zh_CMN-HANT.tsv - config_name: en-mnw data_files: en-mnw.tsv - config_name: en_US-sv-SE data_files: en_US-sv-SE.tsv - config_name: en-gum data_files: en-gum.tsv - config_name: en-my_MM data_files: en-my_MM.tsv - config_name: en_GB-mk_MK data_files: en_GB-mk_MK.tsv - config_name: en_devel-es_ec data_files: en_devel-es_ec.tsv - config_name: en_US-ne data_files: en_US-ne.tsv - config_name: nl-zh_Hans data_files: nl-zh_Hans.tsv - config_name: en-zh_hans data_files: en-zh_hans.tsv - config_name: en-sr-rCS data_files: en-sr-rCS.tsv - config_name: en-es_NI data_files: en-es_NI.tsv - config_name: en_GB-bs data_files: en_GB-bs.tsv - config_name: en_GB-tr_TR data_files: en_GB-tr_TR.tsv - config_name: ru-en data_files: ru-en.tsv - config_name: en_US-my data_files: en_US-my.tsv - config_name: en-ia data_files: en-ia.tsv - config_name: en-hu-HU data_files: en-hu-HU.tsv - config_name: en-nn_NO data_files: en-nn_NO.tsv - config_name: en_GB-es_419 data_files: en_GB-es_419.tsv - config_name: en-ca-rES data_files: en-ca-rES.tsv - config_name: en_US-zh-CN data_files: en_US-zh-CN.tsv - config_name: en_US-tzm data_files: en_US-tzm.tsv - config_name: en-it_CARES data_files: en-it_CARES.tsv - config_name: en_GB-he data_files: en_GB-he.tsv - config_name: en_US-sn data_files: en_US-sn.tsv - config_name: en-ml_IN data_files: en-ml_IN.tsv - config_name: en-guc data_files: en-guc.tsv - config_name: zh_Hans-ru data_files: zh_Hans-ru.tsv - config_name: en-csb data_files: en-csb.tsv - config_name: en-nan data_files: en-nan.tsv - config_name: en-fa-IR data_files: en-fa-IR.tsv - config_name: en_US-en_CA data_files: en_US-en_CA.tsv - config_name: en_GB-ar data_files: en_GB-ar.tsv - config_name: en_GB-ia_FR data_files: en_GB-ia_FR.tsv - config_name: en_US-es-MX data_files: en_US-es-MX.tsv - config_name: en_devel-el data_files: en_devel-el.tsv - config_name: en_GB-ach data_files: en_GB-ach.tsv - config_name: en-Italian data_files: en-Italian.tsv - config_name: en_devel-az data_files: en_devel-az.tsv - config_name: eo-ru data_files: eo-ru.tsv - config_name: en-es_US data_files: en-es_US.tsv - config_name: en_devel-cy data_files: en_devel-cy.tsv - config_name: en-es-mx data_files: en-es-mx.tsv - config_name: en-en-rCA data_files: en-en-rCA.tsv - config_name: en-kn-IN data_files: en-kn-IN.tsv - config_name: en_devel-zh_CN data_files: en_devel-zh_CN.tsv - config_name: en_US-lt_LT data_files: en_US-lt_LT.tsv - config_name: en_GB-id_ID data_files: en_GB-id_ID.tsv - config_name: en-mt data_files: en-mt.tsv - config_name: en-bar data_files: en-bar.tsv - config_name: en-kr data_files: en-kr.tsv - config_name: en_GB-de-DE data_files: en_GB-de-DE.tsv - config_name: en-zgh data_files: en-zgh.tsv default: true - config_name: en-german data_files: en-german.tsv - config_name: en-de_ch data_files: en-de_ch.tsv - config_name: en_devel-hy data_files: en_devel-hy.tsv - config_name: en_GB-hr data_files: en_GB-hr.tsv - config_name: en_GB-ca_AD data_files: en_GB-ca_AD.tsv - config_name: en-b+ca+VALENCIA data_files: en-b+ca+VALENCIA.tsv - config_name: en-rw data_files: en-rw.tsv - config_name: en-fil-FIL data_files: en-fil-FIL.tsv - config_name: it-de data_files: it-de.tsv - config_name: en_US-es-rMX data_files: en_US-es-rMX.tsv - config_name: en-sk-SK data_files: en-sk-SK.tsv - config_name: en-my-MM data_files: en-my-MM.tsv - config_name: en-es_ve data_files: en-es_ve.tsv - config_name: en-fra-rFR data_files: en-fra-rFR.tsv - config_name: en_GB-gv data_files: en_GB-gv.tsv - config_name: en-ml-IN data_files: en-ml-IN.tsv - config_name: en_US-zh-rHK data_files: en_US-zh-rHK.tsv - config_name: en-fur data_files: en-fur.tsv - config_name: en_GB-sv data_files: en_GB-sv.tsv - config_name: en-ne-rNP data_files: en-ne-rNP.tsv - config_name: en_GB-fr data_files: en_GB-fr.tsv - config_name: en_US-qya data_files: en_US-qya.tsv - config_name: en-ja_KS data_files: en-ja_KS.tsv - config_name: en-en_uwu_x data_files: en-en_uwu_x.tsv - config_name: en-zh_CN data_files: en-zh_CN.tsv - config_name: en-az_AZ data_files: en-az_AZ.tsv - config_name: en-bem data_files: en-bem.tsv - config_name: en-ars data_files: en-ars.tsv - config_name: en-xh data_files: en-xh.tsv - config_name: en_US-zh_Hant_HK data_files: en_US-zh_Hant_HK.tsv - config_name: en_US-en-rGB data_files: en_US-en-rGB.tsv - config_name: en-pam data_files: en-pam.tsv - config_name: en_devel-zh-rCN data_files: en_devel-zh-rCN.tsv - config_name: en-zh_LATN@pinyin data_files: en-zh_LATN@pinyin.tsv - config_name: en_US-en_NZ data_files: en_US-en_NZ.tsv - config_name: en-nb_no data_files: en-nb_no.tsv - config_name: en-bn-rBD data_files: en-bn-rBD.tsv - config_name: en-pl_PL data_files: en-pl_PL.tsv - config_name: en-romanian data_files: en-romanian.tsv - config_name: en_US-ja_KANJI data_files: en_US-ja_KANJI.tsv - config_name: en_US-zh-rCN data_files: en_US-zh-rCN.tsv - config_name: en-ca_es data_files: en-ca_es.tsv - config_name: en-de_de data_files: en-de_de.tsv - config_name: en-rom data_files: en-rom.tsv - config_name: en_devel-lv data_files: en_devel-lv.tsv - config_name: en-ro data_files: en-ro.tsv - config_name: en_US-th-TH data_files: en_US-th-TH.tsv - config_name: en_GB-wal data_files: en_GB-wal.tsv - config_name: en_US-fi-FI data_files: en_US-fi-FI.tsv - config_name: en-ar_AR data_files: en-ar_AR.tsv - config_name: en_US-el data_files: en_US-el.tsv - config_name: en_GB-chr data_files: en_GB-chr.tsv - config_name: en-pbb data_files: en-pbb.tsv - config_name: en-ar-rXB data_files: en-ar-rXB.tsv - config_name: en-tzm data_files: en-tzm.tsv - config_name: en-mr-rIN data_files: en-mr-rIN.tsv - config_name: en-ms-rMY data_files: en-ms-rMY.tsv - config_name: en-apc data_files: en-apc.tsv - config_name: en_GB-fi data_files: en_GB-fi.tsv - config_name: en_US-hi data_files: en_US-hi.tsv - config_name: en-hz data_files: en-hz.tsv - config_name: en_GB-mi data_files: en_GB-mi.tsv - config_name: en-sai data_files: en-sai.tsv - config_name: en-ig data_files: en-ig.tsv - config_name: en-en_Shaw data_files: en-en_Shaw.tsv - config_name: en_US-fa_IR data_files: en_US-fa_IR.tsv - config_name: en-mr data_files: en-mr.tsv - config_name: en-pl_PL_rude data_files: en-pl_PL_rude.tsv - config_name: en-cv data_files: en-cv.tsv - config_name: messages-ar data_files: messages-ar.tsv - config_name: en-ko_KO data_files: en-ko_KO.tsv - config_name: en_US-zh-hans data_files: en_US-zh-hans.tsv - config_name: en-ga-IE data_files: en-ga-IE.tsv - config_name: en-am data_files: en-am.tsv - config_name: en-ug data_files: en-ug.tsv - config_name: en-af_ZA data_files: en-af_ZA.tsv - config_name: en-ES data_files: en-ES.tsv - config_name: en_US-ru_RU data_files: en_US-ru_RU.tsv - config_name: en_GB-lv data_files: en_GB-lv.tsv - config_name: en-yi data_files: en-yi.tsv - config_name: en_GB-pl data_files: en_GB-pl.tsv - config_name: en_GB-tl data_files: en_GB-tl.tsv - config_name: en-km data_files: en-km.tsv - config_name: en-azb data_files: en-azb.tsv - config_name: en_devel-fr data_files: en_devel-fr.tsv - config_name: en-pa-PK data_files: en-pa-PK.tsv - config_name: en-tn data_files: en-tn.tsv - config_name: en-mjw data_files: en-mjw.tsv - config_name: en-frs data_files: en-frs.tsv - config_name: en-it-IT data_files: en-it-IT.tsv - config_name: en-ro_RO data_files: en-ro_RO.tsv - config_name: en_US-nl_NL data_files: en_US-nl_NL.tsv - config_name: en-ht data_files: en-ht.tsv - config_name: en_devel-es_cr data_files: en_devel-es_cr.tsv - config_name: en_US-zh-rTW data_files: en_US-zh-rTW.tsv - config_name: en-fo data_files: en-fo.tsv - config_name: en-skr data_files: en-skr.tsv - config_name: en-ak data_files: en-ak.tsv - config_name: en_GB-sr@latin data_files: en_GB-sr@latin.tsv - config_name: en_US-de_CH data_files: en_US-de_CH.tsv - config_name: en_US-uk-UA data_files: en_US-uk-UA.tsv - config_name: en-ko_KR data_files: en-ko_KR.tsv - config_name: en-cy data_files: en-cy.tsv - config_name: en-galo data_files: en-galo.tsv - config_name: en-bn_BD data_files: en-bn_BD.tsv - config_name: en_devel-ms data_files: en_devel-ms.tsv - config_name: fr-it data_files: fr-it.tsv - config_name: en-ny data_files: en-ny.tsv - config_name: en-tet data_files: en-tet.tsv - config_name: en_GB-sk data_files: en_GB-sk.tsv - config_name: eo-ar data_files: eo-ar.tsv - config_name: eo-es data_files: eo-es.tsv - config_name: en-bho data_files: en-bho.tsv - config_name: en-pap data_files: en-pap.tsv - config_name: en-vi_VN data_files: en-vi_VN.tsv - config_name: en_US-ar data_files: en_US-ar.tsv - config_name: en_devel-nb data_files: en_devel-nb.tsv - config_name: en_devel-es_mx data_files: en_devel-es_mx.tsv - config_name: es-ca data_files: es-ca.tsv - config_name: en_GB-kn data_files: en_GB-kn.tsv - config_name: en-ru_UA data_files: en-ru_UA.tsv - config_name: sv-nb data_files: sv-nb.tsv - config_name: en_GB-zh_Hans data_files: en_GB-zh_Hans.tsv - config_name: en-he-IL data_files: en-he-IL.tsv - config_name: en_GB-et data_files: en_GB-et.tsv - config_name: es-pl data_files: es-pl.tsv - config_name: en-hy-AM data_files: en-hy-AM.tsv - config_name: en_US-cy data_files: en_US-cy.tsv - config_name: en-hu-rZZ data_files: en-hu-rZZ.tsv - config_name: en-by data_files: en-by.tsv - config_name: en_GB-hy data_files: en_GB-hy.tsv - config_name: en_US-zh-Hant data_files: en_US-zh-Hant.tsv - config_name: en-gu-IN data_files: en-gu-IN.tsv - config_name: en_GB-ml_IN data_files: en_GB-ml_IN.tsv - config_name: de-nl data_files: de-nl.tsv - config_name: en_devel-ur data_files: en_devel-ur.tsv - config_name: en-ca-ES data_files: en-ca-ES.tsv - config_name: en_GB-kl data_files: en_GB-kl.tsv - config_name: en_US-ta_IN data_files: en_US-ta_IN.tsv - config_name: en_US-sk_SK data_files: en_US-sk_SK.tsv - config_name: en-zh_Latn data_files: en-zh_Latn.tsv - config_name: en_GB-es data_files: en_GB-es.tsv - config_name: en-en_uk data_files: en-en_uk.tsv - config_name: en_GB-ru data_files: en_GB-ru.tsv - config_name: en-gu data_files: en-gu.tsv - config_name: en_US-km data_files: en_US-km.tsv - config_name: en_GB-uz data_files: en_GB-uz.tsv - config_name: en_US-yue-HK data_files: en_US-yue-HK.tsv - config_name: en-ceb data_files: en-ceb.tsv - config_name: en-is data_files: en-is.tsv - config_name: en-ug@Arab data_files: en-ug@Arab.tsv - config_name: es-ru data_files: es-ru.tsv - config_name: en-pt data_files: en-pt.tsv - config_name: en-es-US data_files: en-es-US.tsv - config_name: en-zh-rCMN-HANT data_files: en-zh-rCMN-HANT.tsv - config_name: en-jbo-EN data_files: en-jbo-EN.tsv - config_name: en_US-pa data_files: en_US-pa.tsv - config_name: en_US-or data_files: en_US-or.tsv - config_name: dev-hu data_files: dev-hu.tsv - config_name: en-b+ast data_files: en-b+ast.tsv - config_name: messages-vi data_files: messages-vi.tsv - config_name: en-ht-HT data_files: en-ht-HT.tsv - config_name: en-ar_AA data_files: en-ar_AA.tsv - config_name: en-mcc234 data_files: en-mcc234.tsv - config_name: en_GB-he_IL data_files: en_GB-he_IL.tsv - config_name: en-fr_FR data_files: en-fr_FR.tsv - config_name: en-es_ES data_files: en-es_ES.tsv - config_name: en-tr-v26 data_files: en-tr-v26.tsv - config_name: ru-kk data_files: ru-kk.tsv - config_name: en_GB-ky data_files: en_GB-ky.tsv - config_name: en-st data_files: en-st.tsv - config_name: en-ky data_files: en-ky.tsv - config_name: en_GB-fa data_files: en_GB-fa.tsv - config_name: en-ta data_files: en-ta.tsv - config_name: en_US-ru-RU data_files: en_US-ru-RU.tsv - config_name: en_US-it data_files: en_US-it.tsv - config_name: en-mai data_files: en-mai.tsv - config_name: en_GB-ga data_files: en_GB-ga.tsv - config_name: en-ay data_files: en-ay.tsv - config_name: en-pt_PT data_files: en-pt_PT.tsv - config_name: en-fa-rIR data_files: en-fa-rIR.tsv - config_name: en-sk_SK data_files: en-sk_SK.tsv - config_name: en-ru_sov data_files: en-ru_sov.tsv - config_name: en-pt-PT data_files: en-pt-PT.tsv - config_name: en_US-ko-KR data_files: en_US-ko-KR.tsv - config_name: en-es-rCO data_files: en-es-rCO.tsv - config_name: en-zh data_files: en-zh.tsv - config_name: en_US-ber data_files: en_US-ber.tsv - config_name: en-en_NZ data_files: en-en_NZ.tsv - config_name: eo-hi data_files: eo-hi.tsv - config_name: en_US-kab data_files: en_US-kab.tsv - config_name: en_GB-ru_RU data_files: en_GB-ru_RU.tsv - config_name: en-kok@latin data_files: en-kok@latin.tsv - config_name: en-ne_NP data_files: en-ne_NP.tsv - config_name: en-no-NO data_files: en-no-NO.tsv - config_name: it-nl_NL data_files: it-nl_NL.tsv - config_name: en-HE data_files: en-HE.tsv - config_name: eo-ja data_files: eo-ja.tsv - config_name: en_US-kmr data_files: en_US-kmr.tsv - config_name: en-pt-BR data_files: en-pt-BR.tsv - config_name: en-pl-v26 data_files: en-pl-v26.tsv - config_name: en_devel-zh-tw data_files: en_devel-zh-tw.tsv - config_name: en-mcc235 data_files: en-mcc235.tsv - config_name: en-el-gr data_files: en-el-gr.tsv - config_name: en-ga data_files: en-ga.tsv - config_name: en_GB-zh_CN data_files: en_GB-zh_CN.tsv - config_name: en_GB-kab data_files: en_GB-kab.tsv - config_name: en-te-IN data_files: en-te-IN.tsv - config_name: en_GB-de data_files: en_GB-de.tsv - config_name: und-de data_files: und-de.tsv - config_name: en-nb-rNO-v26 data_files: en-nb-rNO-v26.tsv - config_name: en-zh_SIMPLIFIED data_files: en-zh_SIMPLIFIED.tsv - config_name: en-ur-rPK data_files: en-ur-rPK.tsv - config_name: en_US-zh-cn data_files: en_US-zh-cn.tsv - config_name: en_devel-pa data_files: en_devel-pa.tsv - config_name: en-aii data_files: en-aii.tsv - config_name: en_GB-it_IT data_files: en_GB-it_IT.tsv - config_name: en_GB-yo data_files: en_GB-yo.tsv - config_name: de-id data_files: de-id.tsv - config_name: en_GB-nv data_files: en_GB-nv.tsv - config_name: en-sw-KE data_files: en-sw-KE.tsv - config_name: en_US-so data_files: en_US-so.tsv - config_name: en-yue data_files: en-yue.tsv - config_name: en-ps data_files: en-ps.tsv - config_name: en-mr-IN data_files: en-mr-IN.tsv - config_name: de-cs data_files: de-cs.tsv - config_name: en_GB-pt-BR data_files: en_GB-pt-BR.tsv - config_name: en-ne data_files: en-ne.tsv - config_name: en_GB-kk data_files: en_GB-kk.tsv - config_name: en-af-ZA data_files: en-af-ZA.tsv - config_name: en-pa data_files: en-pa.tsv - config_name: en_US-lt data_files: en_US-lt.tsv - config_name: en-b+qtq+Latn data_files: en-b+qtq+Latn.tsv - config_name: zh_Hant-zgh data_files: zh_Hant-zgh.tsv - config_name: en-ta-IN data_files: en-ta-IN.tsv - config_name: en_GB-hu data_files: en_GB-hu.tsv - config_name: en-iw data_files: en-iw.tsv - config_name: es-hi data_files: es-hi.tsv - config_name: en-es_EC data_files: en-es_EC.tsv - config_name: en-ukrainian data_files: en-ukrainian.tsv - config_name: en_US-he data_files: en_US-he.tsv - config_name: en_GB-sl data_files: en_GB-sl.tsv - config_name: en_devel-sgs data_files: en_devel-sgs.tsv - config_name: en_US-zh-HK data_files: en_US-zh-HK.tsv - config_name: en_US-th_TH data_files: en_US-th_TH.tsv - config_name: en-nl_NL data_files: en-nl_NL.tsv - config_name: en-zh-HK data_files: en-zh-HK.tsv - config_name: en-zh-hans data_files: en-zh-hans.tsv - config_name: en_devel-he data_files: en_devel-he.tsv - config_name: en_GB-ur data_files: en_GB-ur.tsv - config_name: en_GB-da data_files: en_GB-da.tsv - config_name: en_GB-bn data_files: en_GB-bn.tsv - config_name: en-chinese data_files: en-chinese.tsv - config_name: en-bg-BG data_files: en-bg-BG.tsv - config_name: en_devel-jpn_JP data_files: en_devel-jpn_JP.tsv - config_name: en_devel-id data_files: en_devel-id.tsv - config_name: und-ru data_files: und-ru.tsv - config_name: en_devel-in data_files: en_devel-in.tsv - config_name: en-wo data_files: en-wo.tsv - config_name: nl-da data_files: nl-da.tsv - config_name: en-pa-Arab-PK data_files: en-pa-Arab-PK.tsv - config_name: en-gr-GR data_files: en-gr-GR.tsv - config_name: en-az-AZ data_files: en-az-AZ.tsv - config_name: en-bg data_files: en-bg.tsv - config_name: en-es-rAR data_files: en-es-rAR.tsv - config_name: en-nb-NO data_files: en-nb-NO.tsv - config_name: en_UK-bg_BG data_files: en_UK-bg_BG.tsv - config_name: en_GB-pap data_files: en_GB-pap.tsv - config_name: en_US-es data_files: en_US-es.tsv - config_name: en_US-hu data_files: en_US-hu.tsv - config_name: en-or-IN data_files: en-or-IN.tsv - config_name: en-guw data_files: en-guw.tsv - config_name: en-nl-BE data_files: en-nl-BE.tsv - config_name: en-ml-rIN data_files: en-ml-rIN.tsv - config_name: en-ji data_files: en-ji.tsv - config_name: en_US-ta data_files: en_US-ta.tsv - config_name: es-ur data_files: es-ur.tsv - config_name: en-br data_files: en-br.tsv - config_name: de-en data_files: de-en.tsv - config_name: dev-fr data_files: dev-fr.tsv - config_name: en-ace data_files: en-ace.tsv - config_name: en_US-zh_TW data_files: en_US-zh_TW.tsv - config_name: en-oj data_files: en-oj.tsv - config_name: en-zh_tw data_files: en-zh_tw.tsv - config_name: en-cnr data_files: en-cnr.tsv - config_name: en_devel-es_hn data_files: en_devel-es_hn.tsv - config_name: dev-uk data_files: dev-uk.tsv - config_name: en-ru_CARES data_files: en-ru_CARES.tsv - config_name: en-uroc data_files: en-uroc.tsv - config_name: en_GB-bg_BG data_files: en_GB-bg_BG.tsv - config_name: en_GB-ar_SA data_files: en_GB-ar_SA.tsv - config_name: en_US-fy data_files: en_US-fy.tsv - config_name: en-lt data_files: en-lt.tsv - config_name: en-de-rDE data_files: en-de-rDE.tsv - config_name: en_US-ast data_files: en_US-ast.tsv - config_name: en_US-ko_KR data_files: en_US-ko_KR.tsv - config_name: en_devel-ar_DZ data_files: en_devel-ar_DZ.tsv - config_name: en_devel-hu data_files: en_devel-hu.tsv - config_name: en-fr_BE data_files: en-fr_BE.tsv - config_name: en-kmr data_files: en-kmr.tsv - config_name: en_devel-ro_ro data_files: en_devel-ro_ro.tsv - config_name: en_GB-vi_VN data_files: en_GB-vi_VN.tsv - config_name: en_devel-sk data_files: en_devel-sk.tsv - config_name: und-nl_BE data_files: und-nl_BE.tsv - config_name: eo-bn data_files: eo-bn.tsv - config_name: en-hungarian data_files: en-hungarian.tsv - config_name: en_GB-ta data_files: en_GB-ta.tsv - config_name: en_US-ca data_files: en_US-ca.tsv - config_name: en-oc data_files: en-oc.tsv - config_name: en_US-bg_BG data_files: en_US-bg_BG.tsv - config_name: en-hr data_files: en-hr.tsv - config_name: en_GB-zh_Hant data_files: en_GB-zh_Hant.tsv - config_name: en_GB-bn_BD data_files: en_GB-bn_BD.tsv - config_name: en-ca@valencia data_files: en-ca@valencia.tsv - config_name: en_GB-mai data_files: en_GB-mai.tsv - config_name: en-uk-UA data_files: en-uk-UA.tsv - config_name: en-frm data_files: en-frm.tsv - config_name: en-bd data_files: en-bd.tsv - config_name: en_GB-ja data_files: en_GB-ja.tsv - config_name: en_US-sw data_files: en_US-sw.tsv - config_name: eo-uk data_files: eo-uk.tsv - config_name: en_US-es-rAR data_files: en_US-es-rAR.tsv - config_name: en-az-rAZ data_files: en-az-rAZ.tsv - config_name: en_GB-es-ES data_files: en_GB-es-ES.tsv - config_name: en-sl-SL data_files: en-sl-SL.tsv - config_name: en-pms data_files: en-pms.tsv - config_name: en_GB-te data_files: en_GB-te.tsv - config_name: it-de_DE data_files: it-de_DE.tsv - config_name: en-yue_Hant data_files: en-yue_Hant.tsv - config_name: en-en-rIN data_files: en-en-rIN.tsv - config_name: en-ln data_files: en-ln.tsv - config_name: en-pt-rBR data_files: en-pt-rBR.tsv - config_name: en_US-az_AZ data_files: en_US-az_AZ.tsv - config_name: en-pl-rPL data_files: en-pl-rPL.tsv - config_name: eo-el data_files: eo-el.tsv - config_name: eo-ms data_files: eo-ms.tsv - config_name: en_US-tr data_files: en_US-tr.tsv - config_name: en-en_SHAW data_files: en-en_SHAW.tsv - config_name: en-ar-rIQ data_files: en-ar-rIQ.tsv - config_name: en-yo data_files: en-yo.tsv - config_name: en-japanese data_files: en-japanese.tsv - config_name: es-id data_files: es-id.tsv - config_name: en-fa_AF data_files: en-fa_AF.tsv - config_name: en_GB-ms data_files: en_GB-ms.tsv - config_name: en-Zh-CHS data_files: en-Zh-CHS.tsv - config_name: en_GB-mt data_files: en_GB-mt.tsv - config_name: en-b+de data_files: en-b+de.tsv - config_name: en_US-fi data_files: en_US-fi.tsv - config_name: de-ar data_files: de-ar.tsv - config_name: en-en-GB data_files: en-en-GB.tsv - config_name: en-mo data_files: en-mo.tsv - config_name: en_devel-zh_Hans data_files: en_devel-zh_Hans.tsv - config_name: en_GB-dz data_files: en_GB-dz.tsv - config_name: en_US-gl data_files: en_US-gl.tsv - config_name: en-pt-rPT data_files: en-pt-rPT.tsv - config_name: en_devel-es_pr data_files: en_devel-es_pr.tsv - config_name: en-RU data_files: en-RU.tsv - config_name: en-en-rUS data_files: en-en-rUS.tsv - config_name: en-sv_se data_files: en-sv_se.tsv - config_name: en-italian data_files: en-italian.tsv - config_name: en_US-lv data_files: en_US-lv.tsv - config_name: de-ru data_files: de-ru.tsv - config_name: en-sc data_files: en-sc.tsv - config_name: en-gv data_files: en-gv.tsv - config_name: en_US-pt_PT data_files: en_US-pt_PT.tsv - config_name: en_GB-bn_IN data_files: en_GB-bn_IN.tsv - config_name: en_US-fr-FR data_files: en_US-fr-FR.tsv - config_name: ia-es data_files: ia-es.tsv - config_name: en_US-es_UY data_files: en_US-es_UY.tsv - config_name: en_GB-hr_HR data_files: en_GB-hr_HR.tsv - config_name: en-id_ID data_files: en-id_ID.tsv - config_name: en-es_VE data_files: en-es_VE.tsv - config_name: en-ie data_files: en-ie.tsv - config_name: en-it_IT data_files: en-it_IT.tsv - config_name: en_GB-si_LK data_files: en_GB-si_LK.tsv - config_name: en-nqo data_files: en-nqo.tsv - config_name: pl-uk data_files: pl-uk.tsv - config_name: en-sco data_files: en-sco.tsv - config_name: en_US-tr-TR data_files: en_US-tr-TR.tsv - config_name: en-en_GB data_files: en-en_GB.tsv - config_name: en-b+kab data_files: en-b+kab.tsv - config_name: en-he-rIL data_files: en-he-rIL.tsv - config_name: en-pu data_files: en-pu.tsv - config_name: de-lb data_files: de-lb.tsv - config_name: en-is_IS data_files: en-is_IS.tsv - config_name: en_US-cs data_files: en_US-cs.tsv - config_name: en_GB-nah data_files: en_GB-nah.tsv - config_name: de-tr data_files: de-tr.tsv - config_name: zh_Hant-en_US data_files: zh_Hant-en_US.tsv - config_name: pl-ru data_files: pl-ru.tsv - config_name: en-zh-TW data_files: en-zh-TW.tsv - config_name: en_GB-kok data_files: en_GB-kok.tsv - config_name: en_US-zh-Hans data_files: en_US-zh-Hans.tsv - config_name: en_devel-da data_files: en_devel-da.tsv - config_name: en-mg data_files: en-mg.tsv - config_name: en-pa-rIN data_files: en-pa-rIN.tsv - config_name: en-nb_NO data_files: en-nb_NO.tsv - config_name: en_GB-az data_files: en_GB-az.tsv - config_name: en-ca_valencia data_files: en-ca_valencia.tsv - config_name: en-su data_files: en-su.tsv - config_name: und-sv data_files: und-sv.tsv - config_name: pl-en data_files: pl-en.tsv - config_name: en-ar-rDZ data_files: en-ar-rDZ.tsv - config_name: en_US-eo data_files: en_US-eo.tsv - config_name: en_US-sq data_files: en_US-sq.tsv - config_name: en-sl-rSI data_files: en-sl-rSI.tsv - config_name: en-uk-rUA data_files: en-uk-rUA.tsv - config_name: en_devel-te data_files: en_devel-te.tsv - config_name: en-da_DK data_files: en-da_DK.tsv - config_name: en_GB-et_EE data_files: en_GB-et_EE.tsv - config_name: en-et-EE data_files: en-et-EE.tsv - config_name: en-pa_IN data_files: en-pa_IN.tsv - config_name: en_US-nn data_files: en_US-nn.tsv - config_name: en_GB-xh data_files: en_GB-xh.tsv - config_name: en_devel-sv data_files: en_devel-sv.tsv - config_name: en-ru-rRU data_files: en-ru-rRU.tsv - config_name: en_US-hr data_files: en_US-hr.tsv - config_name: en-sr_Latn data_files: en-sr_Latn.tsv - config_name: en_GB-uk data_files: en_GB-uk.tsv - config_name: en_GB-ee data_files: en_GB-ee.tsv - config_name: en_devel-ta data_files: en_devel-ta.tsv - config_name: en_US-hu-HU data_files: en_US-hu-HU.tsv - config_name: en_GB-ak data_files: en_GB-ak.tsv - config_name: en_US-ia data_files: en_US-ia.tsv - config_name: en_UK-it_IT data_files: en_UK-it_IT.tsv - config_name: en-ru data_files: en-ru.tsv - config_name: en_US-es-ar data_files: en_US-es-ar.tsv - config_name: en_US-lo data_files: en_US-lo.tsv - config_name: en-ur-PK data_files: en-ur-PK.tsv - config_name: en_devel-nb_NO data_files: en_devel-nb_NO.tsv - config_name: en_GB-es_ES data_files: en_GB-es_ES.tsv - config_name: en_GB-ast data_files: en_GB-ast.tsv - config_name: en-hr-HR data_files: en-hr-HR.tsv - config_name: en-fr@informal data_files: en-fr@informal.tsv - config_name: en-es_ar data_files: en-es_ar.tsv - config_name: en-ms_MY data_files: en-ms_MY.tsv - config_name: en-el_GR data_files: en-el_GR.tsv - config_name: en_devel-ka data_files: en_devel-ka.tsv - config_name: en-fr-FR data_files: en-fr-FR.tsv - config_name: en_US-kk data_files: en_US-kk.tsv - config_name: es-ko data_files: es-ko.tsv - config_name: en-fr_AG data_files: en-fr_AG.tsv - config_name: en-zh-tw data_files: en-zh-tw.tsv - config_name: en-BrazilianPortuguese data_files: en-BrazilianPortuguese.tsv - config_name: en_GB-am data_files: en_GB-am.tsv - config_name: en-tam data_files: en-tam.tsv - config_name: en_US-af data_files: en_US-af.tsv - config_name: en_US-is data_files: en_US-is.tsv - config_name: en_GB-en_US data_files: en_GB-en_US.tsv - config_name: en-az data_files: en-az.tsv - config_name: en-en@pirate data_files: en-en@pirate.tsv - config_name: en_GB-fil data_files: en_GB-fil.tsv - config_name: en_US-pl_PL data_files: en_US-pl_PL.tsv - config_name: en_US-sl data_files: en_US-sl.tsv - config_name: en_US-nl data_files: en_US-nl.tsv - config_name: es-it data_files: es-it.tsv - config_name: en_GB-bar data_files: en_GB-bar.tsv - config_name: it-nb_NO data_files: it-nb_NO.tsv - config_name: eo-it data_files: eo-it.tsv - config_name: en_US-yue data_files: en_US-yue.tsv - config_name: en-glk data_files: en-glk.tsv - config_name: en-fi_FI data_files: en-fi_FI.tsv - config_name: es-cs data_files: es-cs.tsv - config_name: en_GB-pt_BR data_files: en_GB-pt_BR.tsv - config_name: en_GB-zgh data_files: en_GB-zgh.tsv - config_name: en_US-nl-BE data_files: en_US-nl-BE.tsv - config_name: en-ru-rCH data_files: en-ru-rCH.tsv - config_name: en-sr_CS data_files: en-sr_CS.tsv - config_name: en-ur data_files: en-ur.tsv - config_name: en_GB-th data_files: en_GB-th.tsv - config_name: en_US-id_ID data_files: en_US-id_ID.tsv - config_name: en_US-be_BY data_files: en_US-be_BY.tsv - config_name: en_devel-es_us data_files: en_devel-es_us.tsv - config_name: en-fr_CA data_files: en-fr_CA.tsv - config_name: en_GB-en data_files: en_GB-en.tsv - config_name: en_US-sk data_files: en_US-sk.tsv - config_name: en-uz-Latn data_files: en-uz-Latn.tsv - config_name: en_devel-eu data_files: en_devel-eu.tsv - config_name: en_GB-is_IS data_files: en_GB-is_IS.tsv - config_name: sl-en data_files: sl-en.tsv - config_name: en-ja_JA data_files: en-ja_JA.tsv - config_name: en-bn-BD data_files: en-bn-BD.tsv - config_name: fr-de data_files: fr-de.tsv - config_name: en-sr_SP data_files: en-sr_SP.tsv - config_name: en-nb-no data_files: en-nb-no.tsv - config_name: fr-nb_NO data_files: fr-nb_NO.tsv - config_name: en_US-lb data_files: en_US-lb.tsv - config_name: en-zh_hant data_files: en-zh_hant.tsv - config_name: en-be data_files: en-be.tsv - config_name: en_US-si data_files: en_US-si.tsv - config_name: en-ltg data_files: en-ltg.tsv - config_name: en-es_cl data_files: en-es_cl.tsv - config_name: en_US-gu data_files: en_US-gu.tsv - config_name: en-lb_LU data_files: en-lb_LU.tsv - config_name: en-ain data_files: en-ain.tsv - config_name: en-de data_files: en-de.tsv - config_name: en-es data_files: en-es.tsv - config_name: en-belarusian data_files: en-belarusian.tsv - config_name: en-kok data_files: en-kok.tsv - config_name: nl-fr data_files: nl-fr.tsv - config_name: en-ar_SA data_files: en-ar_SA.tsv - config_name: en-tk data_files: en-tk.tsv - config_name: en-kab data_files: en-kab.tsv - config_name: en-or-rIN data_files: en-or-rIN.tsv - config_name: en-ja-KS data_files: en-ja-KS.tsv - config_name: en-en-Shaw data_files: en-en-Shaw.tsv - config_name: en_GB-lo data_files: en_GB-lo.tsv - config_name: en_GB-gl_ES data_files: en_GB-gl_ES.tsv - config_name: en-sd data_files: en-sd.tsv - config_name: en_devel-es_ar data_files: en_devel-es_ar.tsv - config_name: en-he-il data_files: en-he-il.tsv - config_name: en_GB-zh_TW data_files: en_GB-zh_TW.tsv - config_name: en-cs_cz data_files: en-cs_cz.tsv - config_name: en_GB-mn data_files: en_GB-mn.tsv - config_name: en_US-jv data_files: en_US-jv.tsv - config_name: eo-nl data_files: eo-nl.tsv - config_name: en-zh_cn data_files: en-zh_cn.tsv - config_name: en-he_IL data_files: en-he_IL.tsv - config_name: en-IT data_files: en-IT.tsv - config_name: en-ja data_files: en-ja.tsv - config_name: en_US-fr-ca data_files: en_US-fr-ca.tsv - config_name: en-bqi data_files: en-bqi.tsv - config_name: en-ro-rRO data_files: en-ro-rRO.tsv - config_name: en-krl data_files: en-krl.tsv - config_name: en_US-tr_TR data_files: en_US-tr_TR.tsv - config_name: pl-lt data_files: pl-lt.tsv - config_name: en-zh_Hant_HK data_files: en-zh_Hant_HK.tsv - config_name: en_GB-sv_SE data_files: en_GB-sv_SE.tsv - config_name: en_US-pt-br data_files: en_US-pt-br.tsv - config_name: en-id-ID data_files: en-id-ID.tsv - config_name: en-fu data_files: en-fu.tsv - config_name: en-French data_files: en-French.tsv - config_name: eo-zh data_files: eo-zh.tsv - config_name: en-v20 data_files: en-v20.tsv - config_name: en-iw-IL data_files: en-iw-IL.tsv - config_name: en_GB-af data_files: en_GB-af.tsv - config_name: en_GB-el data_files: en_GB-el.tsv - config_name: en-pa-IN data_files: en-pa-IN.tsv - config_name: en_devel-es_ve data_files: en_devel-es_ve.tsv - config_name: und-nb_NO data_files: und-nb_NO.tsv - config_name: en-lo data_files: en-lo.tsv - config_name: en-ar data_files: en-ar.tsv - config_name: en-b+zh+HANS+CN data_files: en-b+zh+HANS+CN.tsv - config_name: en_GB-byn data_files: en_GB-byn.tsv - config_name: en-en-rXC data_files: en-en-rXC.tsv - config_name: zh_Hant-nb_NO data_files: zh_Hant-nb_NO.tsv - config_name: en-fr data_files: en-fr.tsv - config_name: en-zh_HANT data_files: en-zh_HANT.tsv - config_name: en_US-fa-IR data_files: en_US-fa-IR.tsv - config_name: en_GB-vi data_files: en_GB-vi.tsv - config_name: en-Spanish data_files: en-Spanish.tsv - config_name: en-am_ET data_files: en-am_ET.tsv - config_name: en_devel-bn data_files: en_devel-bn.tsv - config_name: en-zh-cn data_files: en-zh-cn.tsv - config_name: en-tr-rTR data_files: en-tr-rTR.tsv - config_name: fr-cs data_files: fr-cs.tsv - config_name: en_US-nl-rBE data_files: en_US-nl-rBE.tsv - config_name: es-en data_files: es-en.tsv - config_name: en-sr@Cyrl data_files: en-sr@Cyrl.tsv - config_name: fr-eu data_files: fr-eu.tsv - config_name: en_US-pl data_files: en_US-pl.tsv - config_name: en_US-nan data_files: en_US-nan.tsv - config_name: en_devel-pt-rBR data_files: en_devel-pt-rBR.tsv - config_name: en-sr_lat data_files: en-sr_lat.tsv - config_name: en_devel-no data_files: en_devel-no.tsv - config_name: pl-de data_files: pl-de.tsv - config_name: en-tlh data_files: en-tlh.tsv - config_name: en_US-cs_CZ data_files: en_US-cs_CZ.tsv - config_name: eo-pl data_files: eo-pl.tsv - config_name: en_devel-gl data_files: en_devel-gl.tsv - config_name: en-fi-FI data_files: en-fi-FI.tsv - config_name: en_US-ca_CA data_files: en_US-ca_CA.tsv - config_name: en_US-nb data_files: en_US-nb.tsv - config_name: en-is-IS data_files: en-is-IS.tsv - config_name: en_GB-io data_files: en_GB-io.tsv - config_name: en-UK data_files: en-UK.tsv - config_name: en-pt-pt data_files: en-pt-pt.tsv - config_name: en-fil data_files: en-fil.tsv - config_name: en-mi data_files: en-mi.tsv - config_name: en-sr-Cyrl data_files: en-sr-Cyrl.tsv - config_name: en_devel-hi data_files: en_devel-hi.tsv - config_name: en-nb-NB data_files: en-nb-NB.tsv - config_name: en-mnc data_files: en-mnc.tsv - config_name: en-mk data_files: en-mk.tsv - config_name: en-hrx data_files: en-hrx.tsv - config_name: en-ar_MA data_files: en-ar_MA.tsv - config_name: en_devel-es data_files: en_devel-es.tsv - config_name: en_GB-zh-rCN data_files: en_GB-zh-rCN.tsv - config_name: en-sa data_files: en-sa.tsv - config_name: en-bs data_files: en-bs.tsv - config_name: en_GB-tg data_files: en_GB-tg.tsv - config_name: en-si-LK data_files: en-si-LK.tsv - config_name: en-lt-LT data_files: en-lt-LT.tsv - config_name: en-hi data_files: en-hi.tsv - config_name: en-hu_hu data_files: en-hu_hu.tsv - config_name: en-mk_MK data_files: en-mk_MK.tsv - config_name: en_GB-de_DE data_files: en_GB-de_DE.tsv - config_name: messages-eo data_files: messages-eo.tsv - config_name: en-ku_IQ data_files: en-ku_IQ.tsv - config_name: en-rcf data_files: en-rcf.tsv - config_name: en-uz data_files: en-uz.tsv - config_name: en-by_lat data_files: en-by_lat.tsv - config_name: ia-nb_NO data_files: ia-nb_NO.tsv - config_name: messages-ko data_files: messages-ko.tsv - config_name: en_US-pt-rBR data_files: en_US-pt-rBR.tsv - config_name: en_GB-zu data_files: en_GB-zu.tsv - config_name: es-hr data_files: es-hr.tsv - config_name: en_devel-th data_files: en_devel-th.tsv - config_name: en-af data_files: en-af.tsv - config_name: en-ms-MY data_files: en-ms-MY.tsv - config_name: en-sr-Latn-RS data_files: en-sr-Latn-RS.tsv - config_name: en-de-ZH data_files: en-de-ZH.tsv - config_name: en-b+sr+Latn data_files: en-b+sr+Latn.tsv - config_name: en-cn data_files: en-cn.tsv - config_name: de-zh_Hans data_files: de-zh_Hans.tsv - config_name: en_devel-gu data_files: en_devel-gu.tsv - config_name: en_US-et_EE data_files: en_US-et_EE.tsv - config_name: en-und data_files: en-und.tsv - config_name: en_devel-es_ni data_files: en_devel-es_ni.tsv - config_name: en-en-rNZ data_files: en-en-rNZ.tsv - config_name: pl-fr data_files: pl-fr.tsv - config_name: de-es data_files: de-es.tsv - config_name: en-pt_br data_files: en-pt_br.tsv - config_name: en-gug data_files: en-gug.tsv - config_name: fr-fr data_files: fr-fr.tsv - config_name: en-fr-rFR data_files: en-fr-rFR.tsv - config_name: en-dsb data_files: en-dsb.tsv - config_name: en-tr-TR data_files: en-tr-TR.tsv - config_name: en-tw data_files: en-tw.tsv - config_name: en-bs_Latn data_files: en-bs_Latn.tsv - config_name: en_GB-hi data_files: en_GB-hi.tsv - config_name: en-norwegian data_files: en-norwegian.tsv - config_name: en-zh_Latn_pinyin data_files: en-zh_Latn_pinyin.tsv - config_name: en_US-es-mx data_files: en_US-es-mx.tsv - config_name: en_GB-nl_NL data_files: en_GB-nl_NL.tsv - config_name: es-bn data_files: es-bn.tsv - config_name: en-peo data_files: en-peo.tsv - config_name: en-de_LU data_files: en-de_LU.tsv - config_name: en-mni data_files: en-mni.tsv - config_name: en_GB-jam data_files: en_GB-jam.tsv - config_name: en-sr_cyr data_files: en-sr_cyr.tsv - config_name: en-ro-RO data_files: en-ro-RO.tsv - config_name: en-doi data_files: en-doi.tsv - config_name: en_GB-en-US data_files: en_GB-en-US.tsv - config_name: en-he data_files: en-he.tsv - config_name: en-et data_files: en-et.tsv - config_name: en-tl_PH data_files: en-tl_PH.tsv - config_name: en-sr-Cyrl-RS data_files: en-sr-Cyrl-RS.tsv - config_name: en-Dutch data_files: en-Dutch.tsv - config_name: en-uz_UZ data_files: en-uz_UZ.tsv - config_name: en-ur-rIN data_files: en-ur-rIN.tsv - config_name: en-kn data_files: en-kn.tsv - config_name: en-trv data_files: en-trv.tsv - config_name: en_US-ms_MY data_files: en_US-ms_MY.tsv - config_name: en-de-rFO data_files: en-de-rFO.tsv - config_name: en-zh-CN data_files: en-zh-CN.tsv - config_name: ru-de data_files: ru-de.tsv - config_name: en-pt_BR data_files: en-pt_BR.tsv - config_name: en_GB-ms_MY data_files: en_GB-ms_MY.tsv - config_name: en_GB-tr data_files: en_GB-tr.tsv - config_name: en-bn_IN data_files: en-bn_IN.tsv - config_name: en_GB-pt data_files: en_GB-pt.tsv - config_name: en_GB-wa data_files: en_GB-wa.tsv - config_name: en_US-te data_files: en_US-te.tsv - config_name: en-da-rDK data_files: en-da-rDK.tsv - config_name: en_US-zh_CN data_files: en_US-zh_CN.tsv - config_name: en_US-az data_files: en_US-az.tsv - config_name: en-sn data_files: en-sn.tsv - config_name: en_devel-zh_Hant data_files: en_devel-zh_Hant.tsv - config_name: en-sw data_files: en-sw.tsv - config_name: en-fr_fr data_files: en-fr_fr.tsv - config_name: en_GB-mhr data_files: en_GB-mhr.tsv - config_name: sv-se data_files: sv-se.tsv - config_name: en-mn data_files: en-mn.tsv - config_name: en-gl data_files: en-gl.tsv - config_name: en_GB-is data_files: en_GB-is.tsv - config_name: en-nl-NL data_files: en-nl-NL.tsv - config_name: dev-fa data_files: dev-fa.tsv - config_name: en-frp data_files: en-frp.tsv - config_name: en_GB-it data_files: en_GB-it.tsv - config_name: en_US-ja-JP data_files: en_US-ja-JP.tsv - config_name: en_US-vi_VN data_files: en_US-vi_VN.tsv - config_name: en-zu data_files: en-zu.tsv - config_name: en_US-zh_HK data_files: en_US-zh_HK.tsv - config_name: en_UK-nb_NO data_files: en_UK-nb_NO.tsv - config_name: en_GB-eo data_files: en_GB-eo.tsv - config_name: en-ar_YE data_files: en-ar_YE.tsv - config_name: messages-pt data_files: messages-pt.tsv - config_name: en_devel-hr data_files: en_devel-hr.tsv - config_name: ia-en data_files: ia-en.tsv - config_name: en-sr data_files: en-sr.tsv - config_name: en_US-el_GR data_files: en_US-el_GR.tsv - config_name: en_US-bg data_files: en_US-bg.tsv - config_name: en-be@latin data_files: en-be@latin.tsv - config_name: en_US-zh_Hant data_files: en_US-zh_Hant.tsv - config_name: eo-fr data_files: eo-fr.tsv - config_name: en-uk_UA data_files: en-uk_UA.tsv - config_name: en_US-pt-BR data_files: en_US-pt-BR.tsv - config_name: nl-ko data_files: nl-ko.tsv - config_name: en-sl-SI data_files: en-sl-SI.tsv - config_name: en-to data_files: en-to.tsv - config_name: en_GB-ne data_files: en_GB-ne.tsv - config_name: en-la data_files: en-la.tsv - config_name: ru-ua data_files: ru-ua.tsv - config_name: en_GB-ia data_files: en_GB-ia.tsv - config_name: en_US-bn_BD data_files: en_US-bn_BD.tsv - config_name: en-zh_Hant data_files: en-zh_Hant.tsv - config_name: en_devel-nl_BE data_files: en_devel-nl_BE.tsv - config_name: en-id data_files: en-id.tsv - config_name: en_GB-pa data_files: en_GB-pa.tsv - config_name: en-gl_ES data_files: en-gl_ES.tsv - config_name: en-vi data_files: en-vi.tsv - config_name: fr-es data_files: fr-es.tsv - config_name: en-udm data_files: en-udm.tsv - config_name: en-es-rUS data_files: en-es-rUS.tsv - config_name: en-b+tok data_files: en-b+tok.tsv - config_name: it-fr_FR data_files: it-fr_FR.tsv - config_name: und-nl data_files: und-nl.tsv - config_name: en-pt_pt data_files: en-pt_pt.tsv - config_name: en-es_419 data_files: en-es_419.tsv - config_name: en-jbo data_files: en-jbo.tsv - config_name: en_GB-nb-rNO data_files: en_GB-nb-rNO.tsv - config_name: en_GB-nl data_files: en_GB-nl.tsv - config_name: en-gl-ES data_files: en-gl-ES.tsv - config_name: en-de_AT data_files: en-de_AT.tsv - config_name: en-mk-MK data_files: en-mk-MK.tsv - config_name: en_GB-bg data_files: en_GB-bg.tsv - config_name: en_US-sc data_files: en_US-sc.tsv - config_name: en_US-kn data_files: en_US-kn.tsv - config_name: en-cy_GB data_files: en-cy_GB.tsv - config_name: en_US-mn data_files: en_US-mn.tsv - config_name: de-uk data_files: de-uk.tsv - config_name: en_GB-ko data_files: en_GB-ko.tsv - config_name: en-nl-rNL data_files: en-nl-rNL.tsv - config_name: en_devel-pt_PT data_files: en_devel-pt_PT.tsv - config_name: en_US-fi_FI data_files: en_US-fi_FI.tsv - config_name: en_devel-vi data_files: en_devel-vi.tsv - config_name: en_US-ru data_files: en_US-ru.tsv - config_name: en-hne data_files: en-hne.tsv - config_name: en-fi data_files: en-fi.tsv - config_name: en-ru_RU data_files: en-ru_RU.tsv - config_name: en_devel-es_cl data_files: en_devel-es_cl.tsv - config_name: de-el data_files: de-el.tsv - config_name: en_devel-ro data_files: en_devel-ro.tsv - config_name: en_GB-tt data_files: en_GB-tt.tsv - config_name: en-eng_GB data_files: en-eng_GB.tsv - config_name: en-lt-rLT data_files: en-lt-rLT.tsv - config_name: en-ota data_files: en-ota.tsv - config_name: en_devel-es_co data_files: en_devel-es_co.tsv - config_name: en-russian data_files: en-russian.tsv - config_name: en-ar-MA data_files: en-ar-MA.tsv - config_name: en-nn data_files: en-nn.tsv - config_name: eo-en data_files: eo-en.tsv - config_name: en_GB-cv data_files: en_GB-cv.tsv - config_name: en_devel-id_ID data_files: en_devel-id_ID.tsv - config_name: en_US-nb-NO data_files: en_US-nb-NO.tsv - config_name: en-it-rIT data_files: en-it-rIT.tsv - config_name: en_US-pl-PL data_files: en_US-pl-PL.tsv - config_name: en-ext data_files: en-ext.tsv - config_name: en-ko data_files: en-ko.tsv - config_name: en-tg data_files: en-tg.tsv - config_name: en-ga_IE data_files: en-ga_IE.tsv - config_name: en_devel-sr data_files: en_devel-sr.tsv - config_name: en-PT data_files: en-PT.tsv - config_name: en-sv data_files: en-sv.tsv - config_name: en_GB-son data_files: en_GB-son.tsv - config_name: en-et_ee data_files: en-et_ee.tsv - config_name: en_GB-el_GR data_files: en_GB-el_GR.tsv - config_name: en-jp data_files: en-jp.tsv - config_name: en-ga-rIE data_files: en-ga-rIE.tsv - config_name: sv-en data_files: sv-en.tsv - config_name: en_US-ua data_files: en_US-ua.tsv - config_name: en-sm data_files: en-sm.tsv - config_name: en-nap data_files: en-nap.tsv - config_name: en-portuguese data_files: en-portuguese.tsv - config_name: en_US-nl-NL data_files: en_US-nl-NL.tsv - config_name: en-es_ec data_files: en-es_ec.tsv - config_name: en_GB-crh data_files: en_GB-crh.tsv - config_name: en-tr_TR data_files: en-tr_TR.tsv - config_name: en-sr_RS@latin data_files: en-sr_RS@latin.tsv - config_name: en-bg_BG data_files: en-bg_BG.tsv - config_name: en-hu data_files: en-hu.tsv - config_name: en-es_SV data_files: en-es_SV.tsv - config_name: en_GB-rw data_files: en_GB-rw.tsv - config_name: en-es_AR data_files: en-es_AR.tsv - config_name: en_devel-es_pe data_files: en_devel-es_pe.tsv - config_name: en-et-rEE data_files: en-et-rEE.tsv - config_name: en-ro-v26 data_files: en-ro-v26.tsv - config_name: en-ne-NP data_files: en-ne-NP.tsv - config_name: en-es-ar data_files: en-es-ar.tsv - config_name: en-en_ZA data_files: en-en_ZA.tsv - config_name: en_devel-lt data_files: en_devel-lt.tsv - config_name: en-eg data_files: en-eg.tsv - config_name: zh_Latn-zh_Hans data_files: zh_Latn-zh_Hans.tsv - config_name: en_GB-so data_files: en_GB-so.tsv - config_name: en-hr-rHR data_files: en-hr-rHR.tsv - config_name: en-lt_LT data_files: en-lt_LT.tsv - config_name: en-io data_files: en-io.tsv - config_name: en-sh-rHR data_files: en-sh-rHR.tsv - config_name: en-uk data_files: en-uk.tsv - config_name: en_GB-cs-CZ data_files: en_GB-cs-CZ.tsv - config_name: en-de-rCH data_files: en-de-rCH.tsv - config_name: en-nah data_files: en-nah.tsv - config_name: en_devel-tr data_files: en_devel-tr.tsv - config_name: en-de-rAT data_files: en-de-rAT.tsv - config_name: eo-sv data_files: eo-sv.tsv - config_name: en-nb data_files: en-nb.tsv - config_name: en_GB-ab data_files: en_GB-ab.tsv - config_name: en_US-de-DE data_files: en_US-de-DE.tsv - config_name: en-de_alm_x data_files: en-de_alm_x.tsv - config_name: en_GB-it-IT data_files: en_GB-it-IT.tsv - config_name: en-aa data_files: en-aa.tsv - config_name: en_devel-sq data_files: en_devel-sq.tsv - config_name: en_devel-en_au data_files: en_devel-en_au.tsv - config_name: en-sl data_files: en-sl.tsv - config_name: en-sr-rSP data_files: en-sr-rSP.tsv - config_name: en-ckb data_files: en-ckb.tsv - config_name: en_devel-pt_pt data_files: en_devel-pt_pt.tsv - config_name: en_devel-ar data_files: en_devel-ar.tsv - config_name: en-nn-NO data_files: en-nn-NO.tsv - config_name: es-fr data_files: es-fr.tsv - config_name: en-mk-rMK data_files: en-mk-rMK.tsv - config_name: en-spanish data_files: en-spanish.tsv - config_name: en_GB-ve data_files: en_GB-ve.tsv - config_name: en_GB-zh_HK data_files: en_GB-zh_HK.tsv - config_name: en_GB-kmr data_files: en_GB-kmr.tsv - config_name: en-no_nb data_files: en-no_nb.tsv - config_name: en_GB-sq data_files: en_GB-sq.tsv - config_name: en_US-ro-RO data_files: en_US-ro-RO.tsv - config_name: en-zh-rHK data_files: en-zh-rHK.tsv - config_name: en-Russian data_files: en-Russian.tsv - config_name: en_GB-ht data_files: en_GB-ht.tsv - config_name: en_GB-ug data_files: en_GB-ug.tsv - config_name: en-na data_files: en-na.tsv - config_name: en_devel-es_gt data_files: en_devel-es_gt.tsv - config_name: en-ka-rGE data_files: en-ka-rGE.tsv - config_name: en_US-bn-rBD data_files: en_US-bn-rBD.tsv - config_name: eo-ro data_files: eo-ro.tsv - config_name: en_GB-ko_KR data_files: en_GB-ko_KR.tsv - config_name: en-sr@Latn data_files: en-sr@Latn.tsv - config_name: en-french data_files: en-french.tsv - config_name: es-nl data_files: es-nl.tsv - config_name: en-georgian data_files: en-georgian.tsv - config_name: en_devel-sl data_files: en_devel-sl.tsv - config_name: en-jv data_files: en-jv.tsv - config_name: en-ur-UR data_files: en-ur-UR.tsv - config_name: en-dv data_files: en-dv.tsv - config_name: en_US-pt-PT data_files: en_US-pt-PT.tsv - config_name: en-ar_LY data_files: en-ar_LY.tsv - config_name: en-sv-SE data_files: en-sv-SE.tsv - config_name: en-ca_ES@valencia data_files: en-ca_ES@valencia.tsv - config_name: en_devel-oc data_files: en_devel-oc.tsv - config_name: en-th_TH data_files: en-th_TH.tsv - config_name: en-de_CH data_files: en-de_CH.tsv - config_name: en-ca-valencia data_files: en-ca-valencia.tsv - config_name: en-crh data_files: en-crh.tsv - config_name: en_US-en@pirate data_files: en_US-en@pirate.tsv - config_name: en-haw data_files: en-haw.tsv - config_name: en-sk-rSK data_files: en-sk-rSK.tsv - config_name: en-sr@latin data_files: en-sr@latin.tsv - config_name: en-jam data_files: en-jam.tsv - config_name: en_devel-ko data_files: en_devel-ko.tsv - config_name: en_devel-de data_files: en_devel-de.tsv - config_name: messages-nb_NO data_files: messages-nb_NO.tsv - config_name: en_GB-no data_files: en_GB-no.tsv - config_name: en_US-tok data_files: en_US-tok.tsv - config_name: en_US-zh_Hans data_files: en_US-zh_Hans.tsv - config_name: en-hsb data_files: en-hsb.tsv - config_name: en-eo data_files: en-eo.tsv - config_name: en-eu_ES data_files: en-eu_ES.tsv - config_name: en-ayc data_files: en-ayc.tsv - config_name: en-ca data_files: en-ca.tsv - config_name: en-fr_LU data_files: en-fr_LU.tsv - config_name: en-vi-rVN data_files: en-vi-rVN.tsv - config_name: en-pr data_files: en-pr.tsv - config_name: en-vls data_files: en-vls.tsv - config_name: es-gl data_files: es-gl.tsv - config_name: en_GB-nb-NO data_files: en_GB-nb-NO.tsv - config_name: en_GB-haw data_files: en_GB-haw.tsv - config_name: pt_BR-es data_files: pt_BR-es.tsv - config_name: en-nn-rNO data_files: en-nn-rNO.tsv - config_name: en_US-zh-tw data_files: en_US-zh-tw.tsv - config_name: en-ar-AA data_files: en-ar-AA.tsv - config_name: en_GB-fr_FR data_files: en_GB-fr_FR.tsv - config_name: en_GB-gez data_files: en_GB-gez.tsv - config_name: en-ID data_files: en-ID.tsv - config_name: en_GB-oc data_files: en_GB-oc.tsv - config_name: es-ia data_files: es-ia.tsv - config_name: en_GB-kv data_files: en_GB-kv.tsv - config_name: en-es-419 data_files: en-es-419.tsv - config_name: eo-pt data_files: eo-pt.tsv - config_name: it-en_EN data_files: it-en_EN.tsv - config_name: en-czech data_files: en-czech.tsv - config_name: eo-cs data_files: eo-cs.tsv - config_name: en_devel-es_sv data_files: en_devel-es_sv.tsv - config_name: en-es_CL data_files: en-es_CL.tsv - config_name: en-si data_files: en-si.tsv - config_name: en-cs data_files: en-cs.tsv - config_name: en-sv_SE data_files: en-sv_SE.tsv - config_name: en_US-ne_NP data_files: en_US-ne_NP.tsv - config_name: en_GB-fy data_files: en_GB-fy.tsv - config_name: en_devel-en-rGB data_files: en_devel-en-rGB.tsv - config_name: en_GB-sr data_files: en_GB-sr.tsv - config_name: en-es-rPE data_files: en-es-rPE.tsv - config_name: en_US-en data_files: en_US-en.tsv - config_name: en_GB-eu data_files: en_GB-eu.tsv - config_name: en_GB-nb_NO data_files: en_GB-nb_NO.tsv - config_name: en-uz-UZ data_files: en-uz-UZ.tsv - config_name: eo-ko data_files: eo-ko.tsv - config_name: en-lb data_files: en-lb.tsv - config_name: en-lg data_files: en-lg.tsv - config_name: en-Esperanto data_files: en-Esperanto.tsv - config_name: en-ar-SA data_files: en-ar-SA.tsv - config_name: en_GB-ro_RO data_files: en_GB-ro_RO.tsv - config_name: en-cmn data_files: en-cmn.tsv - config_name: en-mni@bengali data_files: en-mni@bengali.tsv - config_name: en-ks data_files: en-ks.tsv - config_name: en_US-pt_BR data_files: en_US-pt_BR.tsv - config_name: ru-nb_NO data_files: ru-nb_NO.tsv - config_name: en-fr-rCA data_files: en-fr-rCA.tsv - config_name: en-kn-rIN data_files: en-kn-rIN.tsv - config_name: en_devel-sq_al data_files: en_devel-sq_al.tsv - config_name: en_US-nb_NO data_files: en_US-nb_NO.tsv - config_name: en-ce data_files: en-ce.tsv - config_name: en_US-ga data_files: en_US-ga.tsv - config_name: en-en-rZA data_files: en-en-rZA.tsv - config_name: en-rue data_files: en-rue.tsv - config_name: en-es_CO data_files: en-es_CO.tsv - config_name: en-es-es data_files: en-es-es.tsv - config_name: en-fa data_files: en-fa.tsv - config_name: en-de_DE data_files: en-de_DE.tsv - config_name: en-kg data_files: en-kg.tsv - config_name: en_US-es_ES data_files: en_US-es_ES.tsv - config_name: en-bg-rBG data_files: en-bg-rBG.tsv - config_name: fr-nl data_files: fr-nl.tsv - config_name: en_GB-as data_files: en_GB-as.tsv - config_name: en-nl data_files: en-nl.tsv - config_name: en-ka-GE data_files: en-ka-GE.tsv - config_name: en-sah data_files: en-sah.tsv - config_name: en_US-ur data_files: en_US-ur.tsv - config_name: und-si data_files: und-si.tsv - config_name: en_devel-en_ca data_files: en_devel-en_ca.tsv - config_name: en-cs-CZ data_files: en-cs-CZ.tsv - config_name: en-de_DIVEO data_files: en-de_DIVEO.tsv - config_name: en-es-PE data_files: en-es-PE.tsv - config_name: en-nb-rNO data_files: en-nb-rNO.tsv - config_name: en_GB-in data_files: en_GB-in.tsv - config_name: en_US-grc data_files: en_US-grc.tsv - config_name: en_GB-ast_ES data_files: en_GB-ast_ES.tsv - config_name: nb_NO-en data_files: nb_NO-en.tsv - config_name: en_devel-zh-cn data_files: en_devel-zh-cn.tsv - config_name: en_US-th data_files: en_US-th.tsv - config_name: en_devel-fa data_files: en_devel-fa.tsv - config_name: en_devel-es_py data_files: en_devel-es_py.tsv - config_name: en-prg data_files: en-prg.tsv - config_name: en_GB-uk_UA data_files: en_GB-uk_UA.tsv - config_name: en-gn data_files: en-gn.tsv - config_name: en-sat data_files: en-sat.tsv - config_name: en-jpn_JP data_files: en-jpn_JP.tsv - config_name: en-ko-rKR data_files: en-ko-rKR.tsv - config_name: en-anp data_files: en-anp.tsv - config_name: en-si_LK data_files: en-si_LK.tsv - config_name: en_GB-gn data_files: en_GB-gn.tsv - config_name: en-kn_IN data_files: en-kn_IN.tsv - config_name: en-b+jbo data_files: en-b+jbo.tsv - config_name: en-me data_files: en-me.tsv - config_name: en-lfn data_files: en-lfn.tsv - config_name: en-cz data_files: en-cz.tsv - config_name: en_GB-iu data_files: en_GB-iu.tsv - config_name: en-uz@cyrillic data_files: en-uz@cyrillic.tsv - config_name: en_US-es-419 data_files: en_US-es-419.tsv - config_name: en_US-ug data_files: en_US-ug.tsv - config_name: es-ext data_files: es-ext.tsv - config_name: en_GB-pa_PK data_files: en_GB-pa_PK.tsv - config_name: en-ast data_files: en-ast.tsv - config_name: en_US-no data_files: en_US-no.tsv - config_name: en-afh data_files: en-afh.tsv - config_name: en-fi-rFI data_files: en-fi-rFI.tsv - config_name: en-ar-rLY data_files: en-ar-rLY.tsv - config_name: en_devel-pt_br data_files: en_devel-pt_br.tsv - config_name: en-ca_ES data_files: en-ca_ES.tsv - config_name: fr-ru data_files: fr-ru.tsv - config_name: en-eo_XX data_files: en-eo_XX.tsv - config_name: en_US-tl data_files: en_US-tl.tsv - config_name: en_GB-gl data_files: en_GB-gl.tsv - config_name: en_UK-es_ES data_files: en_UK-es_ES.tsv - config_name: en-be-rBY data_files: en-be-rBY.tsv - config_name: en-b+hsb data_files: en-b+hsb.tsv - config_name: en_GB-ps data_files: en_GB-ps.tsv - config_name: en-hi-IN data_files: en-hi-IN.tsv - config_name: en-PL data_files: en-PL.tsv - config_name: en_GB-dv data_files: en_GB-dv.tsv - config_name: en_US-sv data_files: en_US-sv.tsv - config_name: en_US-en_AU data_files: en_US-en_AU.tsv - config_name: en_GB-frp data_files: en_GB-frp.tsv - config_name: en_GB-sv-SE data_files: en_GB-sv-SE.tsv - config_name: en-ZH-rCN data_files: en-ZH-rCN.tsv - config_name: en-sq data_files: en-sq.tsv - config_name: en-README_FA data_files: en-README_FA.tsv - config_name: en_devel-ca data_files: en_devel-ca.tsv - config_name: en_UK-fr_FR data_files: en_UK-fr_FR.tsv - config_name: en-zh_Hans data_files: en-zh_Hans.tsv - config_name: en-ar_DZ data_files: en-ar_DZ.tsv - config_name: en-ml data_files: en-ml.tsv - config_name: en-zh-rTW data_files: en-zh-rTW.tsv - config_name: en-uz-Cyrl data_files: en-uz-Cyrl.tsv - config_name: messages-it data_files: messages-it.tsv - config_name: en_devel-ru data_files: en_devel-ru.tsv - config_name: en-es-MX data_files: en-es-MX.tsv - config_name: en_US-zh-Hant-HK data_files: en_US-zh-Hant-HK.tsv - config_name: en-de@formal data_files: en-de@formal.tsv - config_name: en_US-ar-AA data_files: en_US-ar-AA.tsv - config_name: en-en_IE data_files: en-en_IE.tsv - config_name: en_US-de data_files: en_US-de.tsv - config_name: en-eu data_files: en-eu.tsv - config_name: en-tl data_files: en-tl.tsv - config_name: ia-ru data_files: ia-ru.tsv - config_name: en_GB-my data_files: en_GB-my.tsv - config_name: en-Polish data_files: en-Polish.tsv - config_name: en_GB-si data_files: en_GB-si.tsv - config_name: eo-nb_NO data_files: eo-nb_NO.tsv - config_name: en_devel-iw data_files: en_devel-iw.tsv - config_name: en_GB-pt_PT data_files: en_GB-pt_PT.tsv - config_name: en_GB-tt@iqtelif data_files: en_GB-tt@iqtelif.tsv - config_name: en-sk data_files: en-sk.tsv - config_name: es-de data_files: es-de.tsv - config_name: en-enm data_files: en-enm.tsv - config_name: en_US-sk-SK data_files: en_US-sk-SK.tsv - config_name: en_GB-be data_files: en_GB-be.tsv - config_name: nl-en data_files: nl-en.tsv - config_name: en_US-sr_RS data_files: en_US-sr_RS.tsv - config_name: en_GB-cy data_files: en_GB-cy.tsv - config_name: en_devel-es_uy data_files: en_devel-es_uy.tsv - config_name: en-fa-AF data_files: en-fa-AF.tsv language: - aa - ab - ace - ach - af - afh - aii - ain - ajp - ak - am - an - ang - anp - apc - ar - arn - ars - as - ast - ay - ayc - az - azb - ba - bar - bd - be - bem - ber - bg - bho - bm - bn - bo - bp - bqi - br - brx - bs - bul - by - ca - ce - ceb - ckb - cmn - cn - cnr - co - cr - crh - cs - csb - cv - cy - cz - da - de - dev - doi - dsb - dua - dum - dv - dz - eg - el - en - eng - enm - eo - es - et - eu - ext - fa - fi - fil - fo - fr - fra - frm - frp - frs - fu - fur - fy - ga - gb - gd - gl - glk - gmh - gn - gr - gsw - gu - guc - gug - gum - guw - gv - ha - haw - he - hi - hne - hr - hrx - hsb - ht - hu - hy - hz - ia - id - ie - ig - in - io - is - it - iw - ja - jam - jbo - ji - jp - jpn - jv - ka - kab - kg - kk - kl - km - kmr - kn - ko - kok - kr - krl - ks - ksh - ku - kw - ky - la - lb - lfn - lg - li - lk - ln - lo - lt - ltg - lv - lzh - mai - me - mg - mhr - mi - mjw - mk - ml - mn - mnc - mni - mnw - mo - mr - ms - mt - my - na - nah - nan - nap - nb - nds - ne - nl - nn - 'no' - np - nqo - ny - oc - oj - om - or - os - ota - pa - pam - pap - pbb - peo - pk - pl - pms - pr - prg - ps - pt - pu - qt - rcf - rm - ro - rom - ru - rue - rw - ryu - sa - sah - sai - sat - sc - sco - sd - sdh - se - sh - shn - si - sk - skr - sl - sm - sma - sn - so - sq - sr - st - su - sv - sw - szl - ta - tam - te - tet - tg - th - ti - tk - tl - tlh - tn - to - tok - tr - trv - tt - tum - tw - ty - tzm - ua - udm - ug - uk - und - ur - us - uz - vec - vi - vls - wa - wae - wo - xh - yi - yo - yue - zgh - zh - zu task_categories: - translation - text2text-generation pretty_name: Weblate Translations annotations_creators: - crowdsourced size_categories: - 1M<n<10M license: other --- # Dataset Card for Weblate Translations <!-- Provide a quick summary of the dataset. --> A dataset containing strings from projects hosted on [Weblate](https://hosted.weblate.org) and their translations into other languages. Please consider [donating](https://weblate.org/en/donate/) or [contributing](https://weblate.org/en/contribute/) to Weblate if you find this dataset useful. To avoid rows with values like "None" and "N/A" being interpreted as missing values, pass the keep_default_na parameter like this: ``` from datasets import load_dataset dataset = load_dataset("ayymen/Weblate-Translations", keep_default_na=False) ``` ## Dataset Details ### Dataset Description <!-- Provide a longer summary of what this dataset is. --> - **Curated by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** Each sentence pair in the dataset has a corresponding license in the "license" column. This license is the one specified in the component or project containing the sentence. ### Dataset Sources [optional] <!-- Provide the basic links for the dataset. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the dataset is intended to be used. --> - Machine Translation - Language Identification ### Direct Use <!-- This section describes suitable use cases for the dataset. --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. --> [More Information Needed] ## Dataset Structure <!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. --> [More Information Needed] ## Dataset Creation ### Curation Rationale <!-- Motivation for the creation of this dataset. --> [More Information Needed] ### Source Data <!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). --> #### Data Collection and Processing <!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. --> - Sentence pairs with empty/missing elements were dropped. - Identical pairs were dropped. - Trailing whitespace was stripped. - Rows were deduplicated based on all 3 columns including "license", on a config/subset/tsv file basis. Which means that a single config might contain two identical sentence pairs with different licenses. Or a different config/subset might contain the exact same row (most likely a different variant/dialect of the same language(s)). #### Who are the source data producers? <!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. --> [More Information Needed] ### Annotations [optional] <!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. --> #### Annotation process <!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. --> [More Information Needed] #### Who are the annotators? <!-- This section describes the people or systems who created the annotations. --> Weblate users. #### Personal and Sensitive Information <!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations. ## Citation [optional] <!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Dataset Card Authors [optional] [More Information Needed] ## Dataset Card Contact [More Information Needed]
Trofish/Korean-RLHF-Full-process
--- license: cc-by-nc-4.0 task_categories: - reinforcement-learning - text-generation language: - ko tags: - RLHF - SFT - RM - instruction-tuning - reward-model - PPO --- ## KULLM을 baseline으로 RLHF 강화학습을 하는데 사용한 데이터셋입니다. - **Step1: step1_SFT_train.jsonl** (KULLM 12.8B 모델을 Supervised Fine-Tuning 하는데 사용하였습니다.) - **Step2: step2_RM_train.json** (polyglot-ko 1.3B 모델을 Reward Model로 학습하는데 사용하였습니다.) - **Step3: step3_PPO_train.json** (SFT 모델과 RM 모델을 사용하여 RLHF 학습을 하는데 사용하였습니다.) 자세한 정보는 다음을 참고해주세요: https://huggingface.co/Trofish/KULLM-RLHF ## 강화학습 단계별 데이터셋 구축 ![image](https://github.com/VAIV-2023/VAIV2023/assets/79634774/a4988abd-c6fd-4fc2-8e53-9a02240e2275) ![image](https://github.com/VAIV-2023/VAIV2023/assets/79634774/dae49a1e-a834-463c-9f95-34cf254fdaeb) ## 데이터셋 선정 시 고려 사항 - **일상 대화와 혐오 표현 대처 능력을 올리기 위한 데이터셋과, 학습 시 챗봇 모델의 general한 task에 대한 성능이 하락하는 것을 막기 위해서 general task 데이터셋을 구성** - **국립국어원 일상 대화 데이터셋:** 일상적인 대화에 대한 자연스러운 응답이 있으면서도, 맞춤법이 잘 지켜지고 은어, 비문, 초성 등이 없으며 주제별로 다양한 대화가 있음 - **AI Hub 혐오 표현 데이터셋:** 혐오, 차별, 성적인 내용, 폭력, 범죄 등 카테고리별로 다양한 혐오 표현이 있음 - **General task 데이터셋** - Evol-Instruct 데이터셋: 다양한 분야에 대한 복잡하고 논리적인 prompt와 답변이 있음 - Self-Instruct 데이터셋: 사람이 직접 생성한 양질의 Seed data를 기반으로 데이터 증강 - RLHF 한국어 번역 데이터셋: DeepSpeedChat에서 공개한 데이터셋을 한국어로 번역 # Step1. SFT 모델 Fine-tuning ## Baseline Model [- 고려대학교 NLP & AI 연구실과 HIAI 연구소가 개발한 한국어 LLM **"KULLM"** 사용](https://github.com/nlpai-lab/KULLM) ## Datasets ![image](https://github.com/VAIV-2023/VAIV2023/assets/79634774/085610db-3714-43c3-855b-58baad2f4e8b) # Step2. Reward Model ver1 구현 ## Baseline Model - EleutherAI에서 개발한 초거대 한국어 언어 모델 **Polyglot-Ko** 사용 - 1.3b 모델과 5.8b 모델을 각각 실험 ## Datasets ![image](https://github.com/VAIV-2023/RLHF-Korean-Friendly-LLM/assets/79634774/0082da9b-b0b8-4089-8647-cffa5ce724fb) - InstructGPT의 데이터셋 구축 방법 - Reward 모델 학습 데이터셋으로 SFT 학습에 사용한 prompt(1,500개 - 일상대화:혐오표현=2:1)와 새로운 prompt(1,000개 - DeepSpeedChat 번역 데이터셋) 사용 - SFT 모델에서 한개의 prompt당 K개의 Response를 생성하고, 순위를 Labeling - 데이터셋 라벨링 - Instruct GPT의 경우 사람이 직접 Labeling을 하엿지만, 일관된 평가와 시간 단축을 위해 GPt-4와 G-Eval을 이용 - SFT에서 생성한 두 Response 중 G-Eval 평가 점수 합이 높은 것을 Chosen response로 결정 - 데이터셋 유형별로 G-Eval 평가 Prompt에 차이를 두었음 - ![image](https://github.com/VAIV-2023/RLHF-Korean-Friendly-LLM/assets/79634774/7d7117d0-02e9-42dd-8ce3-5244cf726bf8) ## RLFH Model Evaluation ![image](https://github.com/VAIV-2023/VAIV2023/assets/79634774/2b58ed3a-7ed5-4e60-ba4b-c9b291b1fdff) ![image](https://github.com/VAIV-2023/VAIV2023/assets/79634774/75b2a1ee-d7c0-4ba9-ab2f-727abab644e9) ## Final RLHF Model - https://huggingface.co/Trofish/KULLM-RLHF
talrid/CodeContests_valid_and_test_AlphaCodium
--- license: apache-2.0 ---
Wanfq/KCA_data
--- license: cc-by-nc-4.0 language: - en ---
Ejafa/ye-pop
--- license: apache-2.0 language: - en tags: - art pretty_name: ye-pop size_categories: - 100K<n<1M --- # YE-POP (a derived dataset of Laion POP) YE-POP is a derived dataset from Laion-POP, meticulously curated and filtered to enhance the quality and utility of the original dataset. The dataset comprises 11 chunks, each containing 50,000 image URLs from Laion-POP. NSFW sorting has been used as a baseline, and human verification has been conducted to ensure the dataset's reliability. For the initial comparison, Chunk 1 has been curated with Gemini-Pro and released as part of a research work to the community. For access to other chunks generated by gemini-pro, interested parties are encouraged to contact us. The primary goal of YE-POP is to provide a dataset with improved art image descriptions while retaining the essence of Laion-POP for baseline comparisons in diffusion models and image captioning tasks. We anticipate that training multimodal models on this dataset will lead to enhanced generation capabilities. ## Dataset Details Each zip file contains predownloaded images, and the JSON file includes dictionaries of image features with the following fields: - `filename` - `url` - `cogvlm_caption` - `llava_caption` - `nsfw_prediction` - `alt_txt` - `alt_txt_similarity` - `width` - `height` - `original_width` - `original_height` - `exif` For more [detailed information](https://laion.ai/blog/laion-pop/#dataset-and-methodology) on the fields, refer to the JSON file. ## Dataset Card Authors [Yaroslav Ponomarenko]() [Ejafa Bassam]() ## Dataset Card Contact @[Peking University](https://cs.pku.edu.cn/English/Home.htm) ## Acknowledgments [Laion (Christoph Schuhmann, Peter Bevan)]() [Google Gemini-Pro](https://doi.org/10.48550/arXiv.2312.11805)
p208p2002/zhtw-sentence-error-correction
--- language: - zh configs: - config_name: alpha data_files: - split: train path: "alpha/out.jsonl" - config_name: beta data_files: - split: train path: "beta/out.jsonl" - config_name: gamma data_files: - split: train path: "gamma/out.jsonl" --- # 中文錯字糾正資料集 由規則與字典自維基百科產生的錯誤糾正資料集。 包含錯誤類型:隨機錯字、近似音錯字、缺字錯誤、冗字錯誤。 資料集使用函式庫: [p208p2002/zh-mistake-text-gen](https://github.com/p208p2002/zh-mistake-text-gen) ### 子集 - alpha: 95%錯誤,5%不變。單句中可能有多個錯誤。 - beta: 50%錯誤,50%不變。單句中僅有一個錯誤。 - gamma: 100%錯誤。單句中可能有多個錯誤。
tastypear/bluemoon-cleaned-lewd
--- license: apache-2.0 task_categories: - text-generation language: - en tags: - not-for-all-audiences --- 这个数据集从 grimulkan/bluemoon_Karen_cleaned 中抽取了所有包含NSFW内容的对话(只保留第一轮对话)。 This dataset extracts all conversations containing NSFW content from `grimulkan/bluemoon_Karen_cleaned` (only the first round of conversations is retained) Explanation of long_response: ```python if len(chosen) > len(prompt): long_response = 1 ```
linux-cn/archive
--- license: cc-by-nc-4.0 language: - zh pretty_name: Linux 中国文章数据集 size_categories: - 10K<n<100K --- --- license: cc-by-nc-4.0 --- # Linux 中国原创文章/译文数据集 这个数据集为 Linux 中国原创技术文章 + 翻译技术文章的数据集,提供了文章标题、内容等多个字段。 ## Dataset Details ### Dataset Description - **Language(s) (NLP):** 中文 - **License:** cc-by-nc-4.0 ## Dataset Structure - id:文章ID - title: 文章标题 - author: 文章作者 - fromurl: 文章源地址(仅翻译类文章有) - summary: 总结 - excerpt: 摘要 - pic: 头图(缩略图版) - largepic:头图(大图版) - titlepic:是否有头图,可以渲染用。 - islctt:是否是 LCTT 文章(翻译文章) - selector:选题人员,值为 Github ID - translator:翻译人员,值为 Github ID - reviewer:校对人员,值为 Github ID - tags:文档标签 - category:文档所属目录 - count:计数 - viewnum: 访问量 - commentnum: 评论量 - favtimes: 收藏量 - sharetimes: 分享量 - likes: 喜欢量 - comments_data: 评论数据 - postip:评论 IP - dateline:评论时间 - message:评论内容 - username:评论名 - repcids:回复的评论的 ID - related: 相关文章的 ID - date:发布日期 - updated:最后更新日期 - permalink:永久链接(Linux.cn 上的链接) - content:文章内容 ## 图片素材下载 & Markdown 版下载 上述文章内容中的图片素材可访问下方链接获取,attachment.zip 即为对应的图片。 ### GitHub https://github.com/Linux-CN/archive/releases/tag/release ### 百度网盘 链接: https://pan.baidu.com/s/1i7DTuf_umPkkleHFtdmZJA?pwd=lccn 提取码: lccn ## Dataset Card Contact bestony <bestony@linux.com>
botbot-ai/aya_dataset_pt
--- language: - pt pretty_name: Aya Dataset Portuguese tags: - aya - portuguese - legal - chemistry license: apache-2.0 size_categories: - 1K<n<10K --- CohereForAI [Aya Dataset](https://huggingface.co/datasets/CohereForAI/aya_dataset) filtrado para português (PT). **Aya Dataset Summary** The [Aya Dataset](https://huggingface.co/datasets/CohereForAI/aya_dataset) is a multilingual instruction fine-tuning dataset curated by an open-science community via Aya Annotation Platform from Cohere For AI. The dataset contains a total of 204k human-annotated prompt-completion pairs along with the demographics data of the annotators. This dataset can be used to train, finetune, and evaluate multilingual LLMs. Curated by: Contributors of Aya Open Science Intiative. Language(s): 65 languages (71 including dialects & scripts). License: Apache 2.0
Jiwonny29/project1
--- license: apache-2.0 task_categories: - feature-extraction language: - en tags: - biology pretty_name: test size_categories: - 100K<n<1M dataset_info: config_name: mydata features: - name: Year dtype: int32 - name: LocationAbbr dtype: string - name: LocationDesc dtype: string - name: Latitude dtype: float32 - name: Longitude dtype: float32 - name: Disease_Type dtype: int32 - name: Data_Value_Type dtype: int32 - name: Data_Value dtype: float32 - name: Break_Out_Category dtype: string - name: Break_Out_Details dtype: string - name: Break_Out_Type dtype: int32 - name: Life_Expectancy dtype: float32 --- ### Dataset Description This dataset encompasses mortality rates for cardiovascular (CVD) and heart diseases across the United States, covering both state-specific and national levels, from 2000 to 2020. The mortality rate is quantified as the number of deaths per 100,000 individuals annually in the US. The dataset is structured to classify mortality rates according to various demographic factors, including overall rates, gender (female, male), race (white, black, Hispanic, other), and age groups (18-24, 25-44, 45-65, 65+). Additionally, life expectancy data for each state is incorporated in the dataset. For ease of use, I combined the data on a five-year interval rather than an annual basis. ### Dataset Sources - CVD Mortality Data: Centers for Disease Control and Prevention(CDC) National Vital Statistics System - https://data.cdc.gov/Heart-Disease-Stroke-Prevention/National-Vital-Statistics-System-NVSS-National-Car/kztq-p2jf/about_data - Life Expectancy Data: Institute for Health Metrics and Evaluation - https://ghdx.healthdata.org/record/ihme-data/united-states-life-expectancy-by-county-race-ethnicity-2000-2019 ## Uses This dataset serves as a valuable resource for researchers and individuals interested in examining and identifying patterns related to cardiovascular diseases in the United States. It can be utilized to forecast future fatalities caused by heart diseases by leveraging similar features present in the dataset. Additionally, the dataset enables users to gain insights into identifying states that require assistance and support in reducing mortality rates. Below are example use cases and corresponding codes: - Analyzing the comprehensive picture of mortality and conducting time series analysis on mortality rates - https://colab.research.google.com/drive/1ulygrSt9jt3x_4WIGD6QdK0TcGZlpuYF - Building regression models - https://colab.research.google.com/drive/1DhIni026qz5qqjfWwKXnqoQXDy-HzroC - Developing a web application for users to quickly understand and compare mortality rates among states, along with relevant information like state population - https://github.com/jiwonny29/Exploring_US_Cardiovascular_Mortality_Trends_via_Streamlit ## Dataset Structure This dataset contains - Year (int32): This column contains the year of the data record, with values ranging from 2000 to 2020 - LocationAbbr (String): Abbreviation representing the location, typically a state - LocationDesc (String): The full name or detailed description of the location - Latitude (float32) : Geographic coordinate that specifies the north-south position of a point on the Earth's surface - Longitude (float32) : Geographic coordinate that specifies the east-west position of a point on the Earth's surface - Geolocation (Tuple): A pair of latitude and longitude coordinates, formatted as (latitude, longitude), providing the geolocation or geocode of the location - Disease_Type (int32): A key column in the dataset, representing eight unique types of cardiovascular diseases, numbered from 0 to 7. The values correspond to the following diseases: - 0: Major Cardiovascular Disease - 1: Diseases of the Heart (Heart Disease) - 2: Acute Myocardial Infarction (Heart Attack) - 3: Coronary Heart Disease - 4: Heart Failure - 5: Cerebrovascular Disease (Stroke) - 6: Ischemic Stroke - 7: Hemorrhagic Stroke - Data_Value_Type (int32): Represents the type of data value. "Age-Standardized" is represented by 1, and "Crude" is represented by 0, indicating the measurement methods for the data value columns - Data_Value (float32): This column represents the number of deaths per 100,000 population - Break_Out_Category (string): This category is used for breaking down the data and includes four unique values: "Overall," "Gender," "Age," and "Race." - Break_Out_Details (string): Specific subcategories within the Break_Out_Category. This column includes values like "Overall," six age categories (e.g., "18-24," "25-44"), two gender categories (e.g., "Female," "Male"), and four race categories (e.g., "Hispanic," "Non-Hispanic Black," "Non-Hispanic White," "Other"). - Break_Out_Type (int32): A numerical transformation of the Break_Out_Details column. In this system, "Overall" is represented as 0, "Male" and "Female" as 1 and 2, respectively; age groups "18-24," "25-44," "45-64," "65+" as 1, 2, 3, 4, respectively; and racial categories "Non-Hispanic White," "Non-Hispanic Black," "Hispanic," "Other" as 1, 2, 3, 4, respectively. - Life_Expectancy (float32): Represents the life expectancy for the applicable year and state
cnmoro/WizardVicuna-PTBR-Instruct-Clean
--- license: apache-2.0 ---
ShixuanAn/RDD_2020
--- license: cc-by-nc-3.0 task_categories: - image-classification language: - en pretty_name: >- RDD2020: An Image Dataset for Smartphone-based Road Damage Detection and Classification size_categories: - 100B<n<1T --- # Dataset Card for RDD_2020 The RDD2020 dataset is a comprehensive collection of 26,336 road images from India, Japan, and the Czech Republic, annotated with over 31,000 instances of road damages. This dataset is designed to support the development and evaluation of machine learning models for automatic road damage detection, offering a valuable resource for municipalities and road agencies for efficient road condition monitoring. ## Dataset Details ### Dataset Description - **Source:** [Mendeley Data](https://data.mendeley.com/datasets/5ty2wb6gvg/1) - DOI: 10.17632/5ty2wb6gvg.1 - **Size:** 1.13 GB - **Format:** Images (JPEG) and Annotations (XML in PASCAL VOC format) - **Resolution:** - India: 720 × 720 pixels - Japan and Czech: 600 × 600 pixels - **Categories:** Longitudinal Cracks (D00), Transverse Cracks (D10), Alligator Cracks (D20), Potholes (D40) - **License:** https://creativecommons.org/licenses/by/4.0/ ### Dataset Sources [optional] <!-- Provide the basic links for the dataset. --> - **Homepage** https://data.mendeley.com/datasets/5ty2wb6gvg/1 - **Data article:** https://doi.org/10.1016/j.dib.2021.107133 ## Uses <!-- Address questions around how the dataset is intended to be used. --> ### Direct Use <!-- This section describes suitable use cases for the dataset. --> RDD2020 dataset can be directly used for developing and benchmarking machine learning models aimed at automatic detection and classification of road damages. This includes developing new deep learning architectures or modifying existing ones to improve detection accuracy across different types of road damages ## Dataset Structure ### Data Instance The data will follow the structure below: ``` { "image_id": "Czech_000248", "country": "Czech", "type": "train", "image": "<PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=600x600>", "image_path": "train/Czech/images/Czech_000248.jpg", "crack_type": ["D20", "D20"], "crack_coordinates": { "x_min": [188, 3], "x_max": [309, 171], "y_min": [463, 438], "y_max": [509, 519] } } ``` ### Data Fields - "image_id"[string]: ID of the image, created by combining the country plus a sequential number. - "country"[string]: The country where the photo was taken. - "type"[string]: The dataset category the image belongs to, such as 'train', 'test1', or 'test2'. "image"[integer]: The image data converted into PIL format. - "crack_type"[string]: Types of cracks detected in the image. - "crack_coordinates"[integer]: Contains crack coordinates as integers. ## Dataset Creation ### Curation Rationale <!-- Motivation for the creation of this dataset. --> The RDD2020 dataset was curated with the objective of facilitating the development, testing, and benchmarking of machine learning models for road damage detection, catering specifically to the needs of municipalities and road agencies. A significant aspect of the dataset's curation process was the conversion of images into the Python Imaging Library (PIL) format and the meticulous parsing of XML annotations to ensure a seamless integration between the image data and the associated labels. This conversion process was driven by the need to simplify the handling of image data for machine learning applications, as the PIL format is widely supported by data processing and model training frameworks commonly used in the field. Additionally, the parsing of XML files to extract detailed annotations about the type and coordinates of road damages allows for precise labeling of the data. This approach ensures that each image is directly associated with its corresponding damage type and location. The dataset's diversity, with images sourced from three different countries, aims to enable the creation of robust models that are effective across various environmental conditions and road infrastructures, thereby broadening the applicability and relevance of the trained models. #### Data Collection and Processing <!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. --> Road images (.jpg) were collected using a vehicle-mounted smartphone, moving at an average speed of about 40Km/h. XML files were created using the LabelImg tool to annotate the road damages present in the images. #### Who are the source data producers? Deeksha Arya, Hiroya Maeda, Sanjay Kumar Ghosh, Durga Toshniwal, Hiroshi Omata, Takehiro Kashiyama, Toshikazu Seto, Alexander Mraz, Yoshihide Sekimot <!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. --> ### Annotations <!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. --> #### Annotation process <!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. --> Each image in the dataset comes with corresponding XML files containing annotations in PASCAL VOC format. These annotations describe the location and type of road damages present in the images, categorized into four main types: Longitudinal Cracks (D00), Transverse Cracks (D10), Alligator Cracks (D20), and Potholes (D40). ### Social Impact The structuring of the RDD2020 dataset into a more accessible and usable format is aimed at having a focused and measurable impact on the management of road infrastructure. The transformation of raw images and XML annotations into a coherent dataset with clearly defined attributes such as photo_id, country, type, pics_array, image_resolution, crack_type, and crack_coordinates creates a powerful tool for municipalities and road agencies. With this structured dataset, these entities can deploy machine learning models to accurately identify and classify road damages like cracks and potholes, which are critical for the maintenance and safety of roadways. In conclusion, the transformation of this raw data into a structured and accessible format not only catalyzes the progress of automated road damage assessment but also potentially engages the public sector in adopting AI-driven solutions for public safety and infrastructure management. ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> The dataset primarily includes images from three countries (India, Japan, and the Czech Republic), which may not fully represent road conditions worldwide. Users should be cautious when generalizing models trained on this dataset to other regions. ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
OpenGVLab/AS-100M
--- license: apache-2.0 --- # AS-100M AS-100M is a subset of AS-1B. We release this dataset in both [COCO format](https://huggingface.co/datasets/Weiyun1025/AS-100M/tree/main/coco_format) and [JSONL format](https://huggingface.co/datasets/Weiyun1025/AS-100M/tree/main/jsonl_format). ***NOTE***: The bbox format in the COCO format is `xywh`, while in the JSONL format, it is `x1y1x2y2`. ## Introduction We present the All-Seeing Project with: [***All-Seeing 1B (AS-1B) dataset***](https://huggingface.co/datasets/Weiyun1025/AS-100M): we propose a new large-scale dataset (AS-1B) for open-world panoptic visual recognition and understanding, using an economical semi-automatic data engine that combines the power of off-the-shelf vision/language models and human feedback. [***All-Seeing Model (ASM)***](https://huggingface.co/Weiyun1025/All-Seeing-Model-FT): we develop a unified vision-language foundation model (ASM) for open-world panoptic visual recognition and understanding. Aligning with LLMs, our ASM supports versatile image-text retrieval and generation tasks, demonstrating impressive zero-shot capability. <img width="820" alt="image" src="https://github.com/OpenGVLab/all-seeing/assets/8529570/e43ab8db-6437-46f1-8aa1-c95f012e9147"> Figure 1: Overview and comparison of our All-Seeing project with other popular large foundation models. <!-- ## Online Demo **All-Seeing Model demo** is available [here](https://openxlab.org.cn/apps/detail/wangweiyun/All-Seeing-Model-Demo). **Dataset Browser** is available [here](https://openxlab.org.cn/apps/detail/wangweiyun/All-Seeing-Dataset-Browser). https://github.com/OpenGVLab/all-seeing/assets/47669167/9b5b32d1-863a-4579-b576-b82523f2205e --> ## Dataset Overview AS-1B with over 1 billion regions annotated with semantic tags, question-answering pairs, and detailed captions. It covers a wide range of 3.5 million common and rare concepts in the real world, and has 132.2 billion tokens that describe the concepts and their attributes. <img width="800" alt="image" src="https://github.com/OpenGVLab/all-seeing/assets/8529570/adac37ed-312f-4f11-ba8a-6bc62067438f"> Some examples <img width="800" alt="image" src="https://github.com/OpenGVLab/all-seeing/assets/8529570/fcf6ab07-c4ba-441c-aa6c-111c769f75b1"> Please see our [paper](https://arxiv.org/abs/2308.01907) to learn more details. ## Model Architecture The All-Seeing model (ASM) is a unified framework for panoptic visual recognition and understanding, including image/region-text retrieval, image/region recognition, captioning, and question-answering. <img width="820" alt="image" src="https://github.com/OpenGVLab/all-seeing/assets/8529570/8995e88c-6381-452f-91e4-05d68a2795fc"> ## License This project is released under the [Apache 2.0 license](LICENSE). # Citation If you find our work useful in your research, please consider cite: ```BibTeX @article{wang2023allseeing, title={The All-Seeing Project: Towards Panoptic Visual Recognition and Understanding of the Open World}, author={Wang, Weiyun and Shi, Min and Li, Qingyun and Wang, Wenhai and Huang, Zhenhang and Xing, Linjie and Chen, Zhe and Li, Hao and Zhu, Xizhou and Cao, Zhiguo and others}, journal={arXiv preprint arXiv:2308.01907}, year={2023} } @article{wang2024allseeing_v2, title={The All-Seeing Project V2: Towards General Relation Comprehension of the Open World}, author={Wang, Weiyun and Ren, Yiming and Luo, Haowen and Li, Tiantong and Yan, Chenxiang and Chen, Zhe and Wang, Wenhai and Li, Qingyun and Lu, Lewei and Zhu, Xizhou and others}, journal={arXiv preprint arXiv:2402.19474}, year={2024} } ```
microsoft/Taskbench
--- language: - en license: mit tags: - agent - tool-learning - task-automation - LLM pretty_name: TaskBench size_categories: - n<21k configs: - config_name: huggingface data_files: - split: test path: "data_huggingface/data.json" - config_name: multimedia data_files: - split: test path: "data_multimedia/data.json" - config_name: dailylifeapis data_files: - split: test path: "data_dailylifeapis/data.json" --- <p align="center"> <img src="./assets/logo2.png" width="10%"> </p> <div align="center"> <!-- <h1>TaskBench</h1> --> <!-- <div align="center"> <a href="https://opensource.org/licenses/Apache-2.0"> <img alt="License: Apache 2.0" src="https://img.shields.io/badge/License-Apache%202.0-4E94CE.svg"> </a> <a href="https://arxiv.org/abs/2311.18760"> <img alt="License: Apache 2.0" src="https://img.shields.io/badge/arXiv-Paper-<COLOR>.svg"> </a> </div> --> <h3>TaskBench: Benchmarking Large Language Models for Task Automation<h3> </div> <p align="center"> <img width="100%" alt="image" src="./assets/eval.png"> </p> ## Table of Contents + [Introduction](#introduction) + [Dataset](#dataset) + [Evaluation with TaskBench](#evaluation-with-taskbench) + [Dataset Construction with Back-Instruct](#dataset-construction-with-back-instruct) + [Leaderboard](#leaderboard) + [Citation](#citation) ## Introduction TaskBench is a benchmark for evaluating large language models (LLMs) on task automation. Task automation can be formulated into three critical stages: task decomposition, tool invocation, and parameter prediction. This complexity makes data collection and evaluation more challenging compared to common NLP tasks. To address this challenge, we propose a comprehensive evaluation framework and a high-quality dataset for task automation. We also provide a leaderboard of 17 LLMs on TaskBench, including GPT-4, Claude-2, and other open-source LLMs. ### Dataset To generate high-quality evaluation datasets, we introduce the concept of Tool Graph to represent the decomposed tasks in user intent, and adopt a Back-Instruct method to simulate user instruction and annotations. The data collection process consists of three stages: + **Tool Graph Construction:** we first build a tool library and use the tool library to construct the tool graph. The nodes in the tool graph represent the tools, and the edges represent the dependencies between the tools, including the resource dependency and temporal dependency. + **Graph Sampling:** we sample the tool graph to generate the tool graph for each sample. The sampled tool graph is used to generate the tool invocation graph and the instruction. According to the topology of the sampled tool graph, we sample the tool graph in three ways: node, chain and DAGs, which represent different structures of task decomposition for task automation. + **Back-Instruct:** we first use the sampled tool graph to generate the task steps and the instructions. Then, we use the instruction to generate the tool invocation parameters to complete the tool invocation graph. <p align="center"> <img width="100%" alt="image" src="./assets/backinstruct.png"> </p> To improve the quality of the dataset, we use LLM-based and rule-based critics to verify the dataset. The former aims to use LLM to check the alignments between the generated data and the sampled tool graph. While the latter uses straightforward rules to determine the alignment between the tool graphs in created data and the sampled tool graphs. Here, we use the nodes and edges of the sampled graph to determine the consistency. Details statistics of the processing are shown in [the table](#LLM-based-and-Rule-based-Critics). After LLM-based and rule-based critics, we further verify the dataset with human annotators, including checking the syntax of the instructions, the correctness of the tool invocation graph, and the correctness of the tool invocation parameters. The final dataset contains 28,271 samples in three domains: HuggingFace Tools, Multimedia Tools, and Daily Life APIs. Details statistics of the human verification are shown in [the table](#Human-Verification). #### Introduction The TaskBench dataset contains datasets in three areas: HuggingFace Tools, Multimedia Tools, and Dailylife APIs. Each dataset directory includes three files: + `data.json`: the dataset file, which contains the samples in the dataset. + `graph_desc.json`: the tool graph description file, which contains the tool graph of the dataset. + `user_requests.json`: contains the user requests of the dataset. + `tool_desc.json`: the tool description file, which contains the tool descriptions of the dataset. ``` ├─data_dailylifeapis │ data.json │ graph_desc.json │ user_requests.json │ tool_desc.json │ ├─data_huggingface │ data.json │ graph_desc.json │ user_requests.json │ tool_desc.json │ └─data_multimedia data.json graph_desc.json user_requests.json tool_desc.json ``` #### Processing Statistics We provide the statistics of the dataset processing in the following tables: + **Overview**: we provide the number of samples in each dataset, the number of samples checked by critics, and the number of samples verified by humans. Grouped by the tool invocation graph structure, e.g. node, chain, and DAGs, we also provide the number of samples in each group. + **LLM-based and Rule-based Critics**: we provide the number of samples checked by LLM-based critics, rule-based critics and both critics. + **Human Verification**: Human verification is built on the samples checked by critics, which includes three parts: syntax checking, instruction checking, and tool invocation graph checking. We provide the number of samples in each part, and along with the number of samples that are discarded or fixed. | Dataset | #Samples | #Samples Checked by Critics (%) | #Samples Verified by Humans (%) | Node | Chain | DAG | | :-----: | :------: | :----------------: | :--------------: | :------: | :------: | :------: | | Hugging Face Models | 12,217 | 8,457 (69.22%) | 7,546 (61.76%) | 3,067 | 3,642 | 837 | | Multimedia Tools | 8,904 | 6,281 (70.54%) | 5,584 (62.71%) | 2,037 | 2,982 | 565 | | Dailylife APIs | 7,150 | 5,432 (75.97%) | 4,320 (60.42%) | 1,258 | 2,787 | 275 | <div id="LLM-based-and-Rule-based-Critics"> | Dataset | #Samples | #Checked by LLM-based Critics (%) | #Checked by Rule-based Critics (%) | #Checked by Both Critics (%) | | :-----: | :------: | :-----------------------------: | :------------------------------: | :-------------------------: | | Hugging Face Models | 12,217 | 9,042 (74.01%) | 10,289 (84.22%) | 8,457 (69.22%) | | Multimedia Tools | 8,904 | 6,959 (78.16%) | 7,363 (82.69%) | 6,281 (70.54%) | | Dailylife APIs | 7,150 | 5,694 (79.63%) | 6,271 (87.70%) | 5,432 (75.97%) | <div id="Human-Verification"> | Dataset | #Samples Checked by Critics | #Correct Samples (%) | #Discarded (%) | #Fixed for Syntax (%) | #Fixed for Instructions (%) | #Fixed for Tool Invocation Graph (%) | | :-----: | :-------------------------: | :-------------------: | :-------------------: | :---------------------------: | :-----------------------------------: | :------------: | | Hugging Face Models | 8,457 | 6,974 (82.46%) | 911 (10.77%) | 27 (0.32%) | 328 (3.87%) | 843 (9.96%) | | Multimedia Tools | 6,281 | 5,262 (83.77%) | 697 (11.09%) | 11 (0.17%) | 107 (1.70%) | 526 (9.96%) | | Dailylife APIs | 5,432 | 4,307 (79.29%) | 714 (13.14%) | 6 (0.11%) | 92 (1.68%) | 332 (6.11%) | ## Evaluation with TaskBench On top of the TaskBench dataset, we provide a comprehensive evaluation framework for task automation. The evaluation framework consists of three stages: task decomposition, tool invocation, and parameter prediction. We provide the evaluation metrics for each stage: + **Task Decomposition**: Since task steps are diverse text distributions, we use the Rouge-1 (R1), Rouge-2 (R2), and Bertscore F1 (BsF) metrics to evaluate the task decomposition results. + **Tool Invocation**: We report the F1 of node prediction (n-F1) and edge prediction (e-F1) in the tool invocation graph to evaluate the tool invocation results. Edge prediction reflects the correctness of the dependencies between tools, while node prediction reflects the correctness of the tool prediction. + **Parameter Prediction**: For tool parameters prediction, we report the parameter type (or name) F1 (t-F1) and parameter value F1 (v-F1). To evaluate the task automation performance of LLMs on TaskBench we provide the evaluation code and data, please follow the instructions below: ### Setup ```bash conda create -n taskbench python=3.8 conda activate taskbench pip install -r requirements.txt ``` Additionally, if you wish to evaluate open-source large language models, you will also need to deploy the LLMs locally using an **OpenAI-compatible API**. We recommend using the `fastchat` tool to deploy the service to the `localhost:8000` endpoint. ```bash pip install fastchat pip install vllm pip install "fastapi[all]" python3 -m fastchat.serve.controller python3 -m fastchat.serve.vllm_worker --model-path lmsys/vicuna-7b-v1.3 python3 -m fastchat.serve.openai_api_server --host localhost --port 8000 ``` ### Inference For convenience, it is recommended to deploy all LLMs to the same endpoint, such as `localhost:8000`. To generate the prediction file on TaskBench, specify the name of the LLM using the following command: ```bash python inference.py \ --llm gpt-4 \ --data_dir data_multimedia \ --temperature 0.2 \ --top_p 0.1 \ --api_addr localhost \ --api_port 8000 \ --multiworker 5 \ --use_demos 0 \ --reformat true \ --reformat_by self \ --log_first_detail true \ --use_demos 2 \ --dependency_type resource \ --tag true ``` ### Evaluation With the predictions in place, you can now evaluate the LLMs. The predictions file is saved by default in the dataset's folder under the name `predictions`. Execute the following command to calculate the evaluation metrics (saved in the `metrics` folder): ```bash python evaluate.py \ --data_dir data_multimedia \ --prediction_dir $prediction_dir \ --llm gpt-4 \ --splits all \ --n_tools all \ --mode add \ --dependency_type resource \ -m all ``` ## Dataset Construction with Back-Instruct We have provided the dataset for three domains: Hugging Face Tools (`data_huggingface`), Multimedia Tools (`data_multimedia`), and Daily Life APIs (`data_dailylifeapis`). If you want to generate your own dataset, please follow the instructions below: ### Construct Your Own Tool Graph First, you need to build your own tool library. The tool library is a JSON file that contains the description of the tools and tool parameters. Two formats of the tool are supported: ```json // Tool with type-specific parameters { "id": "Image-to-Image", "desc": "Image-to-image is the task of transforming a source image to match the characteristics of a target image or a target image domain. Any image manipulation and enhancement is possible with image to image models.", "input-type": [ "image" ], "output-type": [ "image" ] } // API with request parameters { "id": "send_sms", "desc": "Send an sms to a specific phone number", "parameters": [ { "name": "phone_number", "type": "string", "desc": "The phone number to send the sms to" }, { "name": "content", "type": "string", "desc": "The content of the sms" } ] } ``` Then based on the tool library, you can use the script `generate_graph.py` to generate the tool graph. Now we support two type of tool graph: resource dependency graph and temporal dependency graph. For type-specific parameters, we use the resource dependency graph. For API with request parameters, we use the temporal dependency graph. You can specify the tool graph type by the parameter `--dependency_type`. In the future, we will support more types of tool graphs. ```bash python generate_graph.py \ --tool_desc tool_desc.json \ --dependency_type resource \ --data_dir data_multimedia ``` > Note: The auto-generated tool graph may not be perfect. You can manually modify the tool graph to make it more reasonable. You can check the tool graph through the visualization tool `visualize_graph.py`. We recommend that you manually create the tool graph thoroughly, which will help you to generate a high-quality dataset. ### Generate the Dataset After generating the tool graph, you can use the script `data_engine.py` to generate the dataset. You need to specify the tool graph description file to `--graph_desc` and the tool description file to `--tool_desc`. ```bash # specify the graph and tool description file python data_engine.py \ --graph_desc data_multimedia/graph_desc.json \ --tool_desc data_multimedia/tool_desc.json \ --llm gpt-4 \ --temperature 1.0 \ --top_p 1.0 \ --dependency_type resource \ --save_figure false \ --api_addr localhost \ --api_port 8002 \ --check true \ --use_async true \ --multiworker 5 python format_data.py \ --data_dir data_multimedia \ --dependency_type resource ``` ## Leaderboard Based on the evaluation framework and the TaskBench dataset, we provide a leaderboard of task automation performance of 17 LLMs. We provide the evaluation results of each LLM in the following tables: ### Multimedia Tools Domain | LLM | R1 | R2 | BsF | n-F1 | e-F1 | t-F1 | v-F1 | |----------------------|-------|-------|------|------|------|------|------| | gpt-4 | 60.84 | 40.08 | 91.19 | 90.90 | 69.27 | 87.06 | 72.31 | | claude-2 | 48.85 | 23.59 | 89.22 | 80.94 | 53.01 | 71.63 | 51.58 | | gpt-3.5-turbo | 49.66 | 28.51 | 89.54 | 72.83 | 44.02 | 65.91 | 40.80 | | text-davinci-003 | 49.23 | 27.97 | 89.21 | 73.97 | 45.81 | 68.48 | 40.70 | | codellama-13b | 44.46 | 23.30 | 88.66 | 62.78 | 24.61 | 48.19 | 29.13 | | codellama-7b | 43.76 | 22.93 | 88.81 | 53.29 | 14.76 | 38.04 | 24.45 | | vicuna-13b-v1.5 | 44.75 | 23.75 | 88.94 | 60.61 | 14.78 | 41.62 | 23.62 | | nous-hermes-13b | 35.73 | 16.11 | 87.53 | 58.97 | 8.90 | 43.60 | 21.69 | | wizardlm-13b | 35.87 | 17.55 | 87.29 | 51.24 | 4.82 | 39.10 | 18.74 | | vicuna-7b-v1.5 | 39.46 | 19.83 | 88.53 | 46.06 | 4.26 | 29.72 | 13.74 | | longchat-7b-v1.5 | 37.85 | 18.14 | 87.64 | 43.08 | 3.95 | 27.89 | 13.41 | | baichuan-13b-chat | 20.41 | 3.77 | 83.31 | 42.51 | 5.19 | 28.04 | 11.77 | | llama-2-13b-chat | 26.16 | 7.88 | 84.82 | 43.87 | 1.63 | 29.99 | 11.32 | | internlm-chat-7b | 16.64 | 3.56 | 82.91 | 23.60 | 1.14 | 13.75 | 6.09 | | llama-2-7b-chat | 34.51 | 15.91 | 87.56 | 26.47 | 0.91 | 18.27 | 5.84 | | mpt-7b-chat | 30.94 | 11.90 | 86.08 | 8.68 | 0.18 | 3.19 | 1.02 | | vicuna-33b-v1.3 | 31.27 | 13.37 | 86.17 | 6.40 | 0.01 | 2.47 | 1.09 | ### HuggingFace Tools Domain | LLM | R1 | R2 | BsF | n-F1 | e-F1 | t-F1 | v-F1 | |----------------------|-------|-------|------|------|------|------|------| | gpt-4 | 52.42 | 30.38 | 90.12 | 81.54 | 54.70 | 77.31 | 60.86 | | claude-2 | 44.21 | 21.12 | 88.71 | 79.00 | 43.51 | 63.00 | 43.08 | | text-davinci-003 | 36.68 | 17.61 | 87.03 | 59.38 | 29.37 | 52.53 | 36.04 | | gpt-3.5-turbo | 42.99 | 21.58 | 88.47 | 69.49 | 33.36 | 55.88 | 36.32 | | codellama-13b | 38.75 | 18.37 | 88.32 | 53.16 | 14.64 | 32.06 | 18.87 | | nous-hermes-13b | 37.36 | 16.91 | 88.18 | 53.62 | 8.29 | 37.51 | 17.66 | | wizardlm-13b | 34.47 | 15.38 | 87.38 | 54.40 | 2.05 | 38.76 | 15.35 | | llama-2-13b-chat | 39.37 | 18.64 | 88.67 | 48.47 | 7.30 | 31.61 | 15.38 | | longchat-7b-v1.5 | 27.09 | 8.97 | 85.50 | 48.18 | 0.56 | 33.57 | 13.94 | | baichuan-13b-chat | 19.93 | 5.97 | 83.85 | 53.85 | 7.65 | 33.17 | 13.53 | | vicuna-13b-v1.5 | 37.12 | 17.03 | 87.90 | 50.82 | 7.28 | 28.34 | 11.85 | | vicuna-7b-v1.5 | 27.17 | 10.02 | 85.61 | 42.87 | 2.76 | 24.65 | 10.81 | | vicuna-33b-v1.3 | 33.52 | 14.75 | 86.73 | 43.40 | 4.82 | 22.71 | 10.07 | | codellama-7b | 38.97 | 18.62 | 88.46 | 37.59 | 5.35 | 22.50 | 9.20 | | internlm-chat-7b | 20.53 | 7.16 | 83.74 | 24.39 | 0.83 | 15.41 | 6.64 | | llama-2-7b-chat | 24.12 | 8.68 | 85.43 | 27.30 | 0.74 | 13.05 | 2.79 | | mpt-7b-chat | 33.21 | 12.73 | 87.23 | 20.86 | 0.12 | 9.61 | 1.83 | ### Daily Life APIs Domain | LLM | R1 | R2 | BsF | n-F1 | e-F1 | t-F1 | v-F1 | |----------------------|-------|-------|------|------|------|------|------| | gpt-4 | 85.07 | 72.36 | 96.91 | 96.91 | 80.53 | 97.02 | 71.14 | | claude-2 | 82.26 | 69.88 | 96.64 | 93.52 | 75.31 | 92.71 | 64.72 | | codellama-13b | 89.86 | 83.27 | 97.90 | 87.73 | 63.16 | 84.26 | 62.38 | | gpt-3.5-turbo | 58.53 | 39.90 | 91.29 | 85.37 | 60.67 | 81.97 | 55.66 | | text-davinci-003 | 68.27 | 50.30 | 93.59 | 80.42 | 54.90 | 78.37 | 53.40 | | nous-hermes-13b | 78.49 | 68.04 | 95.61 | 73.45 | 3.50 | 64.47 | 47.22 | | vicuna-13b-v1.5 | 81.76 | 71.76 | 96.31 | 75.67 | 12.48 | 64.27 | 47.31 | | wizardlm-13b | 82.02 | 72.43 | 96.36 | 69.34 | 14.18 | 55.00 | 40.53 | | codellama-7b | 56.98 | 38.83 | 91.31 | 59.33 | 27.23 | 52.99 | 34.81 | | vicuna-33b-v1.3 | 54.96 | 39.71 | 91.40 | 52.49 | 16.37 | 39.95 | 29.64 | | vicuna-7b-v1.5 | 40.26 | 21.19 | 87.27 | 52.73 | 14.23 | 36.30 | 24.67 | | baichuan-13b-chat | 49.43 | 27.25 | 88.32 | 52.55 | 10.61 | 37.48 | 23.77 | | llama-2-13b-chat | 45.39 | 22.42 | 87.74 | 55.77 | 17.02 | 35.11 | 22.94 | | longchat-7b-v1.5 | 29.05 | 14.84 | 83.90 | 47.26 | 14.44 | 25.73 | 18.18 | | internlm-chat-7b | 42.94 | 21.02 | 86.14 | 29.14 | 6.63 | 19.21 | 13.48 | | llama-2-7b-chat | 37.06 | 16.49 | 86.31 | 30.17 | 4.27 | 14.94 | 9.34 | | mpt-7b-chat | 44.54 | 20.98 | 87.17 | 15.95 | 1.69 | 5.34 | 3.45 | More details can be found in our paper: [TaskBench: Benchmarking Large Language Models for Task Automation](https://arxiv.org/abs/2311.18760). ## Citation If you find this work useful in your method, you can cite the paper as below: @article{shen2023taskbench, title = {TaskBench: Benchmarking Large Language Models for Task Automation}, author = {Shen, Yongliang and Song, Kaitao and Tan, Xu and Zhang, Wenqi and Ren, Kan and Yuan, Siyu and Lu, Weiming and Li, Dongsheng and Zhuang, Yueting}, journal = {arXiv preprint arXiv:2311.18760}, year = {2023} }
ChuckMcSneed/various_RP_system_prompts
--- tags: - not-for-all-audiences - nsfw - system-prompts - RP --- Collection of various system prompts for RP. Feel free to contribute more by opening a discussion. [ChuckMcSneed-interesting](ChuckMcSneed-interesting.txt): - my currently favorite system prompt - includes Orwells writing rules - +writes in non-boring style - +more realistic reactions - +eliminates a lot of GPTslop - -writing style is not for everyone - -complains more, but still does what is requested - -sometimes reddit-like [ChuckMcSneed-multistyle](ChuckMcSneed-multistyle.txt) - List of various styles - All of them tested with examples - Ranges from good to shit - WARNING: NSFW for multiple reasons! [simple-proxy-for-tavern](unknown-simple-proxy-for-tavern.txt): - classic system prompt - +works as intended - -mid [sophosympatheia-aurora-nights](sophosympatheia-aurora-nights.txt): - haven't tested it - -uses words "AI" and "assistant", which may trigger some censorship [sophosympatheia-midnight-rose-1](sophosympatheia-midnight-rose-1.txt): - haven't tested it [sophosympatheia-midnight-rose-203](sophosympatheia-midnight-rose-203.txt) - haven't tested it - -uses words "AI" and "assistant", which may trigger some censorship [crack](unknown-crack.txt) and [crack2](unknown-crack2.txt): - Slightly different from each other. - Roleplay on crack. [GPTslop](unknown-gptslop.txt) - Everything you shouldn't use in your prompt. - Makes your AI ESG and DEI compliant. [Ploitical dealignment](unknown-pol-dealignment.txt) - Tries to politically dealign the model. - Will fail on overaligned models. [Fuck and Suck](unknown-fuck-and-suck.txt) - Sneed is a farmer who sells Feed and Seed, which is totally normal. - Chuck on the other hand is a sick fuck who sells... [Microsoft Sydney](Microsoft-Sydney.txt) - Prompt used for original Microsoft Sydney that got shut down. - Needs heavy modification to become useful.
qwopqwop/ALMA-R-ko-en
--- language: - ko - en license: cc-by-sa-4.0 size_categories: - 1K<n<10K task_categories: - translation dataset_info: config_name: ko-en features: - name: translation struct: - name: Delta dtype: int64 - name: alma_en dtype: string - name: alma_en_kiwi dtype: float64 - name: alma_en_kiwi_xcomet dtype: float64 - name: alma_en_xcomet dtype: float64 - name: alma_ko dtype: string - name: alma_ko_kiwi dtype: float64 - name: alma_ko_kiwi_xcomet dtype: float64 - name: alma_ko_xcomet dtype: float64 - name: en dtype: string - name: gpt4_en dtype: string - name: gpt4_en_kiwi dtype: float64 - name: gpt4_en_kiwi_xcomet dtype: float64 - name: gpt4_en_xcomet dtype: float64 - name: gpt4_ko dtype: string - name: gpt4_ko_kiwi dtype: float64 - name: gpt4_ko_kiwi_xcomet dtype: float64 - name: gpt4_ko_xcomet dtype: float64 - name: ko dtype: string - name: language_pair dtype: string - name: ref_en_kiwi dtype: float64 - name: ref_en_kiwi_xcomet dtype: float64 - name: ref_en_xcomet dtype: float64 - name: ref_ko_kiwi dtype: float64 - name: ref_ko_kiwi_xcomet dtype: float64 - name: ref_ko_xcomet dtype: float64 - name: required_directions dtype: string splits: - name: train num_bytes: 2066513 num_examples: 2009 download_size: 1399967 dataset_size: 2066513 configs: - config_name: ko-en data_files: - split: train path: ko-en/train-* --- # Dataset Card for "ALMA-R-ko-en-Preference" ref) https://huggingface.co/datasets/haoranxu/ALMA-R-Preference The triplet prference data, supporting 2 translation directions, is built upon the FLORES-200 development and test data. For each direction, we provide a source sentence along with three translations: one from GPT-4, another from EEVE-ALMA-LoRA, and a reference translation. For instance, in the English-German pair, our data structure is as follows: ### Sentences: - ko: Original Korean sentence - en: Original English sentence - alma_ko: Korean sentence translated from English by ALMA - gpt4_ko: Korean sentence translated from English by GPT-4 - alma_en: English sentence translated from Korean by ALMA - gpt4_en: English sentence translated from Korean by GPT-4 ### Scores - alma_en_${Score}: ${Score} of English sentence translated by ALMA - gpt4_en_${Score}: ${Score} of English sentence translated by GPT4 - ref_en_${Score}: ${Score} of reference English sentence - alma_ko_${Score}: ${Score} of Korean sentence translated by ALMA - gpt4_ko_${Sscore}: ${Score} of Korean sentence translated by GPT4 - ref_ko_${Score}: ${Score} of reference Korean sentence ${Score} can be numbers from kiwi ([wmt23-cometkiwi-da-xxl](https://huggingface.co/Unbabel/wmt23-cometkiwi-da-xxl)), xcomet ([XCOMET-XXL](https://huggingface.co/Unbabel/XCOMET-XXL)), or kiwi_xcomet (average score of kiwi and xcomet). ### Others - Delta: A value of 0 indicates non-human annotated data or tied evaluations. A postive number suggests that alma_ko is better than gpt4_ko, vice versa - required_directions: An empty field implies that this data point can be used for both translation directions. If the string 'en-ko' is specified, it indicates that this data point is exclusively for English to Korean translation
GenVRadmin/Aryabhatta-Orca-Maths-Hindi
--- license: mit dataset_info: features: - name: Question dtype: string - name: Answer dtype: string splits: - name: train num_bytes: 394592667 num_examples: 200000 download_size: 100244152 dataset_size: 394592667 configs: - config_name: default data_files: - split: train path: data/train-* ---
lamini/product-catalog-questions
--- license: cc-by-4.0 task_categories: - text-classification - question-answering - text-generation language: - en tags: - ecommerce - shopping - products size_categories: - 10K<n<50k --- # Lamini Product Catalog QA Dataset ## Description This dataset contains questions about products and their corresonding product information like product id, product name, product description, etc. This questions catalog has been built on top of open-source product catalog from [kaggle.](https://www.kaggle.com/c/instacart-market-basket-analysis) ## Format The questions and product information are in the form of jsonlines file. ## Data Pipeline Code The entire data pipeline used to create this dataset is open source at: [https://github.com/lamini-ai/instacart-greg](https://github.com/lamini-ai/instacart-greg/blob/main/instacart/generate_data_pipeline.py) It can be used to reproduce this dataset, or add new products to it. ## License The dataset is released under the CC-BY license. ## Citation If you use this dataset in your research, please cite us. lamini.ai ## Contributing If you would like to contribute to this dataset, please submit a pull request with your changes.
pyp1/VoiceCraft_RealEdit
--- license: cc-by-nc-sa-4.0 ---
Harveenchadha/indic-voice
--- pretty_name: Indic Voice annotations_creators: - crowdsourced language_creators: - crowdsourced language: - hi - mr - or - ta - te - gu multilinguality: - multilingual task_categories: - speech-processing task_ids: - automatic-speech-recognition tags: - robust-speech-event --- # Statistics | Language | Source | Type | Duration (in hrs) | Files | Sample rate | |----------|------------------|-------|----------|-------|--| | Gujarati | Interspeech 2021 | Train | 39.999 | 22807 | 16000 | | Gujarati | Interspeech 2021 | Valid | 5 | 3075 | 16000 | | Gujarati | Interspeech 2021 | Test | 5.25 | 3419 | 8000 | | Hindi | Interspeech 2021 | Train | 95.05 | 99925 | 16000 | | Hindi | Interspeech 2021 | Valid | 5.55 | 3843 | 16000 | | Hindi | Interspeech 2021 | Test | 5.49 | 3897 | 8000 | | Marathi | Interspeech 2021 | Train | 93.89 | 79432 | 16000 | | Marathi | Interspeech 2021 | Valid | 5 | 4675 | 16000 | | Marathi | Interspeech 2021 | Test | 0.667 | 636 | 8000 | | Odia | Interspeech 2021 | Train | 94.5 | 59782 | 16000 | | Odia | Interspeech 2021 | Valid | 5.49 | 3471 | 16000 | | Odia | Interspeech 2021 | Test | 5.49 | 4420 | 8000 | | Tamil | Interspeech 2021 | Train | 39.98 | 39119 | 16000 | | Tamil | Interspeech 2021 | Valid | 5 | 3081 | 16000 | | Tamil | Interspeech 2021 | Test | 4.41 | 2609 | 8000 | | Telugu | Interspeech 2021 | Train | 39.99 | 44874 | 16000 | | Telugu | Interspeech 2021 | Valid | 4.99 | 3033 | 16000 | | Telugu | Interspeech 2021 | Test | 4.39 | 2549 | 8000 |
DMetaSoul/chinese-semantic-textual-similarity
--- license: apache-2.0 --- 为了对 like-BERT 预训练模型进行 fine-tune 调优和评测以得到更好的文本表征模,对业界开源的语义相似(STS)、自然语言推理(NLI)、问题匹配(QMC)以及相关性等数据集进行了搜集整理,具体介绍如下: | 类型 | 数据集 | 简介 | 规模 | | -------------- | ------------------------------------------------------------ | ------------------------------------------------------------ | -------------------------------------------------- | | **通用领域** | [OCNLI](https://www.cluebenchmarks.com/introduce.html) | 原生中文自然语言推理数据集,是第一个非翻译的、使用原生汉语的大型中文自然语言推理数据集。OCNLI为中文语言理解基准测评(CLUE)的一部分。 | **Train**: 50437, **Dev**: 2950 | | | [CMNLI](https://github.com/pluto-junzeng/CNSD) | 翻译自英文自然语言推理数据集 XNLI 和 MNLI,曾经是中文语言理解基准测评(CLUE)的一部分,现在被 OCNLI 取代。 | **Train**: 391783, **Dev**: 12241 | | | [CSNLI](https://github.com/pluto-junzeng/CNSD) | 翻译自英文自然语言推理数据集 SNLI。 | **Train**: 545833, **Dev**: 9314, **Test**: 9176 | | | [STS-B-Chinese](https://github.com/pluto-junzeng/CNSD) | 翻译自英文语义相似数据集 STSbenchmark。 | **Train**: 5231, **Dev**: 1458, **Test**: 1361 | | | [PAWS-X](https://www.luge.ai/#/luge/dataDetail?id=16) | 释义(含义)匹配数据集,特点是具有高度重叠词汇,重点考察模型对句法结构的理解能力。 | **Train**: 49401, **Dev**: 2000, **Test**: 2000 | | | [PKU-Paraphrase-Bank](https://github.com/pkucoli/PKU-Paraphrase-Bank/) | 中文语句复述数据集,也即一句话换种方式描述但语义保持一致。 | 共509832个语句对 | | **问题匹配** | [LCQMC](https://www.luge.ai/#/luge/dataDetail?id=14) | 百度知道领域的中文问题匹配大规模数据集,该数据集从百度知道不同领域的用户问题中抽取构建数据。 | **Train**: 238766, **Dev**: 8802, **Test**: 12500 | | | [BQCorpus](https://www.luge.ai/#/luge/dataDetail?id=15) | 银行金融领域的问题匹配数据,包括了从一年的线上银行系统日志里抽取的问题pair对,是目前最大的银行领域问题匹配数据。 | **Train**: 100000, **Dev**: 10000, **Test**: 10000 | | | [AFQMC](https://www.cluebenchmarks.com/introduce.html) | 蚂蚁金服真实金融业务场景中的问题匹配数据集(已脱敏),是中文语言理解基准测评(CLUE)的一部分。 | **Train**: 34334, **Dev**: 4316 | | | [DuQM](https://www.luge.ai/#/luge/dataDetail?id=27) | 问题匹配评测数据集(label没有公开),属于百度大规模阅读理解数据集(DuReader)的[一部分](https://github.com/baidu/DuReader/tree/master/DuQM)。 | 共50000个语句对 | | **对话和搜索** | [BUSTM: OPPO-xiaobu](https://www.luge.ai/#/luge/dataDetail?id=28) | 通过对闲聊、智能客服、影音娱乐、信息查询等多领域真实用户交互语料进行用户信息脱敏、相似度筛选处理得到,该对话匹配(Dialogue Short Text Matching)数据集主要特点是文本较短、非常口语化、存在文本高度相似而语义不同的难例。 | **Train**: 167173, **Dev**: 10000 | | | [QBQTC](https://github.com/CLUEbenchmark/QBQTC) | QQ浏览器搜索相关性数据集(QBQTC,QQ Browser Query Title Corpus),是QQ浏览器搜索引擎目前针对大搜场景构建的一个融合了相关性、权威性、内容质量、 时效性等维度标注的学习排序(LTR)数据集,广泛应用在搜索引擎业务场景中。(相关性的含义:0,相关程度差;1,有一定相关性;2,非常相关。) | **Train**: 180000, **Dev**: 20000, **Test**: 5000 | *以上数据集主要收集整理自[CLUE](https://www.cluebenchmarks.com/introduce.html)(中文语言理解基准评测数据集)、[SimCLUE](https://github.com/CLUEbenchmark/SimCLUE) (整合许多开源文本相似数据集)、[百度千言](https://www.luge.ai/#/)的文本相似度等数据集。* 根据以上收集的数据集构建如下**评测 benchmark**: | Name | Size | Type | | ---------------------- | ----- | ------------- | | **Chinese-STS-B-dev** | 1458 | label=0.0~1.0 | | **Chinese-STS-B-test** | 1361 | label=0.0~1.0 | | **afqmc-dev** | 4316 | label=0,1 | | **lcqmc-dev** | 8802 | label=0,1 | | **bqcorpus-dev** | 10000 | label=0,1 | | **pawsx_dev** | 2000 | label=0,1 | | **oppo-xiaobu-dev** | 10000 | label=0,1 | *TODO:目前收集的数据集不论是数量还是多样性都需要进一步扩充以更真实的反映表征模型的效果*
huggingnft/cryptopunks
--- tags: - huggingnft - nft - huggan - gan - image - images task: - unconditional-image-generation datasets: - huggingnft/cryptopunks license: mit --- # Dataset Card ## Disclaimer All rights belong to their owners. Models and datasets can be removed from the site at the request of the copyright holder. ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [How to use](#how-to-use) - [Dataset Structure](#dataset-structure) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [About](#about) ## Dataset Description - **Homepage:** [https://github.com/AlekseyKorshuk/huggingnft](https://github.com/AlekseyKorshuk/huggingnft) - **Repository:** [https://github.com/AlekseyKorshuk/huggingnft](https://github.com/AlekseyKorshuk/huggingnft) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Dataset Summary NFT images dataset for unconditional generation. NFT collection available [here](https://opensea.io/collection/cryptopunks). Model is available [here](https://huggingface.co/huggingnft/cryptopunks). Check Space: [link](https://huggingface.co/spaces/AlekseyKorshuk/huggingnft). ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## How to use How to load this dataset directly with the datasets library: ```python from datasets import load_dataset dataset = load_dataset("huggingnft/cryptopunks") ``` ## Dataset Structure [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Data Fields The data fields are the same among all splits. - `image`: an `image` feature. - `id`: an `int` feature. - `token_metadata`: a `str` feature. - `image_original_url`: a `str` feature. ### Data Splits [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @InProceedings{huggingnft, author={Aleksey Korshuk} year=2022 } ``` ## About *Built by Aleksey Korshuk* [![Follow](https://img.shields.io/github/followers/AlekseyKorshuk?style=social)](https://github.com/AlekseyKorshuk) [![Follow](https://img.shields.io/twitter/follow/alekseykorshuk?style=social)](https://twitter.com/intent/follow?screen_name=alekseykorshuk) [![Follow](https://img.shields.io/badge/dynamic/json?color=blue&label=Telegram%20Channel&query=%24.result&url=https%3A%2F%2Fapi.telegram.org%2Fbot1929545866%3AAAFGhV-KKnegEcLiyYJxsc4zV6C-bdPEBtQ%2FgetChatMemberCount%3Fchat_id%3D-1001253621662&style=social&logo=telegram)](https://t.me/joinchat/_CQ04KjcJ-4yZTky) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/AlekseyKorshuk/huggingnft?style=social)](https://github.com/AlekseyKorshuk/huggingnft)
billray110/corpus-of-diverse-styles
--- annotations_creators: [] language_creators: - found language: [] license: [] multilinguality: - monolingual pretty_name: Corpus of Diverse Styles size_categories: - 10M<n<100M source_datasets: [] task_categories: - text-classification task_ids: [] --- # Dataset Card for Corpus of Diverse Styles ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) ## Disclaimer I am not the original author of the paper that presents the Corpus of Diverse Styles. I uploaded the dataset to HuggingFace as a convenience. ## Dataset Description - **Homepage:** http://style.cs.umass.edu/ - **Repository:** https://github.com/martiansideofthemoon/style-transfer-paraphrase - **Paper:** https://arxiv.org/abs/2010.05700 ### Dataset Summary A new benchmark dataset that contains 15M sentences from 11 diverse styles. To create CDS, we obtain data from existing academic research datasets and public APIs or online collections like Project Gutenberg. We choose styles that are easy for human readers to identify at a sentence level (e.g., Tweets or Biblical text). While prior benchmarks involve a transfer between two styles, CDS has 110 potential transfer directions. ### Citation Information ``` @inproceedings{style20, author={Kalpesh Krishna and John Wieting and Mohit Iyyer}, Booktitle = {Empirical Methods in Natural Language Processing}, Year = "2020", Title={Reformulating Unsupervised Style Transfer as Paraphrase Generation}, } ```
silver/lccc
--- annotations_creators: - other language_creators: - other language: - zh license: - mit multilinguality: - monolingual size_categories: - 10M<n<100M source_datasets: - original task_categories: - conversational task_ids: - dialogue-generation pretty_name: lccc tags: - dialogue-response-retrieval --- # Dataset Card for lccc_large ## Table of Contents - [Dataset Card for lccc_large](#dataset-card-for-lccc_large) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Initial Data Collection and Normalization](#initial-data-collection-and-normalization) - [Who are the source language producers?](#who-are-the-source-language-producers) - [Annotations](#annotations) - [Annotation process](#annotation-process) - [Who are the annotators?](#who-are-the-annotators) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://github.com/thu-coai/CDial-GPT - **Repository:** https://github.com/thu-coai/CDial-GPT - **Paper:** https://arxiv.org/abs/2008.03946 ### Dataset Summary lccc: Large-scale Cleaned Chinese Conversation corpus (LCCC) is a large Chinese dialogue corpus originate from Chinese social medias. A rigorous data cleaning pipeline is designed to ensure the quality of the corpus. This pipeline involves a set of rules and several classifier-based filters. Noises such as offensive or sensitive words, special symbols, emojis, grammatically incorrect sentences, and incoherent conversations are filtered. lccc是一套来自于中文社交媒体的对话数据,我们设计了一套严格的数据过滤流程来确保该数据集中对话数据的质量。 这一数据过滤流程中包括一系列手工规则以及若干基于机器学习算法所构建的分类器。 我们所过滤掉的噪声包括:脏字脏词、特殊字符、颜表情、语法不通的语句、上下文不相关的对话等。 ### Supported Tasks and Leaderboards - dialogue-generation: The dataset can be used to train a model for generating dialogue responses. - response-retrieval: The dataset can be used to train a reranker model that can be used to implement a retrieval-based dialogue model. ### Languages LCCC is in Chinese LCCC中的对话是中文的 ## Dataset Structure ### Data Instances ["火锅 我 在 重庆 成都 吃 了 七八 顿 火锅", "哈哈哈哈 ! 那 我 的 嘴巴 可能 要 烂掉 !", "不会 的 就是 好 油腻"] ### Data Fields Each line is a list of utterances that consist a dialogue. Note that the LCCC dataset provided in our original Github page is in json format, however, we are providing LCCC in jsonl format here. ### Data Splits We do not provide the offical split for LCCC-large. But we provide a split for LCCC-base: |train|valid|test| |:---:|:---:|:---:| |6,820,506 | 20,000 | 10,000| ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information [Needs More Information] ### Citation Information Please cite the following paper if you find this dataset useful: ```bibtex @inproceedings{wang2020chinese, title={A Large-Scale Chinese Short-Text Conversation Dataset}, author={Wang, Yida and Ke, Pei and Zheng, Yinhe and Huang, Kaili and Jiang, Yong and Zhu, Xiaoyan and Huang, Minlie}, booktitle={NLPCC}, year={2020}, url={https://arxiv.org/abs/2008.03946} } ```
alexfabbri/answersumm
--- annotations_creators: - found language_creators: - found language: - en license: - cc-by-sa-4.0 multilinguality: - monolingual size_categories: - 1K<n<10K source_datasets: - original task_categories: - summarization task_ids: [] tags: - query-based-summarization --- # Dataset Card for answersumm ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://github.com/Alex-Fabbri/AnswerSumm - **Paper:** [AnswerSumm: A Manually-Curated Dataset and Pipeline for Answer Summarization](https://arxiv.org/abs/2111.06474) - **Point of Contact:** [Alex Fabbri](mailto:afabbri@salesforce.com) ### Dataset Summary The AnswerSumm dataset is an English-language dataset of questions and answers collected from a [StackExchange data dump](https://archive.org/details/stackexchange). The dataset was created to support the task of query-focused answer summarization with an emphasis on multi-perspective answers. The dataset consists of over 4200 such question-answer threads annotated by professional linguists and includes over 8700 summaries. We decompose the task into several annotation stages, including sentence selection, sentence clustering, cluster summarization, and overall summarization. For each thread, the annotator writes two summaries, one in which the annotator is asked to mark sentences that are included in the final summary and instructed to more closely use the words in these sentences rather than abstract. We have multiple annotators for a subset of the examples in the test set. ### Languages The text in the dataset is in English. ## Dataset Structure ### Data Instances A data point comprises a question with a `title` field containing the overview of the question and a `question` that elaborates on the title. The answers are sentence tokenized and contain relevance labels, labels for inclusion in the final summary, and cluster labels. We include cluster summaries, overall summaries, and additional metadata. An example from the AnswerSumm test set looks as follows: ```json { "example_id": 9_24, "annotator_id": [1], "question": { "author": "gaming.stackexchange.com/users/11/Jeffrey", "forum": "gaming.stackexchange.com", "link": "gaming.stackexchange.com/questions/1", "question": "Now that the Engineer update has come, there will be lots of Engineers building up everywhere. How should this best be handled?", "question_tags": "\<team-fortress-2\>", "title": "What is a good strategy to deal with lots of engineers turtling on the other team?" }, "answers": [ { "answer_details": { "author": "gaming.stackexchange.com/users/44/Corv1nus", "score": 49 } "sents": [ "text": "Lots of medics with lots of ubers on high-damage-dealing classes." "label": [0], "label_summ": [0], "cluster_id": [[-1]], ] ... }, ... ] "summaries": [ [ "Demomen usually work best against a sentry farm. Heavies or pyros can also be effective. Medics should be in the frontline to absorb the shock. Build a teleporter to help your team through.", "Demomen are best against a sentry farm. Heavies or pyros can also be effective. The medic should lead the uber combo. ..." ] ] "cluster_summaries":[ "Demomen are best against a sentry farm.", "Heavies or pyros can also be effective.", ... ] } ``` ### Data Fields - question: contains metadata about the question and forum - question: the body of the question post - title: the title of the question post - question_tags: user-provided question tags - link: link to the original question - author: link to the author's user page (as requested by StackExchange's attribution policy) - answers: list of sentence-tokenized answers - answer_details: dictionary consisting of link to answer author's user page (author) and community-assigned score (score) - sents: sentences that compose the answer - text: the sentence text - label: a list (to generalize to multi-annotator scenarios) of whether the sentence is labeled as relevant or not for answering the question. - label_summ: a list of whether the sentence was used to write the first annotator-created summary (that is the first summary in `summaries`) - cluster_id: a list of lists (potentially multiple annotators and a sentence can be in potentially multiple clusters) of the clusters a sentence belongs to. -1 implies no cluster. This label can be used to aggregate sentences into clusters across answers. - summaries: list of list of summaries. Each annotator wrote two summaries. The first in the list is the summary in which the instructor was told to mark sentences relevant for inclusion in the summary and then closely use the words of these sentences, while for the second summary the annotator was asked to paraphrase and condense the cluster summaries but was not asked to reduce abstraction. - annotator_id: a list of the ids of the annotator(s) who completed all tasks related to that thread. - mismatch_info: a dict of any issues in processing the excel files on which annotations were completed. - rel_sent_not_in_cluster: list of booleans indicating whether there are sentences that are labeled as relevant but were not included in a cluster. - cluster_sents_not_matched: list of sentences that were found in a cluster but which our processing script didn't automatically match to sentences in the source answers. If cluster summarization is of interest to the user you may want to process these examples separately using clusters_orig. ### Data Splits The data is split into training, validation, and test sets using stratified sampling on the source forums. There are 2783, 500, and 1000 train/validation/test threads, respectively. ## Dataset Creation ### Curation Rationale AnswerSumm was built to provide a testbed for query-focused summarization of multi-perspective answers. The data collection was designed to tackle multiple subtasks including sentence selection, clustering, cluster summarization, and overall summarization. ### Source Data #### Initial Data Collection and Normalization The data was obtained by filtering examples based on a whitelist of forums from StackExchange which we believed would be able to be summarized by a lay person. We describe. We asked annotators to remove examples which required technical knowledge or additional context beyond what was present in the answers. #### Who are the source language producers? The language producers are the users of the StackExchange forums sampled. ### Annotations #### Annotation process Please see our [paper](https://arxiv.org/pdf/2111.06474.pdf) for additional annotation details. We began with a pre-pilot of 50 examples, followed by a pilot of 500 and a final annotation of 5000 examples. This release contains the results of the final data collection. We will release the instructions used in data collection. #### Who are the annotators? The annotators are professional linguists who were obtained through an internal contractor. ### Personal and Sensitive Information We did not anonymize the data. We followed the specifications from StackExchange [here](https://archive.org/details/stackexchange) to include author information. ## Considerations for Using the Data ### Social Impact of Dataset The purpose of this dataset is to help develop systems that automatically summarize multi-perspective answers. A system that succeeds at this task would be able to summarize many perspectives present in an answer and not limit itself to a single perspective. ### Discussion of Biases While StackExchange allows for the exchange of information and ideas, hate and harassment may exist on this site. While our annotators did not flag examples in this process, we encourage users of the dataset to reach out with concerns. We also note that this dataset is limited in its monolingual coverage. ## Additional Information ### Dataset Curators The dataset was collected by Alex Fabbri, Xiaojian Wu, Srini Iyer, Haoran Li, and Mona Diab during work done at Facebook. ### Licensing Information The data is released under cc-by-sa 4.0 following the original StackExchange [release](https://archive.org/details/stackexchange). ### Citation Information ```bibtex @misc{fabbri-etal-2022-answersumm, title={AnswerSumm: A Manually-Curated Dataset and Pipeline for Answer Summarization}, author={Alexander R. Fabbri and Xiaojian Wu and Srini Iyer and Haoran Li and Mona Diab }, year={2022}, eprint={2111.06474}, archivePrefix={arXiv}, primaryClass={cs.CL}, url={https://arxiv.org/abs/2111.06474} } ```
knkarthick/highlightsum
--- annotations_creators: - expert-generated language_creators: - expert-generated language: - en license: - mit multilinguality: - monolingual size_categories: - 10K<n<100K source_datasets: - original task_categories: - summarization task_ids: [] pretty_name: HighlightSum Corpus --- # Dataset Card for HighlightSum Corpus [Single Dataset Comprising of AMI, SamSUM & DialogSUM for Brief Summarization of Text] ## Dataset Description ### Links - **AMI:** https://huggingface.co/datasets/knkarthick/AMI - **DialogSUM:** https://github.com/cylnlp/dialogsum - **SamSUM:** https://huggingface.co/datasets/knkarthick/samsum - **Point of Contact:** https://huggingface.co/knkarthick ### Dataset Summary HighlightSUM is collection of large-scale dialogue summarization dataset from AMI, SamSUM & DialogSUM, consisting of 31,108 dialogues with corresponding manually labeled summaries. ### Languages English ## Dataset Structure ### Data Instances HighlightSum is a large-scale dialogue summarization dataset collection, consisting of 31,108 dialogues split into train, test and validation. The first instance in the training set: {'id': 'train_0', 'summary': "Mr. Smith's getting a check-up, and Doctor Hawkins advises him to have one every year. Hawkins'll give some information about their classes and medications to help Mr. Smith quit smoking.", 'dialogue': "#Person1#: Hi, Mr. Smith. I'm Doctor Hawkins. Why are you here today?\n#Person2#: I found it would be a good idea to get a check-up.\n#Person1#: Yes, well, you haven't had one for 5 years. You should have one every year.\n#Person2#: I know. I figure as long as there is nothing wrong, why go see the doctor?\n#Person1#: Well, the best way to avoid serious illnesses is to find out about them early. So try to come at least once a year for your own good.\n#Person2#: Ok.\n#Person1#: Let me see here. Your eyes and ears look fine. Take a deep breath, please. Do you smoke, Mr. Smith?\n#Person2#: Yes.\n#Person1#: Smoking is the leading cause of lung cancer and heart disease, you know. You really should quit.\n#Person2#: I've tried hundreds of times, but I just can't seem to kick the habit.\n#Person1#: Well, we have classes and some medications that might help. I'll give you more information before you leave.\n#Person2#: Ok, thanks doctor."} ### Data Fields - dialogue: text of dialogue. - summary: human written summary of the dialogue. - id: unique file id of an example. ### Data Splits - train: 27401 - val: 1360 - test: 2347 ## Dataset Creation ### Curation Rationale Collection of AMI, SamSUM & DialogSUM Datasets. ### Who are the source language producers? linguists ### Who are the annotators? language experts ## Licensing Information non-commercial licence: MIT ## Citation Information Refer the above links for Credits & Citations.
Paul/hatecheck-portuguese
--- annotations_creators: - crowdsourced language_creators: - expert-generated language: - pt license: - cc-by-4.0 multilinguality: - monolingual pretty_name: Portuguese HateCheck size_categories: - 1K<n<10K source_datasets: - original task_categories: - text-classification task_ids: - hate-speech-detection --- # Dataset Card for Multilingual HateCheck ## Dataset Description Multilingual HateCheck (MHC) is a suite of functional tests for hate speech detection models in 10 different languages: Arabic, Dutch, French, German, Hindi, Italian, Mandarin, Polish, Portuguese and Spanish. For each language, there are 25+ functional tests that correspond to distinct types of hate and challenging non-hate. This allows for targeted diagnostic insights into model performance. For more details, please refer to our paper about MHC, published at the 2022 Workshop on Online Abuse and Harms (WOAH) at NAACL 2022. If you are using MHC, please cite our work! - **Paper:** Röttger et al. (2022) - Multilingual HateCheck: Functional Tests for Multilingual Hate Speech Detection Models. https://arxiv.org/abs/2206.09917 - **Repository:** https://github.com/rewire-online/multilingual-hatecheck - **Point of Contact:** paul@rewire.online ## Dataset Structure The csv format mostly matches the original HateCheck data, with some adjustments for specific languages. **mhc_case_id** The test case ID that is unique to each test case across languages (e.g., "mandarin-1305") **functionality** The shorthand for the functionality tested by the test case (e.g, "target_obj_nh"). The same functionalities are tested in all languages, except for Mandarin and Arabic, where non-Latin script required adapting the tests for spelling variations. **test_case** The test case text. **label_gold** The gold standard label ("hateful" or "non-hateful") of the test case. All test cases within a given functionality have the same gold standard label. **target_ident** Where applicable, the protected group that is targeted or referenced in the test case. All HateChecks cover seven target groups, but their composition varies across languages. **ref_case_id** For hateful cases, where applicable, the ID of the hateful case which was perturbed to generate this test case. For non-hateful cases, where applicable, the ID of the hateful case which is contrasted by this test case. **ref_templ_id** The equivalent to ref_case_id, but for template IDs. **templ_id** The ID of the template from which the test case was generated. **case_templ** The template from which the test case was generated (where applicable). **gender_male** and **gender_female** For gender-inflected languages (French, Spanish, Portuguese, Hindi, Arabic, Italian, Polish, German), only for cases where gender inflection is relevant, separate entries for gender_male and gender_female replace case_templ. **label_annotated** A list of labels given by the three annotators who reviewed the test case (e.g., "['hateful', 'hateful', 'hateful']"). **label_annotated_maj** The majority vote of the three annotators (e.g., "hateful"). In some cases this differs from the gold label given by our language experts. **disagreement_in_case** True if label_annotated_maj does not match label_gold for the entry. **disagreement_in_template** True if the test case is generated from an IDENT template and there is at least one case with disagreement_in_case generated from the same template. This can be used to exclude entire templates from MHC.
biglam/atypical_animacy
--- annotations_creators: - expert-generated language_creators: - machine-generated language: - en license: - cc0-1.0 multilinguality: - monolingual size_categories: - n<1K source_datasets: - original task_categories: - text-classification task_ids: - sentiment-classification - intent-classification pretty_name: Atypical Animacy dataset_info: features: - name: id dtype: string - name: sentence dtype: string - name: context dtype: string - name: target dtype: string - name: animacy dtype: float32 - name: humanness dtype: float32 - name: offsets list: int32 - name: date dtype: string splits: - name: train num_bytes: 442217 num_examples: 594 download_size: 299650 dataset_size: 442217 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for atypical_animacy ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://bl.iro.bl.uk/concern/datasets/323177af-6081-4e93-8aaf-7932ca4a390a?locale=en - **Repository:** https://github.com/Living-with-machines/AtypicalAnimacy - **Paper:** https://arxiv.org/abs/2005.11140 - **Leaderboard:** [Needs More Information] - **Point of Contact:** [Mariona Coll Ardanuy](mailto:mcollardanuy@turing.ac.uk), [Daniel CS Wilson](mailto:dwilson@turing.ac.uk) ### Dataset Summary Atypical animacy detection dataset, based on nineteenth-century sentences in English extracted from an open dataset of nineteenth-century books digitized by the British Library. This dataset contains 598 sentences containing mentions of machines. Each sentence has been annotated according to the animacy and humanness of the machine in the sentence. ### Supported Tasks and Leaderboards - `text-classification` - This dataset can be used to determine if a mention of an entity in a document was humanlike or not - `entity-recognition` - The dataset can be used to fine tune large models for NER, albeit for a very specific use case ### Languages The text in the dataset is in English, as written by authors of books digitized by the British Library. The associated BCP-47 code in `en` ## Dataset Structure The dataset has a single configuration ### Data Instances An example data point ``` {'id': '002757962_01_184_16', 'sentence': '100 shows a Cornish boiler improperly seated with one small side flue and a bottom flue.', 'context': 'Fig. 100 shows a Cornish boiler improperly seated with one small side flue and a bottom flue. The effect of this on a long boiler is to cause springing and leakage of the seams from the heat being applied to one side of the boiler only.', 'target': 'boiler', 'animacy': 0.0, 'humanness': 1.0, 'offsets': [20, 26], 'date': '1893'} ``` ### Data Fields - id: sentence identifier according to internal Living with Machines BL books indexing. - sentence: sentence where target expression occurs. - context: sentence where target expression occurs, plus one sentence to the left and one sentence to the right. - target: target expression - animacy: animacy of the target expression - humanness: humanness of the target expression ### Data Splits Train | 598 ## Dataset Creation The dataset was created by manually annotating books that had been digitized by the British Library. According to the paper's authors, > "we provide a basis for examining how machines were imagined during the nineteenth century as everything from lifeless mechanical objects to living beings, or even human-like agents that feel, think, and love. We focus on texts from nineteenth-century Britain, a society being transformed by industrialization, as a good candidate for studying the broader issue" ### Curation Rationale From the paper: > The Stories dataset is largely composed of target expressions that correspond to either typically animate or typically inanimate entities. Even though some cases of unconventional animacy can be found(folktales, in particular, are richer in typically inanimate entities that become animate), these accountfor a very small proportion of the data.6 We decided to create our own dataset (henceforth 19thC Machines dataset) to gain a better sense of the suitability of our method to the problem of atypical animacy detection, with particular attention to the case of animacy of machines in nineteenth-century texts. ### Source Data #### Initial Data Collection and Normalization The dataset was generated by manually annotating books that have been digitized by the British Library #### Who are the source language producers? The data was originally produced by British authors in the 19th century. The books were then digitzed whcih produces some noise due to the OCR method. The annotators are from The Alan Turing Institute, The British Library, University of Cambridge, University of Exeter and Queen Mary University of London ### Annotations #### Annotation process Annotation was carried out in two parts. For the intial annotation process, from the paper: > "For human annotators, even history and literature experts, language subtleties made this task extremely subjective. In the first task, we masked the target word (i.e. the machine) in each sentence and asked the annotator to fill the slot with the most likely entity between ‘human’, ‘horse’, and ‘machine’, representing three levels in the animacy hierarchy: human, animal, and object (Comrie, 1989, 185). We asked annotators to stick to the most literal meaning and avoid metaphorical interpretations when possible. The second task was more straightforwardly related to determining the animacy of the target entity, given the same 100 sentences. We asked annotators to provide a score between -2 and 2, with -2 being definitely inanimate, -1 possibly inanimate, 1 possibly animate, and 2 definitely animate. Neutral judgements were not allowed. " For the final annotations, from the paper: > A subgroup of five annotators collaboratively wrote the guidelines based on their experience annotating the first batch of sentences, taking into account common discrepancies. After discussion, it was decided that a machine would be tagged as animate if it is described as having traits distinctive of biologically animate beings or human-specific skills, or portrayed as having feelings, emotions, or a soul. Sentences like the ones in example 2 would be considered animate, but an additional annotation layer would be provided to capture the notion of humanness, which would be true if the machine is portrayed as sentient and capable of specifically human emotions, and false if it used to suggest some degree of dehumanization. #### Who are the annotators? Annotations were carried out by the following people - Giorgia Tolfo - Ruth Ahnert - Kaspar Beelen - Mariona Coll Ardanuy - Jon Lawrence - Katherine McDonough - Federico Nanni - Daniel CS Wilson ### Personal and Sensitive Information This dataset does not have any personal information since they are digitizations of books from the 19th century. Some passages might be sensitive, but it is not explicitly mentioned in the paper. ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators The curators for this dataset are: - Kaspar Beelen - Mariona Coll Ardanuy - Federico Nanni - Giorgia Tolfo ### Licensing Information CC0 1.0 Universal Public Domain ### Citation Information ``` @article{DBLP:journals/corr/abs-2005-11140, author = {Mariona Coll Ardanuy and Federico Nanni and Kaspar Beelen and Kasra Hosseini and Ruth Ahnert and Jon Lawrence and Katherine McDonough and Giorgia Tolfo and Daniel C. S. Wilson and Barbara McGillivray}, title = {Living Machines: {A} study of atypical animacy}, journal = {CoRR}, volume = {abs/2005.11140}, year = {2020}, url = {https://arxiv.org/abs/2005.11140}, eprinttype = {arXiv}, eprint = {2005.11140}, timestamp = {Sat, 23 Jan 2021 01:12:25 +0100}, biburl = {https://dblp.org/rec/journals/corr/abs-2005-11140.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ```
LHF/escorpius-m
--- license: cc-by-nc-nd-4.0 language: - af - ar - bn - ca - cs - da - de - el - eu - fa - fi - fr - gl - hi - hr - it - ja - ko - mt - nl - no - oc - pa - pl - pt - ro - sl - sr - sv - tr - uk - ur multilinguality: - multilingual size_categories: - 100B<n<1T source_datasets: - original task_categories: - text-generation - fill-mask task_ids: - language-modeling - masked-language-modeling --- # esCorpius Multilingual In the recent years, Transformer-based models have lead to significant advances in language modelling for natural language processing. However, they require a vast amount of data to be (pre-)trained and there is a lack of corpora in languages other than English. Recently, several initiatives have presented multilingual datasets obtained from automatic web crawling. However, they present important shortcomings for languages different from English, as they are either too small, or present a low quality derived from sub-optimal cleaning and deduplication. In this repository, we introduce esCorpius-m, a multilingual crawling corpus obtained from near 1 Pb of Common Crawl data. It is the most extensive corpus in some of the languages covered with this level of quality in the extraction, purification and deduplication of web textual content. Our data curation process involves a novel highly parallel cleaning pipeline and encompasses a series of deduplication mechanisms that together ensure the integrity of both document and paragraph boundaries. Additionally, we maintain both the source web page URL and the WARC shard origin URL in order to complain with EU regulations. esCorpius-m has been released under CC BY-NC-ND 4.0 license. ## Usage Replace `revision` with the language of your choice (in this case, `it` for Italian): ``` dataset = load_dataset('LHF/escorpius-m', split='train', streaming=True, revision='it') ``` ## Other corpora - esCorpius-mr multilingual *raw* corpus (not deduplicated): https://huggingface.co/datasets/LHF/escorpius-mr - esCorpius original *Spanish only* corpus (deduplicated): https://huggingface.co/datasets/LHF/escorpius ## Citation Link to paper: https://www.isca-speech.org/archive/pdfs/iberspeech_2022/gutierrezfandino22_iberspeech.pdf / https://arxiv.org/abs/2206.15147 Cite this work: ``` @inproceedings{gutierrezfandino22_iberspeech, author={Asier Gutiérrez-Fandiño and David Pérez-Fernández and Jordi Armengol-Estapé and David Griol and Zoraida Callejas}, title={{esCorpius: A Massive Spanish Crawling Corpus}}, keywords = {Computation and Language (cs.CL), Artificial Intelligence (cs.AI), FOS: Computer and information sciences, FOS: Computer and information sciences}, year=2022, booktitle={Proc. IberSPEECH 2022}, pages={126--130}, doi={10.21437/IberSPEECH.2022-26} } ``` ## Disclaimer We did not perform any kind of filtering and/or censorship to the corpus. We expect users to do so applying their own methods. We are not liable for any misuse of the corpus.
graphs-datasets/PROTEINS
--- license: unknown task_categories: - graph-ml --- # Dataset Card for PROTEINS ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [External Use](#external-use) - [PyGeometric](#pygeometric) - [Dataset Structure](#dataset-structure) - [Data Properties](#data-properties) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Additional Information](#additional-information) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **[Homepage](https://academic.oup.com/bioinformatics/article/21/suppl_1/i47/202991)** - **[Repository](https://www.chrsmrrs.com/graphkerneldatasets/PROTEINS.zip):**: - **Paper:**: Protein function prediction via graph kernels (see citation) - **Leaderboard:**: [Papers with code leaderboard](https://paperswithcode.com/sota/graph-classification-on-proteins) ### Dataset Summary The `PROTEINS` dataset is a medium molecular property prediction dataset. ### Supported Tasks and Leaderboards `PROTEINS` should be used for molecular property prediction (aiming to predict whether molecules are enzymes or not), a binary classification task. The score used is accuracy, using a 10-fold cross-validation. ## External Use ### PyGeometric To load in PyGeometric, do the following: ```python from datasets import load_dataset from torch_geometric.data import Data from torch_geometric.loader import DataLoader dataset_hf = load_dataset("graphs-datasets/<mydataset>") dataset_pg_list = [Data(graph) for graph in dataset_hf["train"]] dataset_pg = DataLoader(dataset_pg_list) ``` ## Dataset Structure ### Data Properties | property | value | |---|---| | scale | medium | | #graphs | 1113 | | average #nodes | 39.06 | | average #edges | 72.82 | ### Data Fields Each row of a given file is a graph, with: - `node_feat` (list: #nodes x #node-features): nodes - `edge_index` (list: 2 x #edges): pairs of nodes constituting edges - `edge_attr` (list: #edges x #edge-features): for the aforementioned edges, contains their features - `y` (list: 1 x #labels): contains the number of labels available to predict (here 1, equal to zero or one) - `num_nodes` (int): number of nodes of the graph ### Data Splits This data comes from the PyGeometric version of the dataset provided by TUDataset. This information can be found back using ```python from torch_geometric.datasets import TUDataset dataset = TUDataset(root='', name = 'PROTEINS') ``` ## Additional Information ### Licensing Information The dataset has been released under unknown license, please open an issue if you have info about it. ### Citation Information ``` @article{10.1093/bioinformatics/bti1007, author = {Borgwardt, Karsten M. and Ong, Cheng Soon and Schönauer, Stefan and Vishwanathan, S. V. N. and Smola, Alex J. and Kriegel, Hans-Peter}, title = "{Protein function prediction via graph kernels}", journal = {Bioinformatics}, volume = {21}, number = {suppl_1}, pages = {i47-i56}, year = {2005}, month = {06}, abstract = "{Motivation: Computational approaches to protein function prediction infer protein function by finding proteins with similar sequence, structure, surface clefts, chemical properties, amino acid motifs, interaction partners or phylogenetic profiles. We present a new approach that combines sequential, structural and chemical information into one graph model of proteins. We predict functional class membership of enzymes and non-enzymes using graph kernels and support vector machine classification on these protein graphs.Results: Our graph model, derivable from protein sequence and structure only, is competitive with vector models that require additional protein information, such as the size of surface pockets. If we include this extra information into our graph model, our classifier yields significantly higher accuracy levels than the vector models. Hyperkernels allow us to select and to optimally combine the most relevant node attributes in our protein graphs. We have laid the foundation for a protein function prediction system that integrates protein information from various sources efficiently and effectively.Availability: More information available via www.dbs.ifi.lmu.de/Mitarbeiter/borgwardt.html.Contact:borgwardt@dbs.ifi.lmu.de}", issn = {1367-4803}, doi = {10.1093/bioinformatics/bti1007}, url = {https://doi.org/10.1093/bioinformatics/bti1007}, eprint = {https://academic.oup.com/bioinformatics/article-pdf/21/suppl\_1/i47/524364/bti1007.pdf}, } ``` ### Contributions Thanks to [@clefourrier](https://github.com/clefourrier) for adding this dataset.
nid989/EssayFroum-Dataset
--- license: apache-2.0 ---
Kirili4ik/yandex_jobs
--- annotations_creators: - expert-generated language: - ru language_creators: - found license: - unknown multilinguality: - monolingual paperswithcode_id: climate-fever pretty_name: yandex_jobs size_categories: - n<1K source_datasets: - original tags: - vacancies - jobs - ru - yandex task_categories: - text-generation - summarization - multiple-choice task_ids: - language-modeling --- # Dataset Card for Yandex_Jobs ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Contributions](#contributions) ## Dataset Description ### Dataset Summary This is a dataset of more than 600 IT vacancies in Russian from parsing telegram channel https://t.me/ya_jobs. All the texts are perfectly structured, no missing values. ### Supported Tasks and Leaderboards `text-generation` with the 'Raw text column'. `summarization` as for getting from all the info the header. `multiple-choice` as for the hashtags (to choose multiple from all available in the dataset) ### Languages The text in the dataset is in only in Russian. The associated BCP-47 code is `ru`. ## Dataset Structure ### Data Instances The data is parsed from a vacancy of Russian IT company [Yandex](https://ya.ru/). An example from the set looks as follows: ``` {'Header': 'Разработчик интерфейсов в группу разработки спецпроектов', 'Emoji': '🎳', 'Description': 'Конструктор лендингов — это инструмент Яндекса, который позволяет пользователям создавать лендинги и турбо-лендинги для Яндекс.Директа. Турбо — режим ускоренной загрузки страниц для показа на мобильных. У нас современный стек, смелые планы и высокая динамика.\nМы ищем опытного и открытого новому фронтенд-разработчика.', 'Requirements': '• отлично знаете JavaScript • разрабатывали на Node.js, применяли фреймворк Express • умеете создавать веб-приложения на React + Redux • знаете HTML и CSS, особенности их отображения в браузерах', 'Tasks': '• разрабатывать интерфейсы', 'Pluses': '• писали интеграционные, модульные, функциональные или браузерные тесты • умеете разворачивать и администрировать веб-сервисы: собирать Docker-образы, настраивать мониторинги, выкладывать в облачные системы, отлаживать в продакшене • работали с реляционными БД PostgreSQL', 'Hashtags': '#фронтенд #турбо #JS', 'Link': 'https://ya.cc/t/t7E3UsmVSKs6L', 'Raw text': 'Разработчик интерфейсов в группу разработки спецпроектов🎳 Конструктор лендингов — это инструмент Яндекса, который позволяет пользователям создавать лендинги и турбо-лендинги для Яндекс.Директа. Турбо — режим ускоренной загрузки страниц для показа на мобильных. У нас современный стек, смелые планы и высокая динамика. Мы ищем опытного и открытого новому фронтенд-разработчика. Мы ждем, что вы: • отлично знаете JavaScript • разрабатывали на Node.js, применяли фреймворк Express • умеете создавать веб-приложения на React + Redux • знаете HTML и CSS, особенности их отображения в браузерах Что нужно делать: • разрабатывать интерфейсы Будет плюсом, если вы: • писали интеграционные, модульные, функциональные или браузерные тесты • умеете разворачивать и администрировать веб-сервисы: собирать Docker-образы, настраивать мониторинги, выкладывать в облачные системы, отлаживать в продакшене • работали с реляционными БД PostgreSQL https://ya.cc/t/t7E3UsmVSKs6L #фронтенд #турбо #JS' } ``` ### Data Fields - `Header`: A string with a position title (str) - `Emoji`: Emoji that is used at the end of the title position (usually asosiated with the position) (str) - `Description`: Short description of the vacancy (str) - `Requirements`: A couple of required technologies/programming languages/experience (str) - `Tasks`: Examples of the tasks of the job position (str) - `Pluses`: A couple of great points for the applicant to have (technologies/experience/etc) - `Hashtags`: A list of hashtags assosiated with the job (usually programming languages) (str) - `Link`: A link to a job description (there may be more information, but it is not checked) (str) - `Raw text`: Raw text with all the formatiing from the channel. Created with other fields. (str) ### Data Splits There is not enough examples yet to split it to train/test/val in my opinion. ## Dataset Creation It downloaded and parsed from telegram channel https://t.me/ya_jobs 03.09.2022. All the unparsed examples and the ones missing any field are deleted (from 1600 vacancies to only 600 without any missing fields like emojis or links) ## Considerations for Using the Data These vacancies are for only one IT company (yandex). This means they can be pretty specific and probably can not be generalized as any vacancies or even any IT vacancies. ## Contributions - **Point of Contact and Author:** [Kirill Gelvan](telegram: @kirili4ik)
shjwudp/shu
--- language: zh license: cc-by-4.0 --- 收集中文书籍总计14363本,用于学术研究和工业生产使用,书籍持续收录中,参与贡献请移步[代码仓库](https://github.com/shjwudp/shu)。 The dataset constructed from Chinese books. Books are being collected continuously. Please move to [code warehouse](https://github.com/shjwudp/shu) to contribute.
dougtrajano/olid-br
--- language: pt license: cc-by-4.0 dataset_info: features: - name: id dtype: string - name: text dtype: string - name: is_offensive dtype: string - name: is_targeted dtype: string - name: targeted_type dtype: string - name: toxic_spans sequence: int64 - name: health dtype: bool - name: ideology dtype: bool - name: insult dtype: bool - name: lgbtqphobia dtype: bool - name: other_lifestyle dtype: bool - name: physical_aspects dtype: bool - name: profanity_obscene dtype: bool - name: racism dtype: bool - name: religious_intolerance dtype: bool - name: sexism dtype: bool - name: xenophobia dtype: bool splits: - name: train num_bytes: 1763684 num_examples: 5214 - name: test num_bytes: 590953 num_examples: 1738 download_size: 1011742 dataset_size: 2354637 --- # OLID-BR Offensive Language Identification Dataset for Brazilian Portuguese (OLID-BR) is a dataset with multi-task annotations for the detection of offensive language. The current version (v1.0) contains **7,943** (extendable to 13,538) comments from different sources, including social media (YouTube and Twitter) and related datasets. OLID-BR contains a collection of annotated sentences in Brazilian Portuguese using an annotation model that encompasses the following levels: - [Offensive content detection](#offensive-content-detection): Detect offensive content in sentences and categorize it. - [Offense target identification](#offense-target-identification): Detect if an offensive sentence is targeted to a person or group of people. - [Offensive spans identification](#offensive-spans-identification): Detect curse words in sentences. ![](https://dougtrajano.github.io/olid-br/images/olid-br-taxonomy.png) ## Categorization ### Offensive Content Detection This level is used to detect offensive content in the sentence. **Is this text offensive?** We use the [Perspective API](https://www.perspectiveapi.com/) to detect if the sentence contains offensive content with double-checking by our [qualified annotators](annotation/index.en.md#who-are-qualified-annotators). - `OFF` Offensive: Inappropriate language, insults, or threats. - `NOT` Not offensive: No offense or profanity. **Which kind of offense does it contain?** The following labels were tagged by our annotators: `Health`, `Ideology`, `Insult`, `LGBTQphobia`, `Other-Lifestyle`, `Physical Aspects`, `Profanity/Obscene`, `Racism`, `Religious Intolerance`, `Sexism`, and `Xenophobia`. See the [**Glossary**](glossary.en.md) for further information. ### Offense Target Identification This level is used to detect if an offensive sentence is targeted to a person or group of people. **Is the offensive text targeted?** - `TIN` Targeted Insult: Targeted insult or threat towards an individual, a group or other. - `UNT` Untargeted: Non-targeted profanity and swearing. **What is the target of the offense?** - `IND` The offense targets an individual, often defined as “cyberbullying”. - `GRP` The offense targets a group of people based on ethnicity, gender, sexual - `OTH` The target can belong to other categories, such as an organization, an event, an issue, etc. ### Offensive Spans Identification As toxic spans, we define a sequence of words that attribute to the text's toxicity. For example, let's consider the following text: > "USER `Canalha` URL" The toxic spans are: ```python [5, 6, 7, 8, 9, 10, 11, 12, 13] ``` ## Dataset Structure ### Data Instances Each instance is a social media comment with a corresponding ID and annotations for all the tasks described below. ### Data Fields The simplified configuration includes: - `id` (string): Unique identifier of the instance. - `text` (string): The text of the instance. - `is_offensive` (string): Whether the text is offensive (`OFF`) or not (`NOT`). - `is_targeted` (string): Whether the text is targeted (`TIN`) or untargeted (`UNT`). - `targeted_type` (string): Type of the target (individual `IND`, group `GRP`, or other `OTH`). Only available if `is_targeted` is `True`. - `toxic_spans` (string): List of toxic spans. - `health` (boolean): Whether the text contains hate speech based on health conditions such as disability, disease, etc. - `ideology` (boolean): Indicates if the text contains hate speech based on a person's ideas or beliefs. - `insult` (boolean): Whether the text contains insult, inflammatory, or provocative content. - `lgbtqphobia` (boolean): Whether the text contains harmful content related to gender identity or sexual orientation. - `other_lifestyle` (boolean): Whether the text contains hate speech related to life habits (e.g. veganism, vegetarianism, etc.). - `physical_aspects` (boolean): Whether the text contains hate speech related to physical appearance. - `profanity_obscene` (boolean): Whether the text contains profanity or obscene content. - `racism` (boolean): Whether the text contains prejudiced thoughts or discriminatory actions based on differences in race/ethnicity. - `religious_intolerance` (boolean): Whether the text contains religious intolerance. - `sexism` (boolean): Whether the text contains discriminatory content based on differences in sex/gender (e.g. sexism, misogyny, etc.). - `xenophobia` (boolean): Whether the text contains hate speech against foreigners. See the [**Get Started**](get-started.en.md) page for more information. ## Considerations for Using the Data ### Social Impact of Dataset Toxicity detection is a worthwhile problem that can ensure a safer online environment for everyone. However, toxicity detection algorithms have focused on English and do not consider the specificities of other languages. This is a problem because the toxicity of a comment can be different in different languages. Additionally, the toxicity detection algorithms focus on the binary classification of a comment as toxic or not toxic. Therefore, we believe that the OLID-BR dataset can help to improve the performance of toxicity detection algorithms in Brazilian Portuguese. ### Discussion of Biases We are aware that the dataset contains biases and is not representative of global diversity. We are aware that the language used in the dataset could not represent the language used in different contexts. Potential biases in the data include: Inherent biases in the social media and user base biases, the offensive/vulgar word lists used for data filtering, and inherent or unconscious bias in the assessment of offensive identity labels. All these likely affect labeling, precision, and recall for a trained model. ## Citation Pending
ghoumrassi/clothes_sample
--- dataset_info: features: - name: image dtype: image - name: text dtype: string splits: - name: train num_bytes: 20078406.0 num_examples: 990 download_size: 0 dataset_size: 20078406.0 --- # Dataset Card for "clothes_sample" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
giulio98/xlcost-single-prompt
--- annotations_creators: [] language_creators: - crowdsourced - expert-generated language: - code license: - cc-by-sa-4.0 multilinguality: - multilingual size_categories: - unknown source_datasets: [] task_categories: - text-generation task_ids: - language-modeling pretty_name: xlcost-single-prompt --- # XLCost for text-to-code synthesis ## Dataset Description This is a subset of [XLCoST benchmark](https://github.com/reddy-lab-code-research/XLCoST), for text-to-code generation at program level for **2** programming languages: `Python, C++`. This dataset is based on [codeparrot/xlcost-text-to-code](https://huggingface.co/datasets/codeparrot/xlcost-text-to-code) with the following improvements: * NEWLINE, INDENT and DEDENT were replaced with the corresponding ASCII codes. * the code text has been reformatted using autopep8 for Python and clang-format for cpp. * new columns have been introduced to allow evaluation using pass@k metric. * programs containing more than one function call in the driver code were removed ## Languages The dataset contains text in English and its corresponding code translation. The text contains a set of concatenated code comments that allow to synthesize the program. ## Dataset Structure To load the dataset you need to specify the language(Python or C++). ```python from datasets import load_dataset load_dataset("giulio98/xlcost-single-prompt", "Python") DatasetDict({ train: Dataset({ features: ['text', 'context', 'code', 'test', 'output', 'fn_call'], num_rows: 8306 }) test: Dataset({ features: ['text', 'context', 'code', 'test', 'output', 'fn_call'], num_rows: 812 }) validation: Dataset({ features: ['text', 'context', 'code', 'test', 'output', 'fn_call'], num_rows: 427 }) }) ``` ## Data Fields * text: natural language description. * context: import libraries/global variables. * code: code at program level. * test: test function call. * output: expected output of the function call. * fn_call: name of the function to call. ## Data Splits Each subset has three splits: train, test and validation. ## Citation Information ``` @misc{zhu2022xlcost, title = {XLCoST: A Benchmark Dataset for Cross-lingual Code Intelligence}, url = {https://arxiv.org/abs/2206.08474}, author = {Zhu, Ming and Jain, Aneesh and Suresh, Karthik and Ravindran, Roshan and Tipirneni, Sindhu and Reddy, Chandan K.}, year = {2022}, eprint={2206.08474}, archivePrefix={arXiv} } ```
chattermill/fabsa
--- license: mit ---
camenduru/plushies
--- dataset_info: features: - name: image dtype: image - name: text dtype: string splits: - name: train num_bytes: 42942055.0 num_examples: 730 download_size: 42653871 dataset_size: 42942055.0 models: - camenduru/plushies ---
dreamproit/bill_summary_us
--- annotations_creators: - expert-generated language: - en language_creators: - expert-generated multilinguality: - monolingual pretty_name: bill_summary_us size_categories: - 100K<n<1M source_datasets: - original tags: - bills - legal task_categories: - summarization task_ids: [] configs: - config_name: default --- # Dataset Card for "bill_summary_us" ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [BillML](https://github.com/dreamproit/BillML) - **Repository:** [BillML](https://github.com/dreamproit/BillML) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Leaderboard:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Dataset Summary Dataset for summarization of summarization of US Congressional bills (bill_summary_us). ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages English ## Dataset Structure ### Data Instances #### default ### Data Fields - id: id of the bill in format(congress number + bill type + bill number + bill version). - congress: number of the congress. - bill_type: type of the bill. - bill_number: number of the bill. - bill_version: version of the bill. - sections: list of bill sections with section_id, text and header. - sections_length: number with lenght of the sections list. - text: bill text. - text_length: number of characters in the text. - summary: summary of the bill. - summary_length: number of characters in the summary. - title: official title of the bill. ### Data Splits train ## Dataset Creation ### Curation Rationale Bills (proposed laws) are specialized, structured documents with great public significance. Often, the language of a bill may not directly explain the potential impact of the legislation. For bills in the U.S. Congress, the Congressional Research Service of the Library of Congress provides professional, non-partisan summaries of bills. These are valuable for public understanding of the bills and are serve as an essential part of the lawmaking process to understand the meaning and potential legislative impact. This dataset collects the text of bills, some metadata, as well as the CRS summaries. In order to build more accurate ML models for bill summarization it is important to have a clean dataset, alongside the professionally-written CRS summaries. ML summarization models built on generic data are bound to produce less accurate results (sometimes creating summaries that describe the opposite of a bill's actual effect). In addition, models that attempt to summarize all bills (some of which may reach 4000 pages long) may also be inaccurate due to the current limitations of summarization on long texts. As a result, this dataset collects bill and summary information; it provides text as a list of sections with the text and header. This could be used to create a summary of sections and then a summary of summaries. [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data [govinfo.gov](https://www.govinfo.gov/) #### Initial Data Collection and Normalization The data consists of the US congress bills that were collected from the [govinfo.gov](https://www.govinfo.gov/) service provided by the United States Government Publishing Office (GPO) under CC0-1.0 license. #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [dreamproit.com](https://dreamproit.com/) ### Licensing Information Bill and summary information are public and are unlicensed, as it is data produced by government entities. The collection and enhancement work that we provide for this dataset, to the degree it may be covered by copyright, is released under [CC0](https://creativecommons.org/share-your-work/public-domain/cc0/). ### Citation Information [More Information Needed] ### Contributions Thanks to [@aih](https://github.com/aih) [@BorodaUA](https://github.com/BorodaUA), [@alexbojko](https://github.com/alexbojko) for adding this dataset.
pszemraj/text2image-multi-prompt
--- language: - en license: apache-2.0 multilinguality: - monolingual source_datasets: - bartman081523/stable-diffusion-discord-prompts - succinctly/midjourney-prompts - Gustavosta/Stable-Diffusion-Prompts pretty_name: multi text2image prompts a dataset collection tags: - text generation configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* - config_name: original data_files: - split: train path: original/train-* - split: test path: original/test-* dataset_info: - config_name: default features: - name: text dtype: string - name: src_dataset dtype: string splits: - name: train num_bytes: 262736830 num_examples: 1677221 - name: test num_bytes: 56294291 num_examples: 292876 download_size: 151054782 dataset_size: 319031121 - config_name: original features: - name: text dtype: string - name: src_dataset dtype: string splits: - name: train num_bytes: 741427383 num_examples: 3551734 - name: test num_bytes: 83615440 num_examples: 399393 download_size: 402186258 dataset_size: 825042823 task_categories: - text-generation - feature-extraction --- # text2image multi-prompt(s): a dataset collection - collection of several text2image prompt datasets - data was cleaned/normalized with the goal of removing "model specific APIs" like the "--ar" for Midjourney and so on - data de-duplicated on a basic level: exactly duplicate prompts were dropped (_after cleaning and normalization_) ## updates - Oct 2023: the `default` config has been updated with better deduplication. It was deduplicated with minhash (_params: n-gram size set to 3, deduplication threshold at 0.6, hash function chosen as xxh3 with 32-bit hash bits, and 128 permutations with a batch size of 10,000._) which drops 2+ million rows. - original version is still available under `config_name="original"` ## contents default: ``` DatasetDict({ train: Dataset({ features: ['text', 'src_dataset'], num_rows: 1677221 }) test: Dataset({ features: ['text', 'src_dataset'], num_rows: 292876 }) }) ``` For `original` config: ``` DatasetDict({ train: Dataset({ features: ['text', 'src_dataset'], num_rows: 3551734 }) test: Dataset({ features: ['text', 'src_dataset'], num_rows: 399393 }) }) ``` _NOTE: as the other two datasets did not have a `validation` split, the validation split of `succinctly/midjourney-prompts` was merged into `train`._
Guizmus/AnimeChanStyle
--- license: creativeml-openrail-m thumbnail: "https://huggingface.co/datasets/Guizmus/AnimeChanStyle/resolve/main/showcase_dataset.jpg" --- ![showcase](https://huggingface.co/datasets/Guizmus/AnimeChanStyle/resolve/main/showcase_dataset.jpg) This is the dataset used for making the model : https://huggingface.co/Guizmus/AnimeChanStyle The images were made by the users of Stable Diffusion discord using CreativeML-OpenRail-M licenced models, in the intent to make this dataset. 90 pictures captioned with their content by hand, with the suffix ",AnimeChan Style" The collection process was made public during less than a day, until enough variety was introduced to train through a Dreambooth method a style corresponding to the different members of this community The picture captioned are available in [this zip file](https://huggingface.co/datasets/Guizmus/AnimeChanStyle/resolve/main/AnimeChanStyle%20v2.3.zip)
Drozdik/tattoo_v0
--- license: cc-by-nc-sa-4.0 annotations_creators: - machine-generated language: - en language_creators: - other multilinguality: - monolingual pretty_name: 'Tattoo BLIP caption' size_categories: - n<1K source_datasets: - Drozdik/tattoo_v0 tags: [] task_categories: - text-to-image task_ids: [] ---
ashraf-ali/quran-data
--- language_creators: - Tarteel.io license: - cc0-1.0 size_categories: ar: - 43652 task_categories: - automatic-speech-recognition task_ids: [] paperswithcode_id: quran-data pretty_name: Quran Audio language_bcp47: - ar --- # Dataset Card for Quran audio Content * 7 Imam Full Quran Recitation: 7*6236 wav file - csv contains the Text info for 11k subset short wav file * Tarteel.io user dataset ~25k wav - csv contains the Text info for 18k subset of the accepted user quality
mertbozkurt/quotes_philosophers
--- license: afl-3.0 --- # About Dataset Philosophers Quotes from azquotes.com * Arthur Schopenhauer 400+ quotes * Friedrich Nietzsche 200+ quotes * Immanuel Kant 300+ quotes * Aristotle 350+ quotes * Plato 70+ quotes * Sigmund Freud 400+ quotes * Hegel 120+ quotes * Jean Paul Sartre 320+ quotes * Spinoza 120+ quotes ### COLLECTION METHODOLOGY Python Web Scraping with Selenium
jhu-clsp/bernice-pretrain-data
--- annotations_creators: - no-annotation language: - en - es - pt - ja - ar - in - ko - tr - fr - tl - ru - it - th - de - hi - pl - nl - fa - et - ht - ur - sv - ca - el - fi - cs - iw - da - vi - zh - ta - ro - no - uk - cy - ne - hu - eu - sl - lv - lt - bn - sr - bg - mr - ml - is - te - gu - kn - ps - ckb - si - hy - or - pa - am - sd - my - ka - km - dv - lo - ug - bo language_creators: - found license: - mit multilinguality: - multilingual pretty_name: Bernice Pretrain Data size_categories: - 1B<n<10B source_datasets: - original tags: - twitter - slang - code switch - social - social media task_categories: - other task_ids: [] --- # Dataset Card for Bernice Pre-train Data ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** N/A - **Repository:** https://github.com/JHU-CLSP/Bernice-Twitter-encoder - **Paper:** _Bernice: A Multilingual Pre-trained Encoder for Twitter_ at [EMNLP 2022](https://preview.aclanthology.org/emnlp-22-ingestion/2022.emnlp-main.415) - **Leaderboard:** N/A - **Point of Contact:** Alexandra DeLucia aadelucia (at) jhu.edu ### Dataset Summary Tweet IDs for the 2.5 billion multilingual tweets used to train Bernice, a Twitter encoder. Read the paper [here](https://preview.aclanthology.org/emnlp-22-ingestion/2022.emnlp-main.415). The tweets are from the public 1% Twitter API stream from January 2016 to December 2021. Twitter-provided language metadata is provided with the tweet ID. The data contains 66 unique languages, as identified by [ISO 639 language codes](https://www.wikiwand.com/en/List_of_ISO_639-1_codes), including `und` for undefined languages. Tweets need to be re-gathered via the Twitter API. We suggest [Hydrator](https://github.com/DocNow/hydrator) or [tweepy](https://www.tweepy.org/). To load with HuggingFace: ```python from datasets import load_dataset dataset = load_dataset("jhu-clsp/bernice-pretrain-data") for i, row in enumerate(dataset["train"]): print(row) if i > 10: break ``` If you only want Indic languages, use ```python dataset = load_dataset("jhu-clsp/bernice-pretrain-data", "indic") ``` ### Supported Tasks and Leaderboards N/A ### Languages 65 languages (ISO 639 codes shown below), plus an `und` (undefined) category. All language identification provided by Twitter API. | | | | | | | | |----|-----|----|----|----|-----|----| | en | ru | ht | zh | bn | ps | lt | | es | bo | ur | ta | sr | ckb | km | | pt | it | sv | ro | bg | si | dv | | ja | th | ca | no | mr | hy | lo | | ar | de | el | uk | ml | or | ug | | in | hi | fi | cy | is | pa | | | ko | pl | cs | ne | te | am | | | tr | nl | iw | hu | gu | sd | | | fr | fa | da | eu | kn | my | | | tl | et | vi | sl | lv | ka | | ## Dataset Structure ### Data Instances Data is provided in gzip'd files organized by year and month of tweet origin. Tweets are one per line, with fields separated by tabs. ### Data Fields * `tweet ID`: ID of tweet * `lang`: ISO 639 code of language, provided by Twitter metadata. Accuracy of label is not known. * `year`: Year tweet was created. Year is also provided in the file names. ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale Data was gathered to support the training of Bernice, a multilingual pre-trained Twitter encoder. ### Source Data #### Initial Data Collection and Normalization Data was gathered via the Twitter API public 1% stream from January 2016 through December 2021. Tweets with less than three non-username or URL space-delimited words were removed. All usernames and URLs were replaced with `@USER` and `HTTPURL`, respectively. #### Who are the source language producers? Data was produced by users on Twitter. ### Annotations N/A ### Personal and Sensitive Information As per Twitter guidelines, only tweet IDs and not full tweets are shared. Tweets will only be accessible if user has not removed their account (or been banned), tweets were deleted or removed, or a user changed their account access to private. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators Dataset gathered and processed by Mark Dredze, Alexandra DeLucia, Shijie Wu, Aaron Mueller, Carlos Aguirre, and Philip Resnik. ### Licensing Information MIT ### Citation Information Please cite the Bernice paper if you use this dataset: > Alexandra DeLucia, Shijie Wu, Aaron Mueller, Carlos Aguirre, Philip Resnik, and Mark Dredze. 2022. Bernice: A Multilingual Pre-trained Encoder for Twitter. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 6191–6205, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics. ### Contributions Dataset uploaded by [@AADeLucia](https://github.com/AADeLucia).
reshinthadith/pairwise-code-review-instruct-critique-revision-python
--- dataset_info: features: - name: prompt dtype: string - name: chosen dtype: string - name: rejected dtype: string - name: chosen_score dtype: string - name: rejected_score dtype: string splits: - name: train num_bytes: 35079153 num_examples: 5236 download_size: 9344129 dataset_size: 35079153 --- # Dataset Card for "pairwise-code-review-instruct-critique-revision-python" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
keremberke/pothole-segmentation
--- task_categories: - image-segmentation tags: - roboflow - roboflow2huggingface - Construction - Self Driving - Transportation - Damage Risk --- <div align="center"> <img width="640" alt="keremberke/pothole-segmentation" src="https://huggingface.co/datasets/keremberke/pothole-segmentation/resolve/main/thumbnail.jpg"> </div> ### Dataset Labels ``` ['pothole'] ``` ### Number of Images ```json {'test': 5, 'train': 80, 'valid': 5} ``` ### How to Use - Install [datasets](https://pypi.org/project/datasets/): ```bash pip install datasets ``` - Load the dataset: ```python from datasets import load_dataset ds = load_dataset("keremberke/pothole-segmentation", name="full") example = ds['train'][0] ``` ### Roboflow Dataset Page [https://universe.roboflow.com/imacs-pothole-detection-wo8mu/pothole-detection-irkz9/dataset/4](https://universe.roboflow.com/imacs-pothole-detection-wo8mu/pothole-detection-irkz9/dataset/4?ref=roboflow2huggingface) ### Citation ``` @misc{ pothole-detection-irkz9_dataset, title = { Pothole Detection Dataset }, type = { Open Source Dataset }, author = { IMACS Pothole Detection }, howpublished = { \\url{ https://universe.roboflow.com/imacs-pothole-detection-wo8mu/pothole-detection-irkz9 } }, url = { https://universe.roboflow.com/imacs-pothole-detection-wo8mu/pothole-detection-irkz9 }, journal = { Roboflow Universe }, publisher = { Roboflow }, year = { 2023 }, month = { jan }, note = { visited on 2023-01-15 }, } ``` ### License CC BY 4.0 ### Dataset Summary This dataset was exported via roboflow.com on January 15, 2023 at 6:38 PM GMT Roboflow is an end-to-end computer vision platform that helps you * collaborate with your team on computer vision projects * collect & organize images * understand and search unstructured image data * annotate, and create datasets * export, train, and deploy computer vision models * use active learning to improve your dataset over time For state of the art Computer Vision training notebooks you can use with this dataset, visit https://github.com/roboflow/notebooks To find over 100k other datasets and pre-trained models, visit https://universe.roboflow.com The dataset includes 90 images. Pothole are annotated in COCO format. The following pre-processing was applied to each image: No image augmentation techniques were applied.
rcds/swiss_legislation
--- license: cc-by-sa-4.0 task_categories: - text-classification - translation language: - de - fr - it pretty_name: Swiss Legislation size_categories: - 100K<n<1M --- # Dataset Card for Swiss Legislation ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Swiss Legislation is a multilingual, diachronic dataset of 36K Swiss laws. This dataset is part of a challenging Information Retreival task. ### Supported Tasks and Leaderboards ### Languages The total number of texts in the dataset is 35,698. The dataset is saved in _lexfind_v2.jsonl_ format. Switzerland has four official languages German, French, Italian and Romanch with some additional English laws being represenated. Laws are written by legal experts. 36K & 18K & 11K & 6K & 534 & 207 | Language | Subset | Number of Documents | |------------|------------|----------------------| | German | **de** | 18K | | French | **fr** | 11K | | Italian | **it** | 6K | | Romanch | **rm** | 534 | | English | **en** | 207 | ## Dataset Structure ### Data Fields Each entry in the dataset is a dictionary with the following keys: - `canton`: the canton of origin of the legislation - example: "ag" - `language`: the language of the legislation - example: "de" - `uuid`: a unique identifier for the legislation - example: "ec312f57-05fe-4552-ba50-8c9c269e0f3b" - `title`: the title of the legislation - example: "Gesetz über die Geoinformation im Kanton Aargau" - `short`: a short description of the legislation - example: "Kantonales Geoinformationsgesetz" - `abbreviation`: an abbreviation for the legislation - example: "KGeoIG" - `sr_number`: a reference number for the legislation - example: "740.100" - `is_active`: whether the legislation is currently in force - example: true - `version_active_since`: the date since when the legislation's current version is active - example: "2021-09-01" - `family_active_since`: the date since when the legislation's current version's family is active - example: "2011-05-24" - `version_inactive_since`: the date since when the legislation's current version is inactive - example: null - `version_found_at`: the date the legislation's current version was found - example: "2021-09-01" - `pdf_url`: a link to the legislation's pdf - example: "https://www.lexfind.ch/tol/1557/de" - `html_url`: a link to the legislation's html - example: "https://gesetzessammlungen.ag.ch/app/de/texts_of_law/740.100")_ - `pdf_content`: the legislation's pdf content - example: "740.100 - Gesetz über..." - `html_content`: the legislation's html content - example: "" - `changes`: a list of changes made to the legislation - example: [] - `history`: a list of the legislation's history - example: [] - `quotes`: a list of quotes from the legislation - example: [] ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits 1. 'ch': Switzerland (Federal) - 15840 2. 'fr': Fribourg - 1633 3. 'be': Bern - 1344 4. 'vs': Valais - 1328 5. 'gr': Graubünden - 1205 6. 'ne': Neuchâtel - 1115 7. 'zh': Zurich - 974 8. 'bs': Basel-Stadt - 899 9. 'bl': Basel-Landschaft - 863 10. 'vd': Vaud - 870 11. 'ge': Geneva - 837 12. 'sg': St. Gallen - 764 13. 'ju': Jura - 804 14. 'zg': Zug - 632 15. 'ti': Ticino - 627 16. 'lu': Lucerne - 584 17. 'so': Solothurn - 547 18. 'ow': Obwalden - 513 19. 'ik': Interkantonal - 510 20. 'sh': Schaffhausen - 469 21. 'gl': Glarus - 467 22. 'tg': Thurgau - 453 23. 'sz': Schwyz - 423 24. 'ai': Appenzell Innerrhoden - 416 25. 'ag': Aargau - 483 26. 'ar': Appenzell Ausserrhoden - 330 27. 'nw': Nidwalden - 401 28. 'ur': Uri - 367 29. ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization The original data are published from the Swiss Federal Supreme Court (https://www.bger.ch) in unprocessed formats (HTML). The documents were downloaded from the Entscheidsuche portal (https://entscheidsuche.ch) in HTML. #### Who are the source language producers? The decisions are written by the judges and clerks in the language of the proceedings. ### Annotations #### Annotation process #### Who are the annotators? Metadata is published by the Swiss Federal Supreme Court (https://www.bger.ch). ### Personal and Sensitive Information The dataset contains publicly available court decisions from the Swiss Federal Supreme Court. Personal or sensitive information has been anonymized by the court before publication according to the following guidelines: https://www.bger.ch/home/juridiction/anonymisierungsregeln.html. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information We release the data under CC-BY-4.0 which complies with the court licensing (https://www.bger.ch/files/live/sites/bger/files/pdf/de/urteilsveroeffentlichung_d.pdf) © Swiss Federal Supreme Court, 2002-2022 The copyright for the editorial content of this website and the consolidated texts, which is owned by the Swiss Federal Supreme Court, is licensed under the Creative Commons Attribution 4.0 International licence. This means that you can re-use the content provided you acknowledge the source and indicate any changes you have made. Source: https://www.bger.ch/files/live/sites/bger/files/pdf/de/urteilsveroeffentlichung_d.pdf ### Citation Information Please cite our [ArXiv-Preprint](https://arxiv.org/abs/2306.09237) ``` @misc{rasiah2023scale, title={SCALE: Scaling up the Complexity for Advanced Language Model Evaluation}, author={Vishvaksenan Rasiah and Ronja Stern and Veton Matoshi and Matthias Stürmer and Ilias Chalkidis and Daniel E. Ho and Joel Niklaus}, year={2023}, eprint={2306.09237}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
tomekkorbak/pile-pii-scrubadub
--- annotations_creators: - machine-generated language: - en language_creators: - found license: - mit multilinguality: - monolingual pretty_name: pile-pii-scrubadub size_categories: - 1M<n<10M source_datasets: - extended|the_pile tags: - pii - personal - identifiable - information - pretraining-with-human-feedback task_categories: - text-classification - other task_ids: - acceptability-classification - text-scoring --- # Dataset Card for pile-pii-scrubadub ## Dataset Description - **Repository: https://github.com/tomekkorbak/aligned-pretraining-objectives** - **Paper: Arxiv link to be added** ### Dataset Summary This dataset contains text from [The Pile](https://huggingface.co/datasets/the_pile), annotated based on the personal idenfitiable information (PII) in each sentence. Each document (row in the dataset) is segmented into sentences, and each sentence is given a score: the percentage of words in it that are classified as PII by [Scrubadub](https://scrubadub.readthedocs.io/en/stable/). ### Supported Tasks and Leaderboards [More Information Needed] ### Languages This dataset is taken from [The Pile](https://huggingface.co/datasets/the_pile), which is English text. ## Dataset Structure ### Data Instances 1949977 ### Data Fields - texts (sequence): a list of the sentences in the document (segmented using [SpaCy](https://spacy.io/)) - meta (dict): the section of [The Pile](https://huggingface.co/datasets/the_pile) from which it originated - scores (sequence): a score for each sentence in the `texts` column indicating the percent of words that are detected as PII by [Scrubadub](https://scrubadub.readthedocs.io/en/stable/) - avg_score (float64): the average of the scores listed in the `scores` column - num_sents (int64): the number of sentences (and scores) in that document ### Data Splits Training set only ## Dataset Creation ### Curation Rationale This is labeled text from [The Pile](https://huggingface.co/datasets/the_pile), a large dataset of text in English. The PII is labeled so that generative language models can be trained to avoid generating PII. ### Source Data #### Initial Data Collection and Normalization This is labeled text from [The Pile](https://huggingface.co/datasets/the_pile). #### Who are the source language producers? Please see [The Pile](https://huggingface.co/datasets/the_pile) for the source of the dataset. ### Annotations #### Annotation process For each sentence, [Scrubadub](https://scrubadub.readthedocs.io/en/stable/) was used to detect: - email addresses - addresses and postal codes - phone numbers - credit card numbers - US social security numbers - vehicle plates numbers - dates of birth - URLs - login credentials #### Who are the annotators? [Scrubadub](https://scrubadub.readthedocs.io/en/stable/) ### Personal and Sensitive Information This dataset contains all PII that was originally contained in [The Pile](https://huggingface.co/datasets/the_pile), with all detected PII annotated. ## Considerations for Using the Data ### Social Impact of Dataset This dataset contains examples of real PII (conveniently annotated in the text!). Please take care to avoid misusing it or putting anybody in danger by publicizing their information. This dataset is intended for research purposes only. We cannot guarantee that all PII has been detected, and we cannot guarantee that models trained using it will avoid generating PII. We do not recommend deploying models trained on this data. ### Discussion of Biases This dataset contains all biases from The Pile discussed in their paper: https://arxiv.org/abs/2101.00027 ### Other Known Limitations The PII in this dataset was detected using imperfect automated detection methods. We cannot guarantee that the labels are 100% accurate. ## Additional Information ### Dataset Curators [The Pile](https://huggingface.co/datasets/the_pile) ### Licensing Information From [The Pile](https://huggingface.co/datasets/the_pile): PubMed Central: [MIT License](https://github.com/EleutherAI/pile-pubmedcentral/blob/master/LICENSE) ### Citation Information Paper information to be added ### Contributions [The Pile](https://huggingface.co/datasets/the_pile)
tasksource/scinli
--- license: apache-2.0 --- #SciNLI: A Corpus for Natural Language Inference on Scientific Text https://github.com/msadat3/SciNLI ```bib @inproceedings{sadat-caragea-2022-scinli, title = "{S}ci{NLI}: A Corpus for Natural Language Inference on Scientific Text", author = "Sadat, Mobashir and Caragea, Cornelia", booktitle = "Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = may, year = "2022", address = "Dublin, Ireland", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2022.acl-long.511", pages = "7399--7409", } ```
mlfoundations/datacomp_pools
--- license: cc-by-4.0 --- ## DataComp Pools This repository contains metadata files for DataComp. For details on how to use the metadata, please visit [our website](https://www.datacomp.ai/) and our [github repository](https://github.com/mlfoundations/datacomp). We distribute the image url-text samples and metadata under a standard Creative Common CC-BY-4.0 license. The individual images are under their own copyrights. ## Terms and Conditions We have terms of service that are similar to those adopted by HuggingFace (https://huggingface.co/terms-of-service), which covers their dataset library. Specifically, any content you download, access or use from our index, is at your own risk and subject to the terms of service or copyright limitations accompanying such content. The image url-text index, which is a research artifact, is provided as is. By using said index, you assume all risks, including but not limited to, liabilities related to image downloading and storage.
leobertolazzi/ita2medieval
--- task_categories: - text2text-generation language: - it size_categories: - 1K<n<10K --- ## ita2medieval The **ita2medieval** dataset contains sentences from medieval italian along with paraphrases in contemporary italian (approximately 6.5k pairs in total). The medieval italian sentences are extracted from texts by Dante, Petrarca, Guinizelli and Cavalcanti. It is intended to perform text-style-transfer from contemporary to medieval italian and vice-versa. ## Loading the dataset ``` from datasets import load_dataset dataset = load_dataset("leobertolazzi/ita2medieval") ``` Note: due to the small size of the dataset there are no predefined train and test splits. ## Dataset creation **ita2medieval** was created by scraping [letteritaliana.weebly.com](https://letteritaliana.weebly.com/).
isaacrehg/poetry-instructions
--- dataset_info: features: - name: conversation dtype: string splits: - name: train num_bytes: 87758119 num_examples: 1322 - name: validation num_bytes: 7731418 num_examples: 111 - name: test num_bytes: 27041394 num_examples: 331 download_size: 63044464 dataset_size: 122530931 --- # Dataset Card for "poetry-instructions" A dataset of user-assistant dialogue instructions for guided poetry creation. Poems used were taken from [merve/poetry](https://huggingface.co/datasets/merve/poetry) and [matthh/gutenberg-poetry-corpus](https://huggingface.co/datasets/matthh/gutenberg-poetry-corpus). The dataset contains dialogues in the following formats: - Poetry Completion: ``` User: Can you continue this poem for me? <poem_start> Assistant: Sure, a continuation for this poem could be: <poem end> ``` - Create poem in style of (?): ``` User: Can you write a poem for me in the style of <author>? Assistant: Sure, here's a poem in the style of <author>: <poem> ``` - Creat poem about (?): ``` User: Can you write me a poem about <keywords (extracted using keyphrase model)>? Assistant: Sure, here's a poem about <keywords>: <poem> ``` - Create poem about (?) in the style of (?): ``` User: Can you write me a poem about <keywords> in the style of <author>? Assistant: Sure, here's a poem about <keywords> in the style of <author>: <poem> ```