--- license: apache-2.0 task_categories: - question-answering - conversational language: - en tags: - evaluation pretty_name: MT Bench size_categories: - n<1K --- # MT Bench by LMSYS This set of evaluation prompts is created by the [LMSYS org](https://huggingface.co/lmsys) for better evaluation of chat models. For more information, see the [paper](https://arxiv.org/abs/2306.05685). ### Dataset loading To load this dataset, use 🤗 datasets: ```python from datasets import load_dataset data = load_dataset(HuggingFaceH4/mt_bench_prompts, split="train") ``` ### Dataset creation To create the dataset, we do the following for our internal tooling. * rename `turns` to `prompts`, * add empty `reference` to remaining prompts (for HF Datasets), * Use the following code to load and save as a dataset ```python from datasets import load_dataset import hashlib data = load_dataset("json", data_files="https://huggingface.co/datasets/HuggingFaceH4/mt_bench_prompts/raw/main/raw/question.jsonl", split="train") # %% create_dataset.ipynb 11 def format_example(example): return { "prompt": example["prompt"], "prompt_id": int(hashlib.sha256(''.join(example["prompt"]).encode("utf-8")).hexdigest(), 16) % (10 ** 8), "category": example["category"], "reference": example["reference"], } formatted_ds = data.map(format_example, num_proc=6, remove_columns=data.column_names) # formatted_ds.push_to_hub("HuggingFaceH4/mt_bench_prompts", split="train") ```