--- license: mit dataset_info: features: - name: id dtype: string - name: question dtype: string - name: context dtype: string - name: A dtype: string - name: B dtype: string - name: C dtype: string - name: D dtype: string - name: label dtype: int64 splits: - name: train num_bytes: 63759933.69322235 num_examples: 2517 - name: test num_bytes: 52057383.0 num_examples: 2086 download_size: 19849080 dataset_size: 115817316.69322234 configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* --- This dataset is derived from `tau/scrolls` [dataset](tau/scrolls) by running the following script: ```python import re from datasets import load_dataset quality_dataset = load_dataset("tau/scrolls", "quality") def parse_example(example): text = example["input"] options = dict(re.findall(r"\((A|B|C|D)\) ([^\n]+)", text)) question_part, context = re.split(r"\(D\) [^\n]+\n", text, maxsplit=1) question = re.sub(r"\([A-D]\) [^\n]+\n?", "", question_part).strip() result = {"question": question, "context": context.strip(), **options} if not all(key in result for key in ["A", "B", "C", "D"]): raise ValueError("One or more options (A, B, C, D) are missing!") # get label label = -1 answer = example["output"] if answer is None: answer = "" for idx, option in enumerate([options["A"], options["B"], options["C"], options["D"]]): if answer.strip() == option.strip(): label = idx result["label"] = label return result quality_dataset = quality_dataset.map(parse_example) quality_dataset = quality_dataset.filter(lambda x: x["label"] >= 0) train_ds = quality_dataset["train"].remove_columns(["pid", "input", "output"]) test_ds = quality_dataset["validation"].remove_columns(["pid", "input", "output"]) ``` Specifically, only `quality` subset is kept and processed into MCQ format. The `test` split from original dataset is removed since it doesn't have ground truth labels. Instead, validation split is assigned as test. Number of examples in train: ~2.5k Number of examples in test: ~2.1k This dataset can be used to test performance of a model focusing on long contexts. Input Tokens as per [llama2](bclavie/bert24_32k_tok_llama2) tokenizer: Mean -> 7.4k, SD: 2.3k, Max -> 11.6k --- Relevant sections from the [SCROLLS: Standardized CompaRison Over Long Language Sequences paper](https://arxiv.org/pdf/2201.03533) ``` QuALITY (Pang et al., 2021): A multiplechoice question answering dataset over stories and articles sourced from Project Gutenberg,10 the Open American National Corpus (Fillmore et al., 1998; Ide and Suderman, 2004), and more. Experienced writers wrote questions and distractors, and were incentivized to write answerable, unambiguous questions such that in order to correctly answer them, human annotators must read large portions of the given document. To measure the difficulty of their questions, Pang et al. conducted a speed validation process, where another set of annotators were asked to answer questions given only a short period of time to skim through the document. As a result, 50% of the questions in QuALITY are labeled as hard, i.e. the majority of the annotators in the speed validation setting chose the wrong answer. ```