## Overview Original dataset available [here](https://wellecks.github.io/dialogue_nli/). ## Dataset curation Original `label` column is renamed `original_label`. The original classes are renamed as follows ``` {"positive": "entailment", "negative": "contradiction", "neutral": "neutral"}) ``` and encoded with the following mapping ``` {"entailment": 0, "neutral": 1, "contradiction": 2} ``` and stored in the newly created column `label`. The following splits and the corresponding columns are present in the original files ``` train {'dtype', 'id', 'sentence1', 'sentence2', 'original_label', 'label', 'triple2', 'triple1'} dev {'dtype', 'id', 'sentence1', 'sentence2', 'original_label', 'label', 'triple2', 'triple1'} test {'dtype', 'id', 'sentence1', 'sentence2', 'original_label', 'label', 'triple2', 'triple1'} verified_test {'dtype', 'annotation3', 'sentence1', 'sentence2', 'annotation1', 'annotation2', 'original_label', 'label', 'triple2', 'triple1'} extra_test {'dtype', 'id', 'sentence1', 'sentence2', 'original_label', 'label', 'triple2', 'triple1'} extra_dev {'dtype', 'id', 'sentence1', 'sentence2', 'original_label', 'label', 'triple2', 'triple1'} extra_train {'dtype', 'id', 'sentence1', 'sentence2', 'original_label', 'label', 'triple2', 'triple1'} valid_havenot {'dtype', 'id', 'sentence1', 'sentence2', 'original_label', 'label', 'triple2', 'triple1'} valid_attributes {'dtype', 'id', 'sentence1', 'sentence2', 'original_label', 'label', 'triple2', 'triple1'} valid_likedislike {'dtype', 'id', 'sentence1', 'sentence2', 'original_label', 'label', 'triple2', 'triple1'} ``` Note that I only keep the common columns, which means that I drop "annotation{1, 2, 3}" from `verified_test`. Note that there are some splits with the same instances, as found by matching on "original_label", "sentence1", "sentence2". ## Code to create dataset ```python import pandas as pd from pathlib import Path import json from datasets import Features, Value, ClassLabel, Dataset, DatasetDict, Sequence # load data ds = {} for path in Path(".").rglob("/*.jsonl"): print(path, flush=True) with path.open("r") as fl: data = fl.read() try: d = json.loads(data, encoding="utf-8") except json.JSONDecodeError as error: print(error) df = pd.DataFrame(d) # encode labels df["original_label"] = df["label"] df["label"] = df["label"].map({"positive": "entailment", "negative": "contradiction", "neutral": "neutral"}) df["label"] = df["label"].map({"entailment": 0, "neutral": 1, "contradiction": 2}) ds[path.name.split(".")[0]] = df # prettify names of data splits datasets = { k.replace("dialogue_nli_", "").replace("uu_", "").lower(): v for k, v in ds.items() } datasets.keys() #> dict_keys(['train', 'dev', 'test', 'verified_test', 'extra_test', 'extra_dev', 'extra_train', 'valid_havenot', 'valid_attributes', 'valid_likedislike']) # cast to datasets using only common columns features = Features({ "label": ClassLabel(num_classes=3, names=["entailment", "neutral", "contradiction"]), "sentence1": Value(dtype="string", id=None), "sentence2": Value(dtype="string", id=None), "triple1": Sequence(feature=Value(dtype="string", id=None), length=3), "triple2": Sequence(feature=Value(dtype="string", id=None), length=3), "dtype": Value(dtype="string", id=None), "id": Value(dtype="string", id=None), "original_label": Value(dtype="string", id=None), }) ds = {} for name, df in datasets.items(): if "id" not in df.columns: df["id"] = "" ds[name] = Dataset.from_pandas(df.loc[:, list(features.keys())], features=features) ds = DatasetDict(ds) ds.push_to_hub("dialogue_nli", token="") # check overlap between splits from itertools import combinations for i, j in combinations(ds.keys(), 2): print( f"{i} - {j}: ", pd.merge( ds[i].to_pandas(), ds[j].to_pandas(), on=["original_label", "sentence1", "sentence2"], how="inner", ).shape[0], ) #> train - dev: 58 #> train - test: 98 #> train - verified_test: 90 #> train - extra_test: 0 #> train - extra_dev: 0 #> train - extra_train: 0 #> train - valid_havenot: 0 #> train - valid_attributes: 0 #> train - valid_likedislike: 0 #> dev - test: 19 #> dev - verified_test: 19 #> dev - extra_test: 0 #> dev - extra_dev: 75 #> dev - extra_train: 75 #> dev - valid_havenot: 75 #> dev - valid_attributes: 75 #> dev - valid_likedislike: 75 #> test - verified_test: 12524 #> test - extra_test: 34 #> test - extra_dev: 0 #> test - extra_train: 0 #> test - valid_havenot: 0 #> test - valid_attributes: 0 #> test - valid_likedislike: 0 #> verified_test - extra_test: 29 #> verified_test - extra_dev: 0 #> verified_test - extra_train: 0 #> verified_test - valid_havenot: 0 #> verified_test - valid_attributes: 0 #> verified_test - valid_likedislike: 0 #> extra_test - extra_dev: 0 #> extra_test - extra_train: 0 #> extra_test - valid_havenot: 0 #> extra_test - valid_attributes: 0 #> extra_test - valid_likedislike: 0 #> extra_dev - extra_train: 250946 #> extra_dev - valid_havenot: 250946 #> extra_dev - valid_attributes: 250946 #> extra_dev - valid_likedislike: 250946 #> extra_train - valid_havenot: 250946 #> extra_train - valid_attributes: 250946 #> extra_train - valid_likedislike: 250946 #> valid_havenot - valid_attributes: 250946 #> valid_havenot - valid_likedislike: 250946 #> valid_attributes - valid_likedislike: 250946 ```