""" Copyright (c) 2023, salesforce.com, inc. All rights reserved. SPDX-License-Identifier: Apache License 2.0 For full license text, see the LICENSE file in the repo root or https://www.apache.org/licenses/LICENSE-2.0 """ #!/usr/bin/env python3 import os import json import datasets from datasets import (GeneratorBasedBuilder, BuilderConfig, SplitGenerator, DatasetInfo, Features, Sequence, Value, Version) logger = datasets.logging.get_logger(__name__) datasets.logging.disable_progress_bar() _VERSION = Version("1.0.0") _CITATION = """ @misc{zhang2023dialogstudio, title={DialogStudio: Towards Richest and Most Diverse Unified Dataset Collection for Conversational AI}, author={Jianguo Zhang and Kun Qian and Zhiwei Liu and Shelby Heinecke and Rui Meng and Ye Liu and Zhou Yu and and Huan Wang and Silvio Savarese and Caiming Xiong}, year={2023}, eprint={2307.10172}, archivePrefix={arXiv}, primaryClass={cs.CL} """ DATASETS = { # not pass: "CANARD" "open_domain": [ "chitchat-dataset", "ConvAI2", "AntiScam", "Empathetic", "HH-RLHF", "PLACES3.5", "Prosocial", "SODA" ], "knowledge_grounded": [ "CompWebQ", "CoQA", "CoSQL", "DART", "FeTaQA", "GrailQA", "HybridQA", "MTOP", "MultiModalQA", "SParC", "Spider", "SQA", "ToTTo", "WebQSP", "WikiSQL", "WikiTQ", "wizard_of_internet", "wizard_of_wikipedia" ], "dialogue_summarization": [ "AMI", "CRD3", "DialogSum", "ECTSum", "ICSI", "MediaSum", "QMSum", "SAMSum", "TweetSumm", "ConvoSumm", "SummScreen_ForeverDreaming", "SummScreen_TVMegaSite" ], "natural_language_understanding": [ "ATIS", "ATIS-NER", "BANKING77", "BANKING77-OOS", "CLINC-Single-Domain-OOS-banking", "CLINC-Single-Domain-OOS-credit_cards", "CLINC150", "DSTC8-SGD", "HWU64", "MIT-Movie", "MIT-Restaurant", "RESTAURANTS8K", "SNIPS", "SNIPS-NER", "TOP", "TOP-NER" ], "task_oriented": [ "ABCD", "AirDialogue", "BiTOD", "CaSiNo", "CraigslistBargains", "Disambiguation", "DSTC2-Clean", "FRAMES", "GECOR", "HDSA-Dialog", "KETOD", "KVRET", "MetaLWOZ", "MS-DC", "MuDoCo", "MulDoGO", "MultiWOZ_2.1", "MULTIWOZ2_2", "SGD", "SimJointGEN", "SimJointMovie", "SimJointRestaurant", "STAR", "Taskmaster1", "Taskmaster2", "Taskmaster3", "WOZ2_0" ], "conversational_recommendation": [ "Redial", "DuRecDial-2.0", "OpenDialKG", "SalesBot", ] } _URL = "https://huggingface.co/datasets/Salesforce/dialogstudio/tree/main/" class DialogStudioConfig(BuilderConfig): """BuilderConfig for DialogStudio.""" def __init__(self, extra_features, category, data_name, citation, url, **kwargs): """BuilderConfig for DialogStudio. Args: extra_features: `list[string]`, list of the features that will appear in the feature dict. Should not include "label". data_url: `string`, url to download the zip file from. citation: `string`, citation for the data set. url: `string`, url for information about the data set. label_classes: `list[string]`, the list of classes for the label if the label is present as a string. Non-string labels will be cast to either 'False' or 'True'. **kwargs: keyword arguments forwarded to super. """ super(DialogStudioConfig, self).__init__(version=_VERSION, **kwargs) self.extra_features = extra_features self.category = category self.data_name = data_name self.compressed_file = f"{data_name}.zip" self.citation = citation self.url = url class DialogStudio(GeneratorBasedBuilder): """DialogStudio""" BUILDER_CONFIGS = [] for category, dataset_list in DATASETS.items(): if category in ["task_oriented", "conversational_recommendation"]: extra_features = { "dialog":[ "external knowledge non-flat", "external knowledge", "dst knowledge", "intent knowledge", ], "turn":[ "dst", "dst accumulated", "intent", "external knowledge", "external knowledge non-flat" ] } elif category in ["natural_language_understanding"]: extra_features = {"dialog":[], "turn":["external knowledge", "external knowledge non-flat"]} else: extra_features = {"dialog":[], "turn":[]} for data_name in dataset_list: BUILDER_CONFIGS.append( DialogStudioConfig( name=data_name, data_name=data_name, description="", extra_features=extra_features, category=category, citation=_CITATION, url="https://github.com/salesforce/DialogStudio/tree/main", )) DEFAULT_CONFIG_NAME = "WOZ2_0" def _info(self): features = { "original dialog id": Value("string"), "new dialog id": Value("string"), "dialog index": Value("int32"), "original dialog info": Value("string"), "log": [ { "turn id": Value("int32"), "user utterance": Value("string"), "system response": Value("string"), "dialog history": Value("string"), "original user side information": Value("string"), "original system side information": Value("string"), } ], "prompt": [Value("string")] } if self.config.extra_features["dialog"]: for name in self.config.extra_features["dialog"]: features[name] = Value("string") if self.config.extra_features["turn"]: log_config = { "turn id": Value("int32"), "user utterance": Value("string"), "system response": Value("string"), "dialog history": Value("string"), "original user side information": Value("string"), "original system side information": Value("string"), } for name in self.config.extra_features["turn"]: log_config[name] = Value("string") features["log"] = [log_config] return DatasetInfo( description="", features=Features(features), homepage=self.config.url, citation=_CITATION, ) def _split_generators(self, dl_manager): """ This script assume the datset is not stored in zip file Instead, data is stored in format: . |-task-oriented |-WOZ2.0 |-train |-dialogues_1.json |-dialogues_2.json |-... this script would download the json file one-by-one """ splits = [] file_path_list = {"train":[], "val":[], "test":[]} print("❤️Attention❤️: Dataset download may take some time. We appreciate your patience!") for mode in ["train", "val", "test"]: for file_idx in range(1, 1000000): file_to_download = os.path.join(self.config.category, self.config.data_name, mode, f"dialogues_{file_idx}.json") try: dl_path = dl_manager.download(file_to_download) except: break file_path_list[mode].append(dl_path) if file_path_list["train"]: if any(split.name == datasets.Split.TRAIN for split in splits): raise ValueError("Split 'train' already added.") splits.append( SplitGenerator( name=datasets.Split.TRAIN, gen_kwargs={ "file_path_list": file_path_list["train"], "split": datasets.Split.TRAIN, }, ) ) if file_path_list["val"]: if any(split.name == datasets.Split.VALIDATION for split in splits): raise ValueError("Split 'validation' already added.") splits.append( SplitGenerator( name=datasets.Split.VALIDATION, gen_kwargs={ "file_path_list": file_path_list["val"], "split": datasets.Split.VALIDATION, }, ) ) if file_path_list["test"]: if any(split.name == datasets.Split.TEST for split in splits): raise ValueError("Split 'test' already added.") splits.append( SplitGenerator( name=datasets.Split.TEST, gen_kwargs={ "file_path_list": file_path_list["test"], "split": datasets.Split.TEST, }, ) ) return splits def _load_json(self, file_path): with open(file_path, encoding="utf-8") as f: data = json.loads(f.read()) return data def _generate_examples(self, file_path_list, split): """This function returns the examples in the raw (text) form.""" data = {} for filepath in file_path_list: data.update(self._load_json(filepath)) logger.info(f"generating {len(data)} examples from = {split}") for dial_id, dial in data.items(): if type(dial["log"]) == dict: dial["log"] = [dial["log"]] example = { "original dialog id": dial["original dialog id"], "new dialog id": dial_id, "dialog index": dial["dialog index"], "original dialog info": json.dumps(dial["original dialog info"]), "log": [{ "turn id": turn["turn id"], "user utterance": turn["user utterance"], "system response": turn["system response"], "dialog history": turn["dialog history"], "original user side information": json.dumps(turn["original user side information"]), "original system side information": json.dumps(turn["original system side information"]), } for turn in dial["log"]], "prompt": dial["prompt"] if "prompt" in dial and dial["prompt"] else [""] } if self.config.extra_features["dialog"]: for name in self.config.extra_features["dialog"]: example[name] = json.dumps(dial[name]) if name in dial else "" if self.config.extra_features["turn"]: for idx, turn in enumerate(example["log"]): for name in self.config.extra_features["turn"]: example["log"][idx][name] = json.dumps(dial["log"][idx][name]) if name in dial["log"][idx] else "" yield dial["dialog index"], example