Datasets:
The viewer is disabled because this dataset repo requires arbitrary Python code execution. Please consider
removing the
loading script
and relying on
automated data support
(you can use
convert_to_parquet
from the datasets
library). If this is not possible, please
open a discussion
for direct help.
Dataset Card for Curiosity Dataset
Dataset Summary
Curiosity dataset consists of 14K English dialogs (181K utterances) where users and assistants converse about geographic topics like geopolitical entities and locations. This dataset is annotated with pre-existing user knowledge, message-level dialog acts, grounding to Wikipedia, and user reactions to messages.
Supported Tasks and Leaderboards
text-generation-other-conversational-curiosity
: The dataset can be used to train a model for Conversational Curiosity, which consists in the testing of the hypothesis that engagement increases when users are presented with facts related to what they know. Success on this task is typically measured by achieving a high Accuracy and F1 Score.
Languages
The text in the dataset is in English collected by crowd-souring. The associated BCP-47 code is en
.
Dataset Structure
Data Instances
A typical data point consists of dialogs between an user and an assistant, which is followed by the different attributes of the particular dialog.
An example from the Curiosity Dataset train set looks as follows:
{'annotated': 1,
'aspects': ['Media', 'Politics and government'],
'assistant_dialog_rating': 5,
'assistant_id': 341,
'assistant_other_agent_rating': 5,
'created_time': 1571783665,
'dialog_id': 21922,
'first_aspect': 'Media',
'focus_entity': 'Namibia',
'inferred_steps': 1,
'is_annotated': 0,
'known_entities': ['South Africa', 'United Kingdom', 'Portugal'],
'messages': {'dialog_acts': [['request_topic'],
['inform_response'],
['request_aspect'],
['inform_response'],
['request_followup'],
['inform_response'],
['request_aspect', 'feedback_positive'],
['inform_response'],
['request_followup'],
['inform_response'],
[],
[]],
'facts': [{'fid': [], 'source': [], 'used': []},
{'fid': [77870, 77676, 77816, 77814, 77775, 77659, 77877, 77785, 77867],
'source': [0, 1, 2, 2, 0, 2, 0, 1, 1],
'used': [0, 0, 0, 0, 0, 0, 0, 0, 0]},
{'fid': [], 'source': [], 'used': []},
{'fid': [77725, 77870, 77676, 77863, 77814, 77775, 77659, 77877, 77867],
'source': [2, 0, 1, 1, 2, 0, 2, 0, 1],
'used': [0, 0, 0, 0, 0, 0, 0, 0, 0]},
{'fid': [], 'source': [], 'used': []},
{'fid': [77694, 77661, 77863, 77780, 77671, 77704, 77869, 77693, 77877],
'source': [1, 2, 1, 0, 2, 2, 0, 1, 0],
'used': [0, 0, 0, 0, 0, 0, 0, 0, 1]},
{'fid': [], 'source': [], 'used': []},
{'fid': [77816, 77814, 77864, 77659, 77877, 77803, 77738, 77784, 77789],
'source': [2, 2, 0, 2, 0, 1, 1, 0, 1],
'used': [0, 0, 0, 0, 0, 0, 0, 0, 0]},
{'fid': [], 'source': [], 'used': []},
{'fid': [77694, 77776, 77780, 77696, 77707, 77693, 77778, 77702, 77743],
'source': [1, 0, 0, 2, 1, 1, 0, 2, 2],
'used': [0, 0, 0, 0, 0, 0, 0, 0, 0]},
{'fid': [], 'source': [], 'used': []},
{'fid': [77662, 77779, 77742, 77734, 77663, 77777, 77702, 77731, 77778],
'source': [1, 0, 2, 1, 2, 0, 2, 1, 0],
'used': [0, 0, 0, 0, 0, 0, 0, 0, 1]}],
'liked': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0],
'message': ['Hi. I want information about Namibia.',
'Nmbia is a country in southern Africa.',
'Do you have information about the media there?',
'A mentional amount of foriegn',
'What about it?',
"Media and journalists in Namibia are represented by the Namibia chapter of the Media Institute of 'southern Africa and the Editors Forum of Namibia.",
'Interesting! What can you tell me about the politics and government?',
'Namibia formed the Namibian Defence Force, comprising former enemies in a 23-year bush war.',
'Do you have more information about it?',
"With a small army and a fragile economy , the Namibian government's principal foreign policy concern is developing strengthened ties within the Southern African region.",
"That's all I wanted to know. Thank you!",
'My pleasure!'],
'message_id': ['617343895',
'2842515356',
'4240816985',
'520711081',
'1292358002',
'3677078227',
'1563061125',
'1089028270',
'1607063839',
'113037558',
'1197873991',
'1399017322'],
'sender': [0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1]},
'related_entities': ['Western Roman Empire',
'United Kingdom',
'Portuguese language',
'Southern African Development Community',
'South Africa',
'Kalahari Desert',
'Namib Desert',
'League of Nations',
'Afrikaans',
'Sub-Saharan Africa',
'Portugal',
'South-West Africa',
'Warmbad, Namibia',
'German language',
'NBC'],
'reported': 0,
'second_aspect': 'Politics and government',
'shuffle_facts': 1,
'tag': 'round_2',
'user_dialog_rating': 5,
'user_id': 207,
'user_other_agent_rating': 5}
Data Fields
messages
: List of dialogs between the user and the assistant and their associated attributesdialog_acts
: List of actions performed in the dialogsfacts
: List of facts returned by the assistantfid
: Fact IDsource
: Source for the factused
: Whether facts were used before in the same dialog
liked
: List of values indicating whether each dialog was likedmessage
: List of dialogs (messages) between the user and the assistantmessage_id
: Message IDsender
: Message author ID (numeric)
known_entities
: Rooted facts about entities the user knowsfocus_entity
: Entity in focus in the dialogsdialog_id
: Dialog IDinferred_steps
: Number of inferred stepscreated_time
: Time of creation of the dialogaspects
: List of two aspects which the dialog is aboutfirst_aspect
: First aspectsecond_aspect
: Second aspectshuffle_facts
: Whether facts were shuffledrelated_entities
: List of fifteen related entities to the focus entitytag
: Conversation taguser_id
: User IDassistant_id
: Assistant IDis_annotated
: 0 or 1 (More Information Needed)user_dialog_rating
: 1 - 5 (More Information Needed)user_other_agent_rating
: 1 - 5 (More Information Needed)assistant_dialog_rating
: 1 - 5 (More Information Needed)assistant_other_agent_rating
: 1 - 5 (More Information Needed)reported
: Whether the dialog was reported inappropriateannotated
: 0 or 1 (More Information Needed)
Data Splits
The data is split into a training, validation, test and test_zero set as per the original dataset split.
train | validation | test | test_zero | |
---|---|---|---|---|
Input dialog examples | 10287 | 1287 | 1287 | 1187 |
Dataset Creation
Curation Rationale
[More Information Needed]
Source Data
Initial Data Collection and Normalization
[More Information Needed]
Who are the source language producers?
[More Information Needed]
Annotations
Annotation process
[More Information Needed]
Who are the annotators?
[More Information Needed]
Personal and Sensitive Information
[More Information Needed]
Considerations for Using the Data
Social Impact of Dataset
[More Information Needed]
Discussion of Biases
[More Information Needed]
Other Known Limitations
[More Information Needed]
Additional Information
Dataset Curators
[More Information Needed]
Licensing Information
Attribution-NonCommercial 4.0 International
Citation Information
@inproceedings{rodriguez2020curiosity,
title = {Information Seeking in the Spirit of Learning: a Dataset for Conversational Curiosity},
author = {Pedro Rodriguez and Paul Crook and Seungwhan Moon and Zhiguang Wang},
year = 2020,
booktitle = {Empirical Methods in Natural Language Processing}
}
Contributions
Thanks to @vineeths96 for adding this dataset.
- Downloads last month
- 321