dataset_info:
features:
- name: id
dtype: string
- name: answer
dtype: string
- name: type
dtype: string
- name: level
dtype: string
- name: supporting_facts
sequence:
- name: title
dtype: string
- name: sent_id
dtype: int32
- name: context
sequence:
- name: title
dtype: string
- name: sentences
sequence: string
- name: problem
dtype: string
- name: source
dtype: string
- name: gold_removed
dtype: int64
- name: removed_titles
sequence: string
splits:
- name: train
num_bytes: 226320070.31742346
num_examples: 20000
- name: test
num_bytes: 5716501.080351114
num_examples: 500
download_size: 116725004
dataset_size: 232036571.39777458
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
This dataset is a modified version of the HotPotQA distractor dataset, which contains factual questions requiring multi-hop reasoning. In the original HotPotQA dataset, each example presents ten paragraphs, only two of which contain the information necessary to answer the question; the remaining eight paragraphs include closely related but irrelevant details. Consequently, solving this task requires the model to identify and reason over the pertinent passages. To more strongly develop uncertainty reasoning capability, we construct HotPotQA-Modified, in which we systematically remove either 0, 1, or both of the key paragraphs required to answer each question. This modification introduces varying levels of informational completeness that the model must reason over. The "gold_removed" column indicates the number of relevant paragraphs removed (0,1 or 2). Questions are distributed across three equal groups: one-third have no relevant paragraphs (0/8), one-third have 1 relevant paragraph (1/7), and one-third have both relevant paragraphs (2/6). Each question consistently contains 8 total paragraphs.
To cite the original HotPotQA dataset:
@article{yang2018hotpotqa,
title={HotpotQA: A dataset for diverse, explainable multi-hop question answering},
author={Yang, Zhilin and Qi, Peng and Zhang, Saizheng and Bengio, Yoshua and Cohen, William W and Salakhutdinov, Ruslan and Manning, Christopher D},
journal={arXiv preprint arXiv:1809.09600},
year={2018}
}
To cite our modified dataset:
@article{damani2025beyond,
title={Beyond Binary Rewards: Training LMs to Reason About Their Uncertainty},
author={Damani, Mehul and Puri, Isha and Slocum, Stewart and Shenfeld, Idan and Choshen, Leshem and Kim, Yoon and Andreas, Jacob},
journal={arXiv preprint arXiv:2507.16806},
year={2025}
}