File size: 4,202 Bytes
2ea1065
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
Microsoft Research Sequential Question Answering (SQA) Dataset

-------------------------------------------------------------------------------------------------------
Contact Persons:
	Scott Wen-tau Yih        scottyih@microsoft.com
	Mohit Iyyer              m.iyyer@gmail.com
	Ming-Wei Chang           minchang@microsoft.com
-------------------------------------------------------------------------------------------------------

The SQA dataset was created to explore the task of answering sequences of inter-related questions on HTML tables. A detailed description of the dataset, as well as some experimental studies, can be found in the following paper: 

Mohit Iyyer, Wen-tau Yih, Ming-Wei Chang.
"Answering Complicated Question Intents Expressed in Decomposed Question Sequences."
arXiv preprint arXiv:1611.01242
https://arxiv.org/abs/1611.01242

-------------------------------------------------------------------------------------------------------

Version 1.0: November 9, 2016

-------------------------------------------------------------------------------------------------------
SUMMARY

Recent work in semantic parsing for question answering has focused on long and complicated questions, many of which would seem unnatural if asked in a normal conversation between two humans. In an effort to explore a conversational QA setting, we present a more realistic task: answering sequences of simple but inter-related questions.

We created SQA by asking crowdsourced workers to decompose 2,022 questions from WikiTableQuestions (WTQ)*, which contains highly-compositional questions about tables from Wikipedia. We had three workers decompose each WTQ question, resulting in a dataset of 6,066 sequences that contain 17,553 questions in total. Each question is also associated with answers in the form of cell locations in the tables.

* Panupong Pasupat, Percy Liang. "Compositional Semantic Parsing on Semi-Structured Tables" ACL-2015.
  http://www-nlp.stanford.edu/software/sempre/wikitable/

-------------------------------------------------------------------------------------------------------
LIST OF FILES

train.tsv       -- Training question sequences
test.tsv        -- Testing question sequences
table_csv       -- All the tables used in the questions (originally from WTQ)
random-split-*  -- Five different 80-20 training/dev splits based on training.tsv.  
                   The splits follow those provided in WTQ.
eval.py         -- The evaluation script in Python
rndfake.tsv     -- A fake output file for demonstrating the usage of eval.py
license.docx    -- License
readme.txt      -- This file

-------------------------------------------------------------------------------------------------------
DATA FORMAT

train.tsv, test.tsv, random-split-*
    -- id: question sequence id (the id is consistent with those in WTQ)
	-- annotator: 0, 1, 2 (the 3 annotators who annotated the question intent)
	-- position: the position of the question in the sequence
	-- question: the question given by the annotator
	-- table_file: the associated table
	-- answer_coordinates: the table cell coordinates of the answers (0-based, where 0 is the first row after the table header)
	-- answer_text: the content of the answer cells
	Note that some text fields may contain Tab or LF characters and thus start with quotes. 
	It is recommended to use a CSV parser like the Python CSV package to process the data.
	
table_csv (from WTQ; below is the original description in the WTQ release)
    -- Comma-separated table (The first row is treated as the column header)
       The escaped characters include:
         double quote (`"` => `\"`) and backslash (`\` => `\\`).
         Newlines are represented as quoted line breaks.
		 
rndfake.tsv
    -- A fake output file for the test questions; fields are: id, annotator, position, answer_coordinates

-------------------------------------------------------------------------------------------------------
EVALUATION

$ python eval.py test.tsv rndfake.tsv
Sequence Accuracy = 14.83% (152/1025)
Answer Accuracy =   50.33% (1516/3012)

-------------------------------------------------------------------------------------------------------