gabrielaltay commited on
Commit
e1ba743
1 Parent(s): 545e75f

upload hub_repos/medhop/README.md to hub from bigbio repo

Browse files
Files changed (1) hide show
  1. README.md +65 -0
README.md ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ language:
4
+ - en
5
+ license: cc-by-sa-3.0
6
+ license_bigbio_shortname: CC_BY_SA_3p0
7
+ pretty_name: MedHop
8
+ ---
9
+
10
+
11
+ # Dataset Card for MedHop
12
+
13
+ ## Dataset Description
14
+
15
+ - **Homepage:** http://qangaroo.cs.ucl.ac.uk/
16
+ - **Pubmed:** True
17
+ - **Public:** True
18
+ - **Tasks:** Question Answering
19
+
20
+
21
+ With the same format as WikiHop, this dataset is based on research paper
22
+ abstracts from PubMed, and the queries are about interactions between
23
+ pairs of drugs. The correct answer has to be inferred by combining
24
+ information from a chain of reactions of drugs and proteins.
25
+
26
+
27
+
28
+ ## Citation Information
29
+
30
+ ```
31
+ @article{welbl-etal-2018-constructing,
32
+ title = Constructing Datasets for Multi-hop Reading Comprehension Across Documents,
33
+ author = Welbl, Johannes and Stenetorp, Pontus and Riedel, Sebastian,
34
+ journal = Transactions of the Association for Computational Linguistics,
35
+ volume = 6,
36
+ year = 2018,
37
+ address = Cambridge, MA,
38
+ publisher = MIT Press,
39
+ url = https://aclanthology.org/Q18-1021,
40
+ doi = 10.1162/tacl_a_00021,
41
+ pages = 287--302,
42
+ abstract = {
43
+ Most Reading Comprehension methods limit themselves to queries which
44
+ can be answered using a single sentence, paragraph, or document.
45
+ Enabling models to combine disjoint pieces of textual evidence would
46
+ extend the scope of machine comprehension methods, but currently no
47
+ resources exist to train and test this capability. We propose a novel
48
+ task to encourage the development of models for text understanding
49
+ across multiple documents and to investigate the limits of existing
50
+ methods. In our task, a model learns to seek and combine evidence
51
+ -- effectively performing multihop, alias multi-step, inference.
52
+ We devise a methodology to produce datasets for this task, given a
53
+ collection of query-answer pairs and thematically linked documents.
54
+ Two datasets from different domains are induced, and we identify
55
+ potential pitfalls and devise circumvention strategies. We evaluate
56
+ two previously proposed competitive models and find that one can
57
+ integrate information across documents. However, both models
58
+ struggle to select relevant information; and providing documents
59
+ guaranteed to be relevant greatly improves their performance. While
60
+ the models outperform several strong baselines, their best accuracy
61
+ reaches 54.5 % on an annotated test set, compared to human
62
+ performance at 85.0 %, leaving ample room for improvement.
63
+ }
64
+
65
+ ```