xyma commited on
Commit
1da98c8
1 Parent(s): b457daf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +42 -0
README.md CHANGED
@@ -1,3 +1,45 @@
1
  ---
 
 
 
 
2
  license: apache-2.0
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language: en
3
+ tags:
4
+ - PROP
5
+ - Pretrain4IR
6
  license: apache-2.0
7
+ datasets:
8
+ - msmarco
9
  ---
10
+
11
+
12
+ # PROP-wiki
13
+
14
+ **PROP**, **P**re-training with **R**epresentative w**O**rds **P**rediction, is a new pre-training method tailored for ad-hoc retrieval. PROP is inspired by the classical statistical language model for IR, specifically the query likelihood model, which assumes that the query is generated as the piece of text representative of the “ideal” document. Based on this idea, we construct the representative words prediction (ROP) task for pre-training. The full paper can be found [here](https://arxiv.org/pdf/2010.10137.pdf).
15
+
16
+
17
+ # Citation
18
+ If you find our work useful, please consider citing our paper:
19
+ ```bibtex
20
+ @inproceedings{DBLP:conf/wsdm/MaGZFJC21,
21
+ author = {Xinyu Ma and
22
+ Jiafeng Guo and
23
+ Ruqing Zhang and
24
+ Yixing Fan and
25
+ Xiang Ji and
26
+ Xueqi Cheng},
27
+ editor = {Liane Lewin{-}Eytan and
28
+ David Carmel and
29
+ Elad Yom{-}Tov and
30
+ Eugene Agichtein and
31
+ Evgeniy Gabrilovich},
32
+ title = {{PROP:} Pre-training with Representative Words Prediction for Ad-hoc
33
+ Retrieval},
34
+ booktitle = {{WSDM} '21, The Fourteenth {ACM} International Conference on Web Search
35
+ and Data Mining, Virtual Event, Israel, March 8-12, 2021},
36
+ pages = {283--291},
37
+ publisher = {{ACM}},
38
+ year = {2021},
39
+ url = {https://doi.org/10.1145/3437963.3441777},
40
+ doi = {10.1145/3437963.3441777},
41
+ timestamp = {Wed, 07 Apr 2021 16:17:44 +0200},
42
+ biburl = {https://dblp.org/rec/conf/wsdm/MaGZFJC21.bib},
43
+ bibsource = {dblp computer science bibliography, https://dblp.org}
44
+ }
45
+ ```