xiaowu0162 commited on
Commit
a26bfef
1 Parent(s): 44d98b5

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +24 -0
README.md ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Paper: [Pre-trained Language Models for Keyphrase Generation: A Thorough Empirical Study](https://arxiv.org/abs/2212.10233)
2
+
3
+ ```
4
+ @article{https://doi.org/10.48550/arxiv.2212.10233,
5
+ doi = {10.48550/ARXIV.2212.10233},
6
+ url = {https://arxiv.org/abs/2212.10233},
7
+ author = {Wu, Di and Ahmad, Wasi Uddin and Chang, Kai-Wei},
8
+ keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
9
+ title = {Pre-trained Language Models for Keyphrase Generation: A Thorough Empirical Study},
10
+ publisher = {arXiv},
11
+ year = {2022},
12
+ copyright = {Creative Commons Attribution 4.0 International}
13
+ }
14
+ ```
15
+
16
+ Pre-training Corpus: [RealNews](https://github.com/rowanz/grover/tree/master/realnews)
17
+
18
+ Pre-training Details:
19
+ - Resume from bert-base-uncased
20
+ - Batch size: 512
21
+ - Total steps: 250k
22
+ - Learning rate: 1e-4
23
+ - LR schedule: linear with 4k warmup steps
24
+ - Masking ratio: 15% dynamic masking