Files changed (1) hide show
  1. README.md +41 -30
README.md CHANGED
@@ -1,30 +1,41 @@
1
- ---
2
- language: en
3
- license: apache-2.0
4
- ---
5
-
6
-
7
- HF-version model for PRIMERA: Pyramid-based Masked Sentence Pre-training for Multi-document Summarization (ACL 2022).
8
-
9
- The original code can be found [here](https://github.com/allenai/PRIMER). You can find the script and notebook to train/evaluate the model in the original github repo.
10
-
11
- * Note: due to the difference between the implementations of the original Longformer and the Huggingface LED model, the results of converted models are slightly different. We run a sanity check on both fine-tuned and non fine-tuned models, and show the results below:
12
-
13
- | Model | Rouge-1 | Rouge-2 | Rouge-L |
14
- | --- | ----------- |----------- |----------- |
15
- | PRIMERA | 42.0 | 13.6 | 20.8|
16
- | PRIMERA-hf | 41.7 |13.6 | 20.5|
17
- | PRIMERA(finetuned) | 49.9 | 21.1 | 25.9|
18
- | PRIMERA-hf(finetuned) | 49.9 | 20.9 | 25.8|
19
-
20
- You can use it by
21
- ```
22
- from transformers import (
23
- AutoTokenizer,
24
- LEDConfig,
25
- LEDForConditionalGeneration,
26
- )
27
- tokenizer = AutoTokenizer.from_pretrained('allenai/PRIMERA')
28
- config=LEDConfig.from_pretrained('allenai/PRIMERA')
29
- model = LEDForConditionalGeneration.from_pretrained('allenai/PRIMERA')
30
- ```
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+
3
+ language: en
4
+
5
+ license: apache-2.0
6
+
7
+ ---
8
+
9
+
10
+ # 🏬QAmden🏬: Question-Answering-based Multi-DocumENt model
11
+
12
+ HF-version of the QAmden model fine-tuned over multinews: *Peek Across*: Improving Multi-Document Modeling via Cross-Document Question-Answering (ACL 2023).
13
+
14
+
15
+ You can use it by
16
+
17
+ ```python
18
+ from transformers import (
19
+ AutoTokenizer,
20
+ LEDConfig,
21
+ LEDForConditionalGeneration,
22
+ )
23
+ # load model and tokenizer
24
+ tokenizer = AutoTokenizer.from_pretrained('biu-nlp/QAmden')
25
+ config=LEDConfig.from_pretrained('biu-nlp/QAmden-multinews')
26
+
27
+ model = LEDForConditionalGeneration.from_pretrained('biu-nlp/QAmden-multinews')
28
+ ```
29
+
30
+ The original repo is [here](https://github.com/aviclu/peekacross).
31
+
32
+ If you find our work useful, please cite the paper as:
33
+
34
+ ```python
35
+ @article{caciularu2023peekacross,
36
+ title={Peek Across: Improving Multi-Document Modeling via Cross-Document Question-Answering},
37
+ author={Caciularu, Avi and Peters, Matthew E and Goldberger, Jacob and Dagan, Ido and Cohan, Arman},
38
+ journal={The 61st Annual Meeting of the Association for Computational Linguistics: ACL 2023},
39
+ year={2023}
40
+ }
41
+ ```