StevenTang commited on
Commit
e82063b
1 Parent(s): 96d56dc

Update README

Browse files
Files changed (1) hide show
  1. README.md +42 -0
README.md ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ tags:
6
+ - text-generation
7
+ - text2text-generation
8
+ pipeline_tag: text2text-generation
9
+ widget:
10
+ - text: "Summarize: You may want to stick it to your boss and leave your job, but don't do it if these are your reasons."
11
+ example_title: "Example1"
12
+ - text: "Summarize: Jorge Alfaro drove in two runs, Aaron Nola pitched seven innings of two-hit ball and the Philadelphia Phillies beat the Los Angeles Dodgers 2-1 Thursday, spoiling Clayton Kershaw's first start in almost a month. Hitting out of the No. 8 spot in the ..."
13
+ example_title: "Example2"
14
+ ---
15
+
16
+ # MVP-summarization
17
+ The MVP-summarization model was proposed in [**MVP: Multi-task Supervised Pre-training for Natural Language Generation**](https://github.com/RUCAIBox/MVP/blob/main/paper.pdf) by Tianyi Tang, Junyi Li, Wayne Xin Zhao and Ji-Rong Wen.
18
+
19
+ The detailed information and instructions can be found [https://github.com/RUCAIBox/MVP](https://github.com/RUCAIBox/MVP).
20
+
21
+ ## Model Description
22
+ MVP-summarization is a prompt-based model that MVP is further equipped with prompts pre-trained using labeled summarization datasets. It is a variant (MVP+S) of our main [MVP](https://huggingface.co/RUCAIBox/mvp) model. It follows a Transformer encoder-decoder architecture with layer-wise prompts.
23
+
24
+ MVP-summarization is specially designed for summarization tasks, such as new summarization (CNN/DailyMail, XSum) and dialog summarization (SAMSum).
25
+
26
+ ## Example
27
+ ```python
28
+ >>> from transformers import MvpTokenizer, MvpForConditionalGeneration
29
+
30
+ >>> tokenizer = MvpTokenizer.from_pretrained("RUCAIBox/mvp")
31
+ >>> model = MvpForConditionalGeneration.from_pretrained("RUCAIBox/mvp-summarization")
32
+
33
+ >>> inputs = tokenizer(
34
+ ... "Summarize: You may want to stick it to your boss and leave your job, but don't do it if these are your reasons.",
35
+ ... return_tensors="pt",
36
+ ... )
37
+ >>> generated_ids = model.generate(**inputs)
38
+ >>> tokenizer.batch_decode(generated_ids, skip_special_tokens=True)
39
+ ["Don't do it if these are your reasons"]
40
+ ```
41
+
42
+ ## Citation