dongxiaoqun commited on
Commit
e142f2c
1 Parent(s): daaf4f3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -13
README.md CHANGED
@@ -5,19 +5,6 @@ tags:
5
  inference: True
6
  ---
7
 
8
-
9
- Randeng_Pegasus_523M_Summary model (Chinese),which codes has merged into [Fengshenbang-LM](https://github.com/IDEA-CCNL/Fengshenbang-LM)
10
-
11
- The 523M million parameter randeng_pegasus_large model, training with sampled gap sentence ratios on 180G Chinese data, and stochastically sample important sentences. The pretraining task just same as the paper [PEGASUS: Pre-training with Extracted Gap-sentences for
12
- Abstractive Summarization](https://arxiv.org/pdf/1912.08777.pdf) mentioned.
13
-
14
- Different from the English version of pegasus, considering that the Chinese sentence piece is unstable, we use jieba and Bertokenizer as the tokenizer in chinese pegasus model.
15
-
16
- This model we provided in hugging face hub is only the pretrained model, has not finetuned with download data yet.
17
-
18
- We also pretained a base model, available with [Randeng_Pegasus_238M_Summary](https://huggingface.co/IDEA-CCNL/Randeng_Pegasus_238M_Summary)
19
-
20
-
21
  Task: Summarization
22
 
23
  ## Usage
 
5
  inference: True
6
  ---
7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  Task: Summarization
9
 
10
  ## Usage