Update README.md
Browse files
README.md
CHANGED
@@ -4,13 +4,34 @@ language:
|
|
4 |
license: apache-2.0
|
5 |
---
|
6 |
|
7 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8 |
This is a Chinese transformer-xl model trained on [Wudao dataset](https://resource.wudaoai.cn/home?ind&name=WuDaoCorpora%202.0&id=1394901288847716352)
|
9 |
and finetuned on a denoise dataset constructed by our team. The denoise task is to reconstruct a fluent and clean text from a noisy input which includes random insertion/swap/deletion/replacement/sentence reordering.
|
10 |
|
11 |
-
## Usage
|
|
|
|
|
12 |
|
13 |
-
### load model
|
14 |
```python
|
15 |
from fengshen.models.transfo_xl_denoise.tokenization_transfo_xl_denoise import TransfoXLDenoiseTokenizer
|
16 |
from fengshen.models.transfo_xl_denoise.modeling_transfo_xl_denoise import TransfoXLDenoiseModel
|
@@ -19,7 +40,8 @@ tokenizer = TransfoXLDenoiseTokenizer.from_pretrained('IDEA-CCNL/Bigan-Transform
|
|
19 |
model = TransfoXLDenoiseModel.from_pretrained('IDEA-CCNL/Bigan-Transformer-XL-denoise-1.1B')
|
20 |
```
|
21 |
|
22 |
-
###
|
|
|
23 |
```python
|
24 |
from fengshen.models.transfo_xl_denoise.generate import denoise_generate
|
25 |
input_text = "凡是有成就的人, 都很严肃地对待生命自己的"
|
@@ -27,13 +49,31 @@ res = denoise_generate(model, tokenizer, input_text)
|
|
27 |
print(res) # "有成就的人都很严肃地对待自己的生命。"
|
28 |
```
|
29 |
|
30 |
-
## Citation
|
31 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
32 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
@misc{Fengshenbang-LM,
|
34 |
title={Fengshenbang-LM},
|
35 |
author={IDEA-CCNL},
|
36 |
-
year={
|
37 |
howpublished={\url{https://github.com/IDEA-CCNL/Fengshenbang-LM}},
|
38 |
}
|
39 |
-
```
|
|
|
4 |
license: apache-2.0
|
5 |
---
|
6 |
|
7 |
+
# Bigan-Transformer-XL-denoise-1.1B
|
8 |
+
|
9 |
+
- Github: [Fengshenbang-LM](https://github.com/IDEA-CCNL/Fengshenbang-LM)
|
10 |
+
- Docs: [Fengshenbang-Docs](https://fengshenbang-doc.readthedocs.io/)
|
11 |
+
|
12 |
+
## 简介 Brief Introduction
|
13 |
+
|
14 |
+
以去噪任务为预训练目标的中文Transformer-XL。
|
15 |
+
|
16 |
+
Chinese Transformer-XL with a denoising task as a pre-training target.
|
17 |
+
|
18 |
+
## 模型分类 Model Taxonomy
|
19 |
+
|
20 |
+
| 需求 Demand | 任务 Task | 系列 Series | 模型 Model | 参数 Parameter | 额外 Extra |
|
21 |
+
| :----: | :----: | :----: | :----: | :----: | :----: |
|
22 |
+
| 特殊 Special | 探索 Exploration | 比干 Bigan | Transformer | 1.1B | denoise |
|
23 |
+
|
24 |
+
## 模型信息 Model Information
|
25 |
+
|
26 |
+
|
27 |
+
|
28 |
This is a Chinese transformer-xl model trained on [Wudao dataset](https://resource.wudaoai.cn/home?ind&name=WuDaoCorpora%202.0&id=1394901288847716352)
|
29 |
and finetuned on a denoise dataset constructed by our team. The denoise task is to reconstruct a fluent and clean text from a noisy input which includes random insertion/swap/deletion/replacement/sentence reordering.
|
30 |
|
31 |
+
## 使用 Usage
|
32 |
+
|
33 |
+
### 加载模型 Loading Models
|
34 |
|
|
|
35 |
```python
|
36 |
from fengshen.models.transfo_xl_denoise.tokenization_transfo_xl_denoise import TransfoXLDenoiseTokenizer
|
37 |
from fengshen.models.transfo_xl_denoise.modeling_transfo_xl_denoise import TransfoXLDenoiseModel
|
|
|
40 |
model = TransfoXLDenoiseModel.from_pretrained('IDEA-CCNL/Bigan-Transformer-XL-denoise-1.1B')
|
41 |
```
|
42 |
|
43 |
+
### 使用示例 Usage Examples
|
44 |
+
|
45 |
```python
|
46 |
from fengshen.models.transfo_xl_denoise.generate import denoise_generate
|
47 |
input_text = "凡是有成就的人, 都很严肃地对待生命自己的"
|
|
|
49 |
print(res) # "有成就的人都很严肃地对待自己的生命。"
|
50 |
```
|
51 |
|
52 |
+
## 引用 Citation
|
53 |
+
|
54 |
+
如果您在您的工作中使用了我们的模型,可以引用我们的[论文](https://arxiv.org/abs/2209.02970):
|
55 |
+
|
56 |
+
If you are using the resource for your work, please cite the our [paper](https://arxiv.org/abs/2209.02970):
|
57 |
+
|
58 |
+
```text
|
59 |
+
@article{fengshenbang,
|
60 |
+
author = {Junjie Wang and Yuxiang Zhang and Lin Zhang and Ping Yang and Xinyu Gao and Ziwei Wu and Xiaoqun Dong and Junqing He and Jianheng Zhuo and Qi Yang and Yongfeng Huang and Xiayu Li and Yanghan Wu and Junyu Lu and Xinyu Zhu and Weifeng Chen and Ting Han and Kunhao Pan and Rui Wang and Hao Wang and Xiaojun Wu and Zhongshen Zeng and Chongpei Chen and Ruyi Gan and Jiaxing Zhang},
|
61 |
+
title = {Fengshenbang 1.0: Being the Foundation of Chinese Cognitive Intelligence},
|
62 |
+
journal = {CoRR},
|
63 |
+
volume = {abs/2209.02970},
|
64 |
+
year = {2022}
|
65 |
+
}
|
66 |
```
|
67 |
+
|
68 |
+
也可以引用我们的[网站](https://github.com/IDEA-CCNL/Fengshenbang-LM/):
|
69 |
+
|
70 |
+
You can also cite our [website](https://github.com/IDEA-CCNL/Fengshenbang-LM/):
|
71 |
+
|
72 |
+
```text
|
73 |
@misc{Fengshenbang-LM,
|
74 |
title={Fengshenbang-LM},
|
75 |
author={IDEA-CCNL},
|
76 |
+
year={2021},
|
77 |
howpublished={\url{https://github.com/IDEA-CCNL/Fengshenbang-LM}},
|
78 |
}
|
79 |
+
```
|