GPT2-Poem-Small / README.md
snzhang's picture
Update README.md
c192693
metadata
license: apache-2.0
language:
  - zh
library_name: transformers
pipeline_tag: text-generation
inference:
  parameters:
    temperature: 0.7
    top_p: 0.6
    repetition_penalty: 1.1
    max_new_tokens: 128
    num_return_sequences: 3
    do_sample: true
tags:
  - art
widget:
  - 笔底江山助磅礴
  - (唐诗:秋思)诗词
  - (宋词:浣溪沙)秋
  - (对联)冬

Chinese Poem and Couplt small GPT2 Model

Model description

The model is used to generate Chinese ancient poems and couplets. It is based on the IDEA-CCNL/Wenzhong-GPT2-110M

How to use

You can use the model directly with a pipeline for text generation:

When the parameter skip_special_tokens is True:

>>> from transformers import BertTokenizer, GPT2LMHeadModel,TextGenerationPipeline
>>> tokenizer = BertTokenizer.from_pretrained("snzhang/GPT2-Poem-Small")
>>> model = GPT2LMHeadModel.from_pretrained("snzhang/GPT2-Poem-Small")
>>> text_generator = TextGenerationPipeline(model, tokenizer)   
>>> text_generator("笔底江山助磅礴", max_length=50, do_sample=True)
    [{'generated_text':'笔底江山助磅礴,万卷诗书见成章。'}]

And you can add the prefix "(唐诗:your title)"、"(宋词:your title)" and "(对联)" to make generation more precise.

Training data

Training data contains 71,334 Chinese ancient poems and couplets which are collected by Chinese Poetry and Couplet Dataset

More Details

You can get more details in GPT2-Poem-Small