inference: parameters: max_length: 500 do_sample: True temperature: 0.8

GPT-2

The GPT2 model is pre-trained on the writing style of Vu Trong Phung

How to use the model

from transformers import GPT2Tokenizer, GPT2LMHeadModel

tokenizer = GPT2Tokenizer.from_pretrained("Khoa/VN-Literature-Generation")
model = GPT2LMHeadModel.from_pretrained("Khoa/VN-Literature-Generation")


text = "Mùa thu lá vàng rơi"
input_ids = tokenizer.encode(text, return_tensors='pt')
max_length = 300
model.to('cpu')
sample_outputs = model.generate(input_ids,pad_token_id=tokenizer.eos_token_id,
                                   do_sample=True,
                                   max_length=max_length,
                                   min_length=max_length,
                                   top_k=40,
                                   num_beams=5,
                                   early_stopping=True,
                                   no_repeat_ngram_size=2,
                                   num_return_sequences=3)

for i, sample_output in enumerate(sample_outputs):
    print(">> Generated text {}\n\n{}".format(i+1, tokenizer.decode(sample_output.tolist())))
    print('\n---')

Author

Dong Dang Khoa

Downloads last month
21
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.