| from transformers import GPT2LMHeadModel, GPT2Tokenizer | |
| model_name = "gpt2" # You can also use "gpt2-medium", "gpt2-large", "gpt2-xl" | |
| # Download the tokenizer and model | |
| tokenizer = GPT2Tokenizer.from_pretrained(model_name) | |
| model = GPT2LMHeadModel.from_pretrained(model_name) | |
| # Optional test | |
| input_ids = tokenizer.encode("The world is", return_tensors="pt") | |
| output = model.generate(input_ids, max_length=20) | |
| print(tokenizer.decode(output[0])) | |