Shakespeare Fine-Tuned GPT-2 Model

Model Description

This is a fine-tuned version of the GPT-2 language model trained on the Tiny Shakespeare dataset. The model is optimized to generate text in the style of William Shakespeare, capturing the syntax, vocabulary, and poetic structure characteristic of his works.

Intended Use

The model is designed for educational purposes, creative writing, and experimentation with fine-tuned language models. Potential use cases include:

  • Generating Shakespearean-style text for creative projects.
  • Studying language modeling and fine-tuning techniques.
  • Providing inspiration for poetry or prose in Shakespearean English.

Usage

You can use this model via the Hugging Face Transformers library. Below is an example:

from transformers import AutoTokenizer, AutoModelForCausalLM

# Load model and tokenizer
model_name = "msttftmk/shakespeare-gpt2"
tokenizer = AutoTokenizer.from_pretrained(model_name, use_auth_token=True)
model = AutoModelForCausalLM.from_pretrained(model_name, use_auth_token=True)

# Generate text
input_text = "O gentle fair maiden,"
inputs = tokenizer.encode(input_text, return_tensors="pt")
outputs = model.generate(inputs, max_length=100, temperature=0.7)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Training Details

  • Base Model: GPT-2 (medium)
  • Dataset: Tiny Shakespeare dataset.
  • Fine-Tuning Framework: Hugging Face's Trainer API.
  • Training Parameters:
    • Learning rate: 2e-5
    • Epochs: 3
    • Batch size: 2
    • Max sequence length: 128

Evaluation

  • Validation Split: 10% of the dataset.
  • Evaluation Strategy: Per epoch evaluation during training.
  • Metrics: Loss and perplexity on validation data.

Limitations

  • Style-Restricted: The model generates text exclusively in a Shakespearean style. It is not intended for modern conversational or general-purpose language modeling.
  • Biases: The model inherits any biases present in the training dataset.
  • Dataset Limitations: The Tiny Shakespeare dataset is limited in size and scope, potentially restricting the richness and variability of the generated text.

Ethical Considerations

  • The model should not be used for generating harmful, offensive, or misleading content.
  • Users should ensure proper attribution when using this model for creative projects.

Citation

If you use this model, please cite:

@misc{shakespeare-gpt2,
  author = {Mustafa Tomak},
  title = {Shakespeare Fine-Tuned GPT-2},
  year = {2025},
  url = {https://huggingface.co/mstftmk/shakespeare-gpt2},
}

License

The model is released under the apache-2.0. Users must comply with the terms of use.

Downloads last month
14
Safetensors
Model size
355M params
Tensor type
F32
ยท
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.

Model tree for mstftmk/shakespeare-gpt2

Finetuned
(1342)
this model

Space using mstftmk/shakespeare-gpt2 1