Edit model card

GPT2 Alpaca Clean

This repository contains a fine-tuned version of the GPT2 language model, trained on the alpaca-cleaned dataset. The alpaca-cleaned dataset is a cleaned and filtered version of the Stanford Alpaca dataset. The model showed improvements on instruction related tasks.

graph

Model Details

The GPT2 model was fine-tuned using the Hugging Face Transformers library on the alpaca-cleaned dataset. The fine-tuned model can be used for various natural language processing tasks, such as text generation, summarization, and question answering.

Usage

To use the fine-tuned model, you can load it using the Hugging Face Transformers library. Here's an example:

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "Rjonah321/gpt2-alpaca-clean"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

input_text = "Write a short story about a dog."
input_ids = tokenizer.encode(input_text, return_tensors='pt')

output = model.generate(input_ids, max_length=200, do_sample=True, top_k=50, top_p=0.95, num_return_sequences=1)
generated_text = tokenizer.decode(output[0], skip_special_tokens=True)

print(generated_text)

License

This project is licensed under the MIT License - see the LICENSE file for details.

Acknowledgements

@misc{alpaca_cleaned_2023,
author = {Yahma},
title = {Alpaca-cleaned Dataset},
year = 2023,
url = {https://huggingface.co/datasets/yahma/alpaca-cleaned},
note = {Accessed: 2024-06-19},
license = {CC-BY-4.0}
}
Downloads last month
2
Safetensors
Model size
124M params
Tensor type
BF16
·
Inference API
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.