Edit model card

Model Card for Model ID

Fine tuning (learning/educational) results of GPT2-medium on a customized dataset

Testing different ways to provide a though process withing GPT2

This one seemed like the best results for now..

Model Details

Base model : 'hf_models/gpt2-medium'

Direct Use

from transformers import GPT2LMHeadModel, GPT2Tokenizer

models_folder = "Deeokay/gpt2-javis-stks"

# if you know your device, you can just set "device = 'mps'" 
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")

model = GPT2LMHeadModel.from_pretrained(models_folder)
tokenizer = GPT2Tokenizer.from_pretrained(models_folder)

tokenizer.pad_token = tokenizer.eos_token

prompt = "what is the meaning of life?"

inputs = tokenizer(prompt, return_tensors="pt", padding=True, truncation=True).to(device)

set_seed(42)

sample_output = model.generate(
    **inputs,
    max_new_tokens=250,
    do_sample=True,
    top_k=30,
    temperature=0.7,
)

print("Output:\n" + 80 * '-')
print(tokenizer.decode(sample_output[0], skip_special_tokens=True))

[More Information Needed]

Training Details

Training Data

[More Information Needed]

Training Procedure

Evaluation

Summary

Downloads last month
7
Safetensors
Model size
355M params
Tensor type
F32
·
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.