Edit model card

This text generator is based on OpenAI GPT2 model from HuggingFace Base model went through two step of learning

First - Finetining of base model

On this step model is finetuned on dataset of single sentence from the texts of Dostovesky F.M.

Training parameters:

  • Epoch = 10
  • Learning Rate = 1e-3
  • Optimizer = AdamW
  • Scheduler = OneCycleLR
  • Training env = PyTorch

image.png

image.png

Second - RL

On this step finetuned model went trough reinforcement learning pipline with TRL library.

Training parameters:

  • Epoch = 30
  • Trainer = PPO
  • Query texts = first 100 texts from dataset, trimmed by first 3 words
  • Reward = score from binary classifier multiplied by 10

image.png

image.png

Downloads last month
3
Safetensors
Model size
124M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Space using Roaoch/CyberClassic-Generator 1

Collection including Roaoch/CyberClassic-Generator