Edit model card

wit

This model is a fine-tuned version of igorktech/rugpt3-joker-150k on the baneks dataset for 10 epochs. It achieved 2.0391 overall loss during training. Model evaluation has not been performed.

Model description

The model is a fine-tuned variant of the igorktech/rugpt3-joker-150k architecture with causal language modeling head.

Intended uses & limitations

The model should be used for studying abilities of natural language models to generate jokes.

Training and evaluation data

The model is trained on a list of anecdotes pulled from a few vk communities (see baneks dataset for more details).

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0005
  • train_batch_size: 8
  • eval_batch_size: 16
  • seed: 42
  • gradient_accumulation_steps: 8
  • total_train_batch_size: 64
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_steps: 1000
  • num_epochs: 10

Training results

Train Loss Epoch
2.0391 10

Framework versions

  • Transformers 4.34.0
  • Pytorch 2.1.0
  • Datasets 2.12.0
  • Tokenizers 0.14.1
Downloads last month
11
Safetensors
Model size
125M params
Tensor type
F32
·
Inference API
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for zeio/wit

Finetuned
this model

Dataset used to train zeio/wit

Collection including zeio/wit