Edit model card

tinyllama-tarot-v1

This model is a fine-tuned version of TinyLlama/TinyLlama-1.1B-Chat-v1.0 .

Model description

This model is a language model capable of making predictions based on tarot cards. Trained to respond to questions related to topics such as love, career, and general life, tarot cards are the foundation of its predictions. The model can make predictions based on the selected tarot cards. You can access the tarot cards from the tarot dataset.

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0002
  • train_batch_size: 4
  • eval_batch_size: 8
  • seed: 42
  • gradient_accumulation_steps: 8
  • total_train_batch_size: 32
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: cosine
  • training_steps: 250
  • mixed_precision_training: Native AMP

Training results

Framework versions

  • PEFT 0.8.2
  • Transformers 4.37.2
  • Pytorch 2.1.0+cu121
  • Datasets 2.17.1
  • Tokenizers 0.15.2
Downloads last month
4
Inference Examples
Inference API (serverless) does not yet support peft models for this pipeline type.

Adapter for

Dataset used to train barissglc/tinyllama-tarot-v1