tinyllama-tarot-v1 / README.md
barissglc's picture
Update README.md
bca0eb0 verified
---
license: apache-2.0
library_name: peft
tags:
- trl
- sft
- generated_from_trainer
base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
model-index:
- name: tinyllama-tarot-v1
results: []
pipeline_tag: text2text-generation
widget:
- text: >-
Give me a one paragraph tarot reading if I pull the cards Nine of Cups, King
of Pentacles and Three of Cups.
datasets:
- barissglc/tarot
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# tinyllama-tarot-v1
This model is a fine-tuned version of [TinyLlama/TinyLlama-1.1B-Chat-v1.0](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0) .
## Model description
This model is a language model capable of making predictions based on tarot cards.
Trained to respond to questions related to topics such as love, career, and general life, tarot cards are the foundation of its predictions.
The model can make predictions based on the selected tarot cards. You can access the tarot cards from the tarot dataset.
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 8
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- training_steps: 250
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.17.1
- Tokenizers 0.15.2