Merry's picture
Update README.md
4af3531
---
tags:
- generated_from_trainer
model-index:
- name: pythia-70m-deduped-aid
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# pythia-70m-deduped-aid
![Example](https://cdn.discordapp.com/attachments/1042160561808482304/1070125215910211705/Screenshot_2023-01-31_at_15-34-45_KoboldAI_Client__mnt_mnt1_library-of-alexandria_ai-models_text-generation_my-models_pythia-70m-deduped-aid.png)
## Model description
This model is a finetune of [EleutherAI/pythia-70m-deduped](https://huggingface.co/EleutherAI/pythia-70m-deduped) (from when it was instead `pythia-19m-deduped`), on the [`text_adventures.txt`](https://github.com/Latitude-Archives/AIDungeon/blob/ca098ca7dab480d24e47954c8873b03ba1091ffc/data/text_adventures.txt) dataset originally intended for AI Dungeon 2. Performance will be very poor, as expected by the small model, and generations may be offensive thanks to its training data.
This model was trained for testing purposes and was intended for use with KoboldAI. A temperature of `0.5` and a repetition penalty of `1.01` were tested.
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
### Framework versions
- Transformers 4.26.0.dev0
- Pytorch 1.13.1+cu116
- Datasets 2.8.0
- Tokenizers 0.13.2