finetuned smol 220M
Collection
smol_llama 220M fine-tunes we did
•
6 items
•
Updated
•
1
This model is a fine-tuned version of BEE-spoke-data/smol_llama-220M-openhermes on the Ultrachat_200k dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
1.6447 | 1.0 | 1624 | 1.6579 |
https://wandb.ai/amazingvince/huggingface/runs/5rffzk3x/workspace?workspace=user-amazingvince
Base model
BEE-spoke-data/smol_llama-220M-GQA