Trained for 2 epochs using mpasila/Finnish-ShareGPT-Tiny-V1-1 in 2048 context with LoRA Rank set to 256 with Alpha set to 512.

Prompt format: ChatML

Works better with a system prompt.

Uploaded model

  • Developed by: mpasila
  • License: apache-2.0
  • Finetuned from model : LumiOpen/Viking-7B

This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.

Downloads last month
21
Safetensors
Model size
7.55B params
Tensor type
FP16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for mpasila/Finnish-Chatty-Tiny-V1-1-7B

Base model

LumiOpen/Viking-7B
Finetuned
(31)
this model

Dataset used to train mpasila/Finnish-Chatty-Tiny-V1-1-7B

Collection including mpasila/Finnish-Chatty-Tiny-V1-1-7B