|
--- |
|
license: apache-2.0 |
|
datasets: |
|
- Locutusque/hercules-v1.0 |
|
- Open-Orca/SlimOrca-Dedup |
|
language: |
|
- en |
|
base_model: Locutusque/TinyMistral-248M-v2.5 |
|
tags: |
|
- chemistry |
|
- biology |
|
- not-for-all-audiences |
|
- merge |
|
- code |
|
inference: |
|
parameters: |
|
do_sample: true |
|
renormalize_logits: false |
|
temperature: 0.8 |
|
top_p: 0.14 |
|
top_k: 12 |
|
min_new_tokens: 2 |
|
max_new_tokens: 96 |
|
repetition_penalty: 1.15 |
|
no_repeat_ngram_size: 5 |
|
epsilon_cutoff: 0.002 |
|
widget: |
|
- text: "<|im_start|>user\nWrite me a Python program that calculates the factorial of n. <|im_end|>\n<|im_start|>assistant\n" |
|
--- |
|
# Model description |
|
Fine-tuned Locutusque/TinyMistral-248m-v2.5 on SlimOrca-Dedup and Hercules-v1.0. Averaged a loss of 1.5 during training. This model's performance is excellent considering it's size. |
|
|
|
This model may output X-rated content. You and you alone are responsible for downloading and using the model and it's outputs. You have been warned. |
|
|
|
You can use the ChatML prompt format for this model. |
|
# Evaluation |
|
This model will be submitted to the Open LLM Leaderboard. |