Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
@@ -28,21 +28,21 @@ Quantized GGUF model files for [TinyMistral-248M](https://huggingface.co/Locutus
|
|
28 |
|
29 |
| Name | Quant method | Size |
|
30 |
| ---- | ---- | ---- |
|
31 |
-
| [tinymistral-248m.fp16.gguf](https://huggingface.co/afrideva/TinyMistral-248M-GGUF/resolve/main/tinymistral-248m.fp16.gguf) | fp16 |
|
32 |
-
| [tinymistral-248m.q2_k.gguf](https://huggingface.co/afrideva/TinyMistral-248M-GGUF/resolve/main/tinymistral-248m.q2_k.gguf) | q2_k |
|
33 |
-
| [tinymistral-248m.q3_k_m.gguf](https://huggingface.co/afrideva/TinyMistral-248M-GGUF/resolve/main/tinymistral-248m.q3_k_m.gguf) | q3_k_m |
|
34 |
-
| [tinymistral-248m.q4_k_m.gguf](https://huggingface.co/afrideva/TinyMistral-248M-GGUF/resolve/main/tinymistral-248m.q4_k_m.gguf) | q4_k_m |
|
35 |
-
| [tinymistral-248m.q5_k_m.gguf](https://huggingface.co/afrideva/TinyMistral-248M-GGUF/resolve/main/tinymistral-248m.q5_k_m.gguf) | q5_k_m |
|
36 |
-
| [tinymistral-248m.q6_k.gguf](https://huggingface.co/afrideva/TinyMistral-248M-GGUF/resolve/main/tinymistral-248m.q6_k.gguf) | q6_k |
|
37 |
-
| [tinymistral-248m.q8_0.gguf](https://huggingface.co/afrideva/TinyMistral-248M-GGUF/resolve/main/tinymistral-248m.q8_0.gguf) | q8_0 |
|
38 |
|
39 |
|
40 |
|
41 |
## Original Model Card:
|
42 |
-
A pre-trained language model, based on the Mistral 7B model, has been scaled down to approximately 248 million parameters.
|
43 |
This model should have a context length of around 32,768 tokens.
|
44 |
|
45 |
-
During evaluation on InstructMix, this model achieved an average perplexity score of 6.3.
|
46 |
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
47 |
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Locutusque__TinyMistral-248m)
|
48 |
|
|
|
28 |
|
29 |
| Name | Quant method | Size |
|
30 |
| ---- | ---- | ---- |
|
31 |
+
| [tinymistral-248m.fp16.gguf](https://huggingface.co/afrideva/TinyMistral-248M-GGUF/resolve/main/tinymistral-248m.fp16.gguf) | fp16 | 497.75 MB |
|
32 |
+
| [tinymistral-248m.q2_k.gguf](https://huggingface.co/afrideva/TinyMistral-248M-GGUF/resolve/main/tinymistral-248m.q2_k.gguf) | q2_k | 116.20 MB |
|
33 |
+
| [tinymistral-248m.q3_k_m.gguf](https://huggingface.co/afrideva/TinyMistral-248M-GGUF/resolve/main/tinymistral-248m.q3_k_m.gguf) | q3_k_m | 131.01 MB |
|
34 |
+
| [tinymistral-248m.q4_k_m.gguf](https://huggingface.co/afrideva/TinyMistral-248M-GGUF/resolve/main/tinymistral-248m.q4_k_m.gguf) | q4_k_m | 156.60 MB |
|
35 |
+
| [tinymistral-248m.q5_k_m.gguf](https://huggingface.co/afrideva/TinyMistral-248M-GGUF/resolve/main/tinymistral-248m.q5_k_m.gguf) | q5_k_m | 180.16 MB |
|
36 |
+
| [tinymistral-248m.q6_k.gguf](https://huggingface.co/afrideva/TinyMistral-248M-GGUF/resolve/main/tinymistral-248m.q6_k.gguf) | q6_k | 205.20 MB |
|
37 |
+
| [tinymistral-248m.q8_0.gguf](https://huggingface.co/afrideva/TinyMistral-248M-GGUF/resolve/main/tinymistral-248m.q8_0.gguf) | q8_0 | 265.26 MB |
|
38 |
|
39 |
|
40 |
|
41 |
## Original Model Card:
|
42 |
+
A pre-trained language model, based on the Mistral 7B model, has been scaled down to approximately 248 million parameters. This model has been trained on 7,320,000 examples. This model isn't intended for direct use but for fine-tuning on a downstream task.
|
43 |
This model should have a context length of around 32,768 tokens.
|
44 |
|
45 |
+
During evaluation on InstructMix, this model achieved an average perplexity score of 6.3. This is the final epoch planned for this model.
|
46 |
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
47 |
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Locutusque__TinyMistral-248m)
|
48 |
|