--- license: apache-2.0 language: - el - en model_creator: ilsp base_model: ilsp/Meltemi-7B-Instruct-v1 library_name: gguf prompt_template: | [INST] {prompt} [/INST] quantized_by: ilsp --- # Meltemi 7B Instruct Quantized models ![image/png](https://miro.medium.com/v2/resize:fit:720/format:webp/1*IaE7RJk6JffW8og-MOnYCA.png) ## Description In this repository you can find quantised GGUF variants of [Meltemi-7B-Instruct-v1](https://huggingface.co/ilsp/Meltemi-7B-Instruct-v1) model, created using [llama.cpp](https://github.com/ggerganov/llama.cpp) at the [Institute for Language and Speech Processing](https://www.athenarc.gr/en/ilsp) of [Athena Research & Innovation Center](https://www.athenarc.gr/en). ## Provided files (Use case column taken from the llama.cpp documentation) Based on the information | Name | Quant method | Bits | Size | Appr. RAM required | Use case | | ---- | ---- | ---- | ---- | ---- | ----- | | [meltemi-instruct-v1_q3_K_M.bin](https://huggingface.co/ilsp/Meltemi-7B-Instruct-v1-GGUF/blob/main/meltemi-instruct-v1_q3_K_M.bin) | Q3_K_M | 3 | 3.67 GB| 6.45 GB | small, high quality loss | | [meltemi-instruct-v1_q5_K_M.bin](https://huggingface.co/ilsp/Meltemi-7B-Instruct-v1-GGUF/blob/main/meltemi-instruct-v1_q5_K_M.bin) | Q5_K_M | 5 | 5.31 GB| 8.1 GB | large, low quality loss - recommended |