---
base_model: remyxai/stablelm-zephyr-3B_localmentor
license: apache-2.0
language:
- en
pipeline_tag: text-generation
tags:
- stablelm
- zephyr
- gguf
library_name: llama.cpp
model_creator: remyxai
model_name: stablelm-zephyr-3B_localmentor
model_type: stablelm
prompt_template: |
<|system|>
{system_prompt}
<|user|>
{prompt}
<|assistant|>
quantized_by: mgonzs13
---
# stablelm-zephyr-3B-localmentor-GGUF
**Model creator:** [remyxai](https://huggingface.co/remyxai)
**Original model**: [stablelm-zephyr-3B_localmentor](https://huggingface.co/remyxai/stablelm-zephyr-3B_localmentor)
**GGUF quantization:** `llama.cpp` commit [fadde6713506d9e6c124f5680ab8c7abebe31837](https://github.com/ggerganov/llama.cpp/tree/fadde6713506d9e6c124f5680ab8c7abebe31837)
## Description
Fine-tune with low-rank adapters on 25K conversational turns discussing tech/startup from over 800 podcast episodes.
- **Developed by:** [Remyx.AI](https://huggingface.co/remyxai)
- **License:** apache-2.0
- **Finetuned from model:** [stablelm-zephyr-3b](https://huggingface.co/stabilityai/stablelm-zephyr-3b)
- **Repository**: https://github.com/remyxai/LocalMentor
## Prompt Template
Following the [tokenizer_config.json](https://huggingface.co/remyxai/stablelm-zephyr-3B_localmentor/blob/main/tokenizer_config.json), the prompt template is Zephyr.
```
<|system|>
{system_prompt}
<|user|>
{prompt}
<|assistant|>
```