English
Hrishi
Added README
ea63625
|
raw
history blame
1.8 kB
---
base_model: ehartford/dolphin-2.1-mistral-7b
datasets:
- ehartford/dolphin
- jondurbin/airoboros-2.2.1
language:
- en
license: apache-2.0
model-index:
- name: mlc-chat-dolphin-2.2.1-mistral-7b
results: []
model_creator: Hugging Face H4
model_name: WASM Dolphin 2.2.1
model_type: mistral
prompt_template: '<|im_start|>system
You are Dolphin, a helpful AI assistant.<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant'
---
# Dolphin 2.2.1 (Finetune of Mistral 7B) compiled for WebGPU - q4f32_1
- Original model: [Dolphin 2.1 🐬](https://huggingface.co/ehartford/dolphin-2.1-mistral-7b)
- creator: Eric Hartford: [https://erichartford.com/dolphin](https://erichartford.com/dolphin)
- compiled by: Hrishi Olickel: [say hi on Twitter!](https://twitter.com/hrishioa)
## Description
This is a quantized version of Dolphin 2.1 🐬, one of the best finetunes of [Mistral-7b](https://huggingface.co/mistralai/Mistral-7B-v0.1) out there, ready to be used for on-browser inference over WebGPU.
Compiled with [mlc-llm](https://llm.mlc.ai/).
Very helpful direction provided by [felladrin](https://github.com/felladrin)!
You can use [his example](https://huggingface.co/spaces/Felladrin/Web-LLM-Mistral-7B-OpenOrca) to get quickly started with this model.
## Prompt template: Dolphin
Prompt format:
This model (and all my future releases) use [ChatML](https://github.com/openai/openai-python/blob/main/chatml.md) prompt format.
```
<|im_start|>system
You are Dolphin, a helpful AI assistant.<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant
```
Example:
```
<|im_start|>system
you are an expert dolphin trainer<|im_end|>
<|im_start|>user
What is the best way to train a dolphin to obey me? Please answer step by step.<|im_end|>
<|im_start|>assistant
```