|
--- |
|
base_model: ehartford/dolphin-2.1-mistral-7b |
|
datasets: |
|
- ehartford/dolphin |
|
- jondurbin/airoboros-2.2.1 |
|
language: |
|
- en |
|
license: apache-2.0 |
|
model-index: |
|
- name: mlc-chat-dolphin-2.2.1-mistral-7b |
|
results: [] |
|
model_creator: Hugging Face H4 |
|
model_name: WASM Dolphin 2.2.1 |
|
model_type: mistral |
|
prompt_template: '<|im_start|>system |
|
You are Dolphin, a helpful AI assistant.<|im_end|> |
|
<|im_start|>user |
|
{prompt}<|im_end|> |
|
<|im_start|>assistant' |
|
--- |
|
|
|
# Dolphin 2.2.1 (Finetune of Mistral 7B) compiled for WebGPU - q4f32_1 |
|
|
|
- Original model: [Dolphin 2.1 🐬](https://huggingface.co/ehartford/dolphin-2.1-mistral-7b) |
|
- creator: Eric Hartford: [https://erichartford.com/dolphin](https://erichartford.com/dolphin) |
|
- compiled by: Hrishi Olickel: [say hi on Twitter!](https://twitter.com/hrishioa) |
|
|
|
## Description |
|
|
|
This is a quantized version of Dolphin 2.1 🐬, one of the best finetunes of [Mistral-7b](https://huggingface.co/mistralai/Mistral-7B-v0.1) out there, ready to be used for on-browser inference over WebGPU. |
|
|
|
Compiled with [mlc-llm](https://llm.mlc.ai/). |
|
|
|
Very helpful direction provided by [felladrin](https://github.com/felladrin)! |
|
|
|
You can use [his example](https://huggingface.co/spaces/Felladrin/Web-LLM-Mistral-7B-OpenOrca) to get quickly started with this model. |
|
|
|
## Prompt template: Dolphin |
|
|
|
Prompt format: |
|
This model (and all my future releases) use [ChatML](https://github.com/openai/openai-python/blob/main/chatml.md) prompt format. |
|
|
|
``` |
|
<|im_start|>system |
|
You are Dolphin, a helpful AI assistant.<|im_end|> |
|
<|im_start|>user |
|
{prompt}<|im_end|> |
|
<|im_start|>assistant |
|
|
|
``` |
|
|
|
Example: |
|
|
|
``` |
|
<|im_start|>system |
|
you are an expert dolphin trainer<|im_end|> |
|
<|im_start|>user |
|
What is the best way to train a dolphin to obey me? Please answer step by step.<|im_end|> |
|
<|im_start|>assistant |
|
``` |
|
|