Text Generation
Transformers
Safetensors
llama
4-bit precision
AWQ
Inference Endpoints
generated_from_trainer
conversational
text-generation-inference
Edit model card

cognitivecomputations/dolphin-2.9-llama3-8b AWQ

Curated and trained by Eric Hartford, Lucas Atkins, and Fernando Fernandes, and Cognitive Computations

Discord: https://discord.gg/8fbBeC7ZGx

My appreciation for the sponsors of Dolphin 2.9:

This model is based on Llama-3-8b, and is governed by META LLAMA 3 COMMUNITY LICENSE AGREEMENT

The base model has 8k context, and the full-weight fine-tuning was with 4k sequence length.

It took 2.5 days on 8x L40S provided by Crusoe Cloud

This model was trained FFT on all parameters, using ChatML prompt template format.

example:

<|im_start|>system
You are Dolphin, a helpful AI assistant.<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant
Downloads last month
246
Safetensors
Model size
1.98B params
Tensor type
I32
·
FP16
·
Inference API
Input a message to start chatting with solidrust/dolphin-2.9-llama3-8b-AWQ.
Inference API (serverless) has been turned off for this model.

Quantized from

Datasets used to train solidrust/dolphin-2.9-llama3-8b-AWQ

Collection including solidrust/dolphin-2.9-llama3-8b-AWQ