bartowski's picture
Add exl2 quant link
4a05214 verified
---
license: other
language:
- en
pipeline_tag: text-generation
---
# llama-3-neural-chat-v2.2-8b
<!-- Provide a quick summary of what the model is/does. -->
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6437292ecd93f4c9a34b0d47/6XQuhjWNr6C4RbU9f1k99.png)
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
I fine-tuned llama-3 8B on an approach similar to Intel's neural chat language model. I have slightly modified the data sources so it is stronger in coding, math, and writing. I use both SFT and DPO-Positive.
DPO-Positive dramatically improves performance over DPO.
- **Developed by:** Locutusque
- **Model type:** Built with Meta Llama 3
- **Language(s) (NLP):** Many?
- **License:** Llama 3 license https://huggingface.co/meta-llama/Meta-Llama-3-8B/blob/main/LICENSE
## Quants
GGUF: https://huggingface.co/bartowski/llama-3-neural-chat-v2.2-8B-GGUF
ExLlamaV2: https://huggingface.co/bartowski/llama-3-neural-chat-v2.2-8B-exl2
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
This model has great performance in writing, coding, and math.
## Training Data
Recipe information will be coming soon. This language model's recipe is similar to Intel's Neural Chat.
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
Conversational AI. This model is also very uncensored, it will respond to pretty much any request regardless of the system prompt, use at your own risk.
## Evaluations
| Tasks |Version| Filter |n-shot| Metric |Value | |Stderr|
|---------------------------------|-------|----------------|-----:|-----------|-----:|---|-----:|
|truthfulqa_mc2 | 2|none | 0|acc |0.5232|± |0.0151|
|gsm8k | 3|strict-match | 5|exact_match|0.5974|± |0.0135|
| | |flexible-extract| 5|exact_match|0.5974|± |0.0135|
|agieval_nous |N/A |none | 0|acc_norm |0.3841|± |0.0094|
| | |none | 0|acc |0.3802|± |0.0094|
| - agieval_aqua_rat | 1|none | 0|acc |0.2598|± |0.0276|
| | |none | 0|acc_norm |0.2520|± |0.0273|
| - agieval_logiqa_en | 1|none | 0|acc |0.3441|± |0.0186|
| | |none | 0|acc_norm |0.3687|± |0.0189|
| - agieval_lsat_ar | 1|none | 0|acc |0.2217|± |0.0275|
| | |none | 0|acc_norm |0.2348|± |0.0280|
| - agieval_lsat_lr | 1|none | 0|acc |0.3882|± |0.0216|
| | |none | 0|acc_norm |0.3824|± |0.0215|
| - agieval_lsat_rc | 1|none | 0|acc |0.4944|± |0.0305|
| | |none | 0|acc_norm |0.5019|± |0.0305|
| - agieval_sat_en | 1|none | 0|acc |0.6650|± |0.0330|
| | |none | 0|acc_norm |0.6553|± |0.0332|
| - agieval_sat_en_without_passage| 1|none | 0|acc |0.3981|± |0.0342|
| | |none | 0|acc_norm |0.3981|± |0.0342|
| - agieval_sat_math | 1|none | 0|acc |0.3500|± |0.0322|
| | |none | 0|acc_norm |0.3318|± |0.0318|