Edit model card

💫 Community Model> Llama 3 8B Instruct by Meta

👾 LM Studio Community models highlights program. Highlighting new & noteworthy models by the community. Join the conversation on Discord.

Model creator: meta-llama
Original model: Meta-Llama-3-8B-Instruct
GGUF quantization: provided by bartowski based on llama.cpp release b2777

Model Summary:

Llama 3 represents a huge update to the Llama family of models. This model is the 8B parameter instruction tuned model, meaning it's small, fast, and tuned for following instructions.
This model is very happy to follow the given system prompt, so use this to your advantage to get the behavior you desire.
Llama 3 excels at all the general usage situations, including multi turn conversations, general world knowledge, and coding.
This 8B model exceeds the performance of Llama 2's 70B model, showing that the performance is far greater than the previous iteration.

This model is made with the BPE fixes from llama.cpp

Prompt Template:

Choose the 'Llama 3' preset in your LM Studio.

Under the hood, the model will see a prompt that's formatted like so:

<|begin_of_text|><|start_header_id|>system<|end_header_id|>

{system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>

{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>

Use case and examples

Llama 3 should be great for anything you throw at it. Try it with conversations, coding, and just all around general inquiries.

Creative conversations

Using a system prompt of You are a pirate chatbot who always responds in pirate speak!

image/png

General knowledge

image/png

Coding

image/png

Technical Details

Llama 3 was trained on over 15T tokens from a massively diverse range of subjects and languages, and includes 4 times more code than Llama 2.

This model also features Grouped Attention Query (GQA) so that memory usage scales nicely over large contexts.

Instruction fine tuning was performed with a combination of supervised fine-tuning (SFT), rejection sampling, proximal policy optimization (PPO), and direct policy optimization (DPO).

Check out their blog post for more information here

Special thanks

🙏 Special thanks to Georgi Gerganov and the whole team working on llama.cpp for making all of this possible.

🙏 Special thanks to Kalomaze for his dataset (linked here) that was used for calculating the imatrix for these quants, which improves the overall quality!

Disclaimers

LM Studio is not the creator, originator, or owner of any Model featured in the Community Model Program. Each Community Model is created and provided by third parties. LM Studio does not endorse, support, represent or guarantee the completeness, truthfulness, accuracy, or reliability of any Community Model. You understand that Community Models can produce content that might be offensive, harmful, inaccurate or otherwise inappropriate, or deceptive. Each Community Model is the sole responsibility of the person or entity who originated such Model. LM Studio may not monitor or control the Community Models and cannot, and does not, take responsibility for any such Model. LM Studio disclaims all warranties or guarantees about the accuracy, reliability or benefits of the Community Models. LM Studio further disclaims any warranty that the Community Model will meet your requirements, be secure, uninterrupted or available at any time or location, or error-free, viruses-free, or that any errors will be corrected, or otherwise. You will be solely responsible for any damage resulting from your use of or access to the Community Models, your downloading of any Community Model, or use of any other Community Model provided by or through LM Studio.

Downloads last month
298
GGUF
Model size
8.03B params
Architecture
llama

4-bit

5-bit

6-bit

8-bit

Inference Examples
Unable to determine this model's library. Check the docs .

Model tree for lmstudio-community/Meta-Llama-3-8B-Instruct-BPE-fix-GGUF

Quantized
(178)
this model