💫 Community Model> Llama 3 70B Instruct by Meta

👾 LM Studio Community models highlights program. Highlighting new & noteworthy models by the community. Join the conversation on Discord.

Model creator: meta-llama
Original model: Meta-Llama-3-70B-Instruct
GGUF quantization: provided by bartowski based on llama.cpp release b2777

Model Summary:

Llama 3 represents a huge update to the Llama family of models. This model is the 70B parameter instruction tuned model, with performance reaching and usually exceeding GPT-3.5.
This is a massive milestone, as an open model reaches the performance of a closed model over double its size.
This model is very happy to follow the given system prompt, so use this to your advantage to get the behavior you desire.
Llama 3 excels at all the general usage situations, including multi turn conversations, general world knowledge, and coding.

This model is made with the BPE fixes from llama.cpp

Prompt Template:

Choose the 'Llama 3' preset in your LM Studio.

Under the hood, the model will see a prompt that's formatted like so:

<|begin_of_text|><|start_header_id|>system<|end_header_id|>

{system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>

{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>

Use cases and examples to come.

Technical Details

Llama 3 was trained on over 15T tokens from a massively diverse range of subjects and languages, and includes 4 times more code than Llama 2.

This model also features Grouped Attention Query (GQA) so that memory usage scales nicely over large contexts.

Instruction fine tuning was performed with a combination of supervised fine-tuning (SFT), rejection sampling, proximal policy optimization (PPO), and direct policy optimization (DPO).

Only IQ1_M and IQ2_XS use importance matrix (iMatrix), the rest are made with the standard quant algorithms.

Check out their blog post for more information here

Special thanks

🙏 Special thanks to Georgi Gerganov and the whole team working on llama.cpp for making all of this possible.

🙏 Special thanks to Kalomaze for his dataset (linked here) that was used for calculating the imatrix for the IQ1_M and IQ2_XS quants, which makes them usable even at their tiny size!

Disclaimers

LM Studio is not the creator, originator, or owner of any Model featured in the Community Model Program. Each Community Model is created and provided by third parties. LM Studio does not endorse, support, represent or guarantee the completeness, truthfulness, accuracy, or reliability of any Community Model. You understand that Community Models can produce content that might be offensive, harmful, inaccurate or otherwise inappropriate, or deceptive. Each Community Model is the sole responsibility of the person or entity who originated such Model. LM Studio may not monitor or control the Community Models and cannot, and does not, take responsibility for any such Model. LM Studio disclaims all warranties or guarantees about the accuracy, reliability or benefits of the Community Models. LM Studio further disclaims any warranty that the Community Model will meet your requirements, be secure, uninterrupted or available at any time or location, or error-free, viruses-free, or that any errors will be corrected, or otherwise. You will be solely responsible for any damage resulting from your use of or access to the Community Models, your downloading of any Community Model, or use of any other Community Model provided by or through LM Studio.

Downloads last month
177
GGUF
Model size
70.6B params
Architecture
llama

1-bit

2-bit

3-bit

4-bit

Inference Examples
Unable to determine this model's library. Check the docs .

Model tree for lmstudio-community/Meta-Llama-3-70B-Instruct-BPE-fix-GGUF

Quantized
(46)
this model