bartowski commited on
Commit
9272924
1 Parent(s): 3c7aa71

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +73 -1
README.md CHANGED
@@ -186,6 +186,78 @@ extra_gated_fields:
186
  extra_gated_description: The information you provide will be collected, stored, processed and shared in accordance with the [Meta Privacy Policy](https://www.facebook.com/privacy/policy/).
187
  extra_gated_button_content: Submit
188
  quantized_by: bartowski
 
 
 
 
 
 
 
 
 
189
  ---
190
 
191
- Description coming soon.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
186
  extra_gated_description: The information you provide will be collected, stored, processed and shared in accordance with the [Meta Privacy Policy](https://www.facebook.com/privacy/policy/).
187
  extra_gated_button_content: Submit
188
  quantized_by: bartowski
189
+ lm_studio:
190
+ param_count: 8b
191
+ use_case: general
192
+ release_date: 18-04-2024
193
+ model_creator: meta-llama
194
+ prompt_template: Llama 3
195
+ system_prompt: You are a helpful AI assistant.
196
+ base_model: llama
197
+ original_repo: meta-llama/Meta-Llama-3-8B-Instruct
198
  ---
199
 
200
+ ## 💫 Community Model> Llama 3 8B Instruct by Meta
201
+
202
+ *👾 [LM Studio](https://lmstudio.ai) Community models highlights program. Highlighting new & noteworthy models by the community. Join the conversation on [Discord](https://discord.gg/aPQfnNkxGC)*.
203
+
204
+ **Model creator:** [meta-llama](https://huggingface.co/meta-llama)<br>
205
+ **Original model**: [Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct)<br>
206
+ **GGUF quantization:** provided by [bartowski](https://huggingface.co/bartowski) based on `llama.cpp` PR [6745](https://github.com/ggerganov/llama.cpp/pull/6745)<br>
207
+
208
+ ## Model Summary:
209
+ Llama 3 represents a huge update to the Llama family of models. This model is the 8B parameter instruction tuned model, meaning it's small, fast, and tuned for following instructions.<br>
210
+ This model is very happy to follow the given system prompt, so use this to your advantage to get the behavior you desire.<br>
211
+ Llama 3 excels at all the general usage situations, including multi turn conversations, general world knowledge, and coding.<br>
212
+ This 8B model exceeds the performance of Llama 2's 70B model, showing that the performance is far greater than the previous iteration.
213
+
214
+ ## Prompt Template:
215
+
216
+ Choose the 'Llama 3' preset in your LM Studio.
217
+
218
+ Under the hood, the model will see a prompt that's formatted like so:
219
+
220
+ ```
221
+ <|begin_of_text|><|start_header_id|>system<|end_header_id|>
222
+
223
+ {system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>
224
+
225
+ {prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
226
+
227
+ ```
228
+
229
+ ## Use case and examples
230
+
231
+ Llama 3 should be great for anything you throw at it. Try it with conversations, coding, and just all around general inquiries.
232
+
233
+ ## Creative conversations
234
+
235
+ Using a system prompt of `You are a pirate chatbot who always responds in pirate speak!`
236
+
237
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6435718aaaef013d1aec3b8b/PYIhzOZtKVSHEUq24u3ll.png)
238
+
239
+ ## General knowledge
240
+
241
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6435718aaaef013d1aec3b8b/3XDcR9e10CxcdVhmeco_W.png)
242
+
243
+ ## Coding
244
+
245
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6435718aaaef013d1aec3b8b/l-AHfv39hXG9IPzKqIBpv.png)
246
+
247
+ ## Technical Details
248
+
249
+ Llama 3 was trained on over 15T tokens from a massively diverse range of subjects and languages, and includes 4 times more code than Llama 2.
250
+
251
+ This model also features Grouped Attention Query (GQA) so that memory usage scales nicely over large contexts.
252
+
253
+ Instruction fine tuning was performed with a combination of supervised fine-tuning (SFT), rejection sampling, proximal policy optimization (PPO), and direct policy optimization (DPO).
254
+
255
+ ## Special thanks
256
+
257
+ 🙏 Special thanks to [Georgi Gerganov](https://github.com/ggerganov) and the whole team working on [llama.cpp](https://github.com/ggerganov/llama.cpp/) for making all of this possible.
258
+
259
+ 🙏 Special thanks to [Kalomaze](https://github.com/kalomaze) for his dataset (linked [here](https://github.com/ggerganov/llama.cpp/discussions/5263)) that was used for calculating the imatrix for these quants, which improves the overall quality!
260
+
261
+ ## Disclaimers
262
+
263
+ LM Studio is not the creator, originator, or owner of any Model featured in the Community Model Program. Each Community Model is created and provided by third parties. LM Studio does not endorse, support, represent or guarantee the completeness, truthfulness, accuracy, or reliability of any Community Model. You understand that Community Models can produce content that might be offensive, harmful, inaccurate or otherwise inappropriate, or deceptive. Each Community Model is the sole responsibility of the person or entity who originated such Model. LM Studio may not monitor or control the Community Models and cannot, and does not, take responsibility for any such Model. LM Studio disclaims all warranties or guarantees about the accuracy, reliability or benefits of the Community Models. LM Studio further disclaims any warranty that the Community Model will meet your requirements, be secure, uninterrupted or available at any time or location, or error-free, viruses-free, or that any errors will be corrected, or otherwise. You will be solely responsible for any damage resulting from your use of or access to the Community Models, your downloading of any Community Model, or use of any other Community Model provided by or through LM Studio.