File size: 1,306 Bytes
266997e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
---
language:
- en
tags:
- facebook
- meta
- pytorch
- llama
- llama-2
- mlx
extra_gated_heading: Access Llama 2 on Hugging Face
extra_gated_description: This is a form to enable access to Llama 2 on Hugging Face
  after you have been granted access from Meta. Please visit the [Meta website](https://ai.meta.com/resources/models-and-libraries/llama-downloads)
  and accept our license terms and acceptable use policy before submitting this form.
  Requests will be processed in 1-2 days.
extra_gated_button_content: Submit
extra_gated_fields:
  ? I agree to share my name, email address and username with Meta and confirm that
    I have already been granted download access on the Meta website
  : checkbox
pipeline_tag: text-generation
inference: false
---

# meriamcherif/Meta-Llama-3-8B-quantized

The Model [meriamcherif/Meta-Llama-3-8B-quantized](https://huggingface.co/meriamcherif/Meta-Llama-3-8B-quantized) was converted to MLX format from [NousResearch/Llama-2-7b-chat-hf](https://huggingface.co/NousResearch/Llama-2-7b-chat-hf) using mlx-lm version **0.14.3**.

## Use with mlx

```bash
pip install mlx-lm
```

```python
from mlx_lm import load, generate

model, tokenizer = load("meriamcherif/Meta-Llama-3-8B-quantized")
response = generate(model, tokenizer, prompt="hello", verbose=True)
```