language: | |
- en | |
tags: | |
- meta | |
- pytorch | |
- llama | |
- llama-2 | |
- mlx | |
extra_gated_heading: Access Llama 2 on Hugging Face | |
extra_gated_description: This is a form to enable access to Llama 2 on Hugging Face | |
after you have been granted access from Meta. Please visit the [Meta website](https://ai.meta.com/resources/models-and-libraries/llama-downloads) | |
and accept our license terms and acceptable use policy before submitting this form. | |
Requests will be processed in 1-2 days. | |
extra_gated_button_content: Submit | |
extra_gated_fields: | |
? I agree to share my name, email address and username with Meta and confirm that | |
I have already been granted download access on the Meta website | |
: checkbox | |
pipeline_tag: text-generation | |
inference: false | |
# meriamcherif/Meta-Llama-3-8B-quantized | |
The Model [meriamcherif/Meta-Llama-3-8B-quantized](https://huggingface.co/meriamcherif/Meta-Llama-3-8B-quantized) was converted to MLX format from [NousResearch/Llama-2-7b-chat-hf](https://huggingface.co/NousResearch/Llama-2-7b-chat-hf) using mlx-lm version **0.14.3**. | |
## Use with mlx | |
```bash | |
pip install mlx-lm | |
``` | |
```python | |
from mlx_lm import load, generate | |
model, tokenizer = load("meriamcherif/Meta-Llama-3-8B-quantized") | |
response = generate(model, tokenizer, prompt="hello", verbose=True) | |
``` | |