Edit model card

Directly quantized 4bit model with bitsandbytes.

Unsloth can finetune LLMs with QLoRA 2.2x faster and use 62% less memory!

We have a Google Colab Tesla T4 notebook for Llama 7b here: https://colab.research.google.com/drive/1lBzz5KeZJKXjvivbYvmGarix9Ao6Wxe5?usp=sharing

Downloads last month
2,109
Safetensors
Model size
6.74B params
Tensor type
FP16
·

Space using unsloth/llama-2-7b 1

Collection including unsloth/llama-2-7b