Finetune Mistral, Gemma, Llama 2-5x faster with 70% less memory via Unsloth!

We have a Google Colab Tesla T4 notebook for Mistral v3 7b here: https://colab.research.google.com/drive/1_yNCks4BTD5zOnjozppphh5GzMFaMKq_?usp=sharing

For conversational ShareGPT style and using Mistral v3 Instruct: https://colab.research.google.com/drive/15F1xyn8497_dUbxZP4zWmPZ3PJx1Oymv?usp=sharing

✨ Finetune for Free

All notebooks are beginner friendly! Add your dataset, click "Run All", and you'll get a 2x faster finetuned model which can be exported to GGUF, vLLM or uploaded to Hugging Face.

Unsloth supports Free Notebooks Performance Memory use
Llama-3.2 (3B) ▢️ Start on Colab 2.4x faster 58% less
Llama-3.2 (11B vision) ▢️ Start on Colab 2x faster 60% less
Llama-3.1 (8B) ▢️ Start on Colab 2.4x faster 58% less
Qwen2 VL (7B) ▢️ Start on Colab 1.8x faster 60% less
Qwen2.5 (7B) ▢️ Start on Colab 2x faster 60% less
Phi-3.5 (mini) ▢️ Start on Colab 2x faster 50% less
Gemma 2 (9B) ▢️ Start on Colab 2.4x faster 58% less
Mistral (7B) ▢️ Start on Colab 2.2x faster 62% less
DPO - Zephyr ▢️ Start on Colab 1.9x faster 19% less

Downloads last month
11,707
Safetensors
Model size
3.86B params
Tensor type
F32
Β·
BF16
Β·
U8
Β·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for unsloth/mistral-7b-instruct-v0.2-bnb-4bit

Adapters
29 models
Finetunes
396 models
Quantizations
89 models

Spaces using unsloth/mistral-7b-instruct-v0.2-bnb-4bit 9

Collection including unsloth/mistral-7b-instruct-v0.2-bnb-4bit