Llama-3.3-70B-o1 GGUF Quants

This repository contains the GGUF quants for the Llama-3.3-70B-o1 model. You can use them for inference in local inference servers like ollama or llama.cpp

Downloads last month
133
GGUF
Model size
70.6B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

4-bit

5-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support