Edit model card

This is a quantized GGUF version of Microsoft Phi-2 to 4_0, 8_0 bits and the converted 16 FP model.

(link to the original model : https://huggingface.co/microsoft/phi-2)

Disclamer : make sure to have the latest version of llama.cpp after commit b9e74f9bca5fdf7d0a22ed25e7a9626335fdfa48

Downloads last month
1,824
GGUF
Inference Examples
Inference API (serverless) has been turned off for this model.

Space using kroonen/phi-2-GGUF 1