convert ggml-vicuna-7b-f16 to ggml-vicuna-7b-q4_0
Source: https://huggingface.co/chharlesonfire/ggml-vicuna-7b-f16
No unnecessary changes
Usage:
Download llama.cpp from https://github.com/ggerganov/llama.cpp
make and run llama.cpp and choose model with ggml-vicuna-7b-q4_0.bin
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API:
The model has no library tag.