File size: 405 Bytes
65abf12
 
 
 
 
1
2
3
4
5
Upload bloomz-7b1-q4_0.gguf, converted from [bloomz-7b1](https://huggingface.co/bigscience/bloomz-7b1/tree/main) by llama.cpp's bash [convert-hf-to-gguf.py](https://github.com/ggerganov/llama.cpp/blob/master/convert-hf-to-gguf.py).
```bash
python convert-hf-to-gguf.py --outfile bloomz-7b1.gguf --outtype f16 /mnt/disk1/models/bloomz-7b1/
./build/bin/quantize bloomz-7b1.gguf bloomz-7b1-q4_0.gguf q4_0
```