Configuration Parsing Warning: In config.json: "quantization_config.bits" must be an integer

IceCaffeLatteRP-7b-4.32bpw-h8-exl2

This is a 4.32bpw(4.3295) h8 exl2 quant of IceCaffeLatteRP-7b

Built-in(default) calibration dataset is used.

Downloads last month
0
Inference API
This model can be loaded on Inference API (serverless).