Edit model card

This model is a GPTQ-quantized version of the Codestral-22B-v0.1 model with the FIM fix, designed to provide improved performance and efficiency. The original Codestral-22B-v0.1 model had a tokenization error for the FIM use case, which was addressed by the legraphista/Codestral-22B-v0.1-hf-FIM-fix model. This GPTQ-quantized model builds upon the fixed model, using default AutoGPTQ arguments and 4 bits per weight.

Compatible with vllm

Alternative to dan-kwiat/Codestral-22B-v0.1-hf-FIM-fix-awq

Tips for Code Completion

When using this model or any other Codestral-22B-v0.1 model for code completion in VSCode on Windows, make sure to use LF (Line Feed) instead of CRLF (Carriage Return Line Feed) for better results. This will prevent unrelated '\' characters from appearing in completions.

Downloads last month
2
Safetensors
Model size
3.33B params
Tensor type
I32
·
FP16
·
Inference API
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.