T-Ragx Quantized
Collection
7 items
•
Updated
Merged and GPTQ quantized version of rayliuca/TRagx-internlm2-7b
Note: I'm having some difficulties quantizing the models using GPTQ. Mistral and NeuralOmniBeagle's GPTQ models have significantly degraded output, while quantized TowerInstruct v0.2 was not working out right
While this quantized model for InternLM2 seems to work all right, the translation accuracy is not validated.
These AWQ quantized models are recommended:
Qutanized with nsamples=45 * 3 languages [ja, zh, en] from the c4 dataset
See the original InternLM2 repo https://huggingface.co/internlm/internlm2-7b#open-source-license