Unable to convert ONNX model to INT4/FP16

#15
by Avan2000 - opened

Hi community,

I tried exporting the Qwen to ONNX model using the command -

optimum-cli export onnx --model Qwen/CodeQwen1.5-7B-Chat --framework pt --task text-generation-with-past Qwen/CodeQwen1.5-7B-Chat

and successfully retrieved a .onnx model.
Now while trying to convert it to FP16/INT4 precision, I am facing the error stating it has exceeded protobuff memory capacity of 2GB as -

convertion_error.png

What should be the next step for this conversion ?

Hi Community, Any update on this ?

Sign up or log in to comment