@fxmarty I converted this repo's safetensors model to ONNX (same config), but it's significantly larger than the pytorch equivalent (~500MB vs ~8MB).
I'm using the LlamaOnnxConfig from Optimum - any ideas for what could cause this? :) Thanks!
· Sign up or log in to comment