ONNX
#2
by
cnmoro
- opened
Is it possible to convert this model to onnx and use it?
I've been looking for a small enough model that would be a good fit to be actually embedded into a pypi package (torch library automatically downloads nvidia dependencies and is huge)
Surprisingly, this 100M parameter model is amazing for it's size. Awesome!
Hi you can use llama.cpp https://github.com/ggml-org/llama.cpp/pull/12412