GPUs Requirement For Inference
#42
by
Aillian
- opened
What is the GPUs requirement for inference on this model?
I wrote this down by mistake.
Any answer ?
I am running exl2 5.0bpw with 64k context at this very moment on 90.72 GB