GPUs Requirement For Inference

#42
by Aillian - opened

What is the GPUs requirement for inference on this model?

I wrote this down by mistake.

Any answer ?

I am running exl2 5.0bpw with 64k context at this very moment on 90.72 GB

Sign up or log in to comment