Hardware requirement for inference

#1
by ronghash - opened

Hi Shailja,

I'm pretty interested in this model. If I only want to do inference with the model, how many GPUs and what is version of GPU do I need.

Best,
Zhirong

Hello Zhirong,

I used one a100 for inference without any optimization to memory and GPU parallelization,

and with optimization (fauxpilot), I could use one rtx8000 to do the inferecing using 16B model

This comment has been hidden

Sign up or log in to comment