I want to do llama batch inference so that it can run in multiple parallel environments. Can I implement it using Llama-2-70B-Chat-GPTQ?
· Sign up or log in to comment