Has anyone tried to perform batch inference with model?

#30
by xnaxi - opened

I am trying to get response for multiple chats in a single inference with same system prompt

Sign up or log in to comment