Text Generation
Transformers
PyTorch
Safetensors
English
llama
text-generation-inference
Inference Endpoints

Call w/ LiteLLM

#4
by krrish-litellm - opened

Hi @flavoredquark / @arielnlee

What's the best way for me to deploy this model? I'd love to make a demo of this with LiteLLM - https://github.com/BerriAI/litellm.

Lite currently works with Replicate, Azure, Together.ai and HF Inference Endpoints.

I'm facing issues with HF Inference endpoints due to quota limitations, so curious if you've tried any other provider.

Sign up or log in to comment