Text Generation
Transformers
PyTorch
English
llama
Inference Endpoints
text-generation-inference

Set use_cache to True, otherwise inference performance is poor

#2
by TheBloke - opened
Open Access AI Collective org
No description provided.
winglian changed pull request status to merged

Sign up or log in to comment