Text Generation
Transformers
PyTorch
llama
text-generation-inference
Inference Endpoints

Maximum sequence length for a Llama 2 model is 4096

#3
by TheBloke - opened
No description provided.
WizardLM changed pull request status to merged
This comment has been hidden
This comment has been hidden
This comment has been hidden

Sign up or log in to comment