Fix max tokens in LLM2Vec Models

#96
Files changed (1) hide show
  1. app.py +6 -6
app.py CHANGED
@@ -721,12 +721,12 @@ EXTERNAL_MODEL_TO_SEQLEN = {
721
  "Cohere-embed-multilingual-light-v3.0": 512,
722
  "DanskBERT": 514,
723
  "LASER2": "N/A",
724
- "LLM2Vec-Llama-supervised": 512,
725
- "LLM2Vec-Llama-unsupervised": 512,
726
- "LLM2Vec-Mistral-supervised": 512,
727
- "LLM2Vec-Mistral-unsupervised": 512,
728
- "LLM2Vec-Sheared-Llama-supervised": 512,
729
- "LLM2Vec-Sheared-Llama-unsupervised": 512,
730
  "LaBSE": 512,
731
  "all-MiniLM-L12-v2": 512,
732
  "all-MiniLM-L6-v2": 512,
 
721
  "Cohere-embed-multilingual-light-v3.0": 512,
722
  "DanskBERT": 514,
723
  "LASER2": "N/A",
724
+ "LLM2Vec-Llama-supervised": 4096,
725
+ "LLM2Vec-Llama-unsupervised": 4096,
726
+ "LLM2Vec-Mistral-supervised": 32768,
727
+ "LLM2Vec-Mistral-unsupervised": 32768,
728
+ "LLM2Vec-Sheared-Llama-supervised": 4096,
729
+ "LLM2Vec-Sheared-Llama-unsupervised": 4096,
730
  "LaBSE": 512,
731
  "all-MiniLM-L12-v2": 512,
732
  "all-MiniLM-L6-v2": 512,