bge-large-en-v1.5-sparse
Usage
This is the sparse ONNX variant of the bge-small-en-v1.5 embeddings model accelerated with Sparsify for quantization/pruning and DeepSparseSentenceTransformers for inference.
pip install -U deepsparse-nightly[sentence_transformers]
from deepsparse.sentence_transformers import DeepSparseSentenceTransformer
model = DeepSparseSentenceTransformer('neuralmagic/bge-large-en-v1.5-sparse', export=False)
# Our sentences we like to encode
sentences = ['This framework generates embeddings for each input sentence',
'Sentences are passed as a list of string.',
'The quick brown fox jumps over the lazy dog.']
# Sentences are encoded by calling model.encode()
embeddings = model.encode(sentences)
# Print the embeddings
for sentence, embedding in zip(sentences, embeddings):
print("Sentence:", sentence)
print("Embedding:", embedding.shape)
print("")
For general questions on these models and sparsification methods, reach out to the engineering team on our community Slack.
- Downloads last month
- 155
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Spaces using neuralmagic/bge-large-en-v1.5-sparse 2
Evaluation results
- cos_sim_pearson on MTEB BIOSSEStest set self-reported87.733
- cos_sim_spearman on MTEB BIOSSEStest set self-reported85.644
- euclidean_pearson on MTEB BIOSSEStest set self-reported86.069
- euclidean_spearman on MTEB BIOSSEStest set self-reported85.607
- manhattan_pearson on MTEB BIOSSEStest set self-reported85.691
- manhattan_spearman on MTEB BIOSSEStest set self-reported85.053
- cos_sim_pearson on MTEB SICK-Rtest set self-reported85.618
- cos_sim_spearman on MTEB SICK-Rtest set self-reported80.737
- euclidean_pearson on MTEB SICK-Rtest set self-reported83.937
- euclidean_spearman on MTEB SICK-Rtest set self-reported80.645