Edit model card

This model is a fine-tuned version of BAAI/bge-m3 designed for the following use case:

None

How to Use

This model can be easily integrated into your NLP pipeline for tasks such as text classification, sentiment analysis, entity recognition, and more. Here's a simple example to get you started:

from sentence_transformers import SentenceTransformer
from sentence_transformers.util import cos_sim

model = SentenceTransformer(
    'fine-tuned/SCIDOCS-512-192-gpt-4o-2024-05-13-34642434',
    trust_remote_code=True
)

embeddings = model.encode([
    'first text to embed',
    'second text to embed'
])
print(cos_sim(embeddings[0], embeddings[1]))
Downloads last month
6
Safetensors
Model size
568M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train fine-tuned/SCIDOCS-512-192-gpt-4o-2024-05-13-34642434

Space using fine-tuned/SCIDOCS-512-192-gpt-4o-2024-05-13-34642434 1