This model is a fine-tuned version of BAAI/bge-m3 designed for the following use case:

financial sentiment and QA analysis

How to Use

This model can be easily integrated into your NLP pipeline for tasks such as text classification, sentiment analysis, entity recognition, and more. Here's a simple example to get you started:

from sentence_transformers import SentenceTransformer
from sentence_transformers.util import cos_sim

model = SentenceTransformer(
    'fine-tuned/FiQA2018-256-24-gpt-4o-2024-05-13-898550',
    trust_remote_code=True
)

embeddings = model.encode([
    'first text to embed',
    'second text to embed'
])
print(cos_sim(embeddings[0], embeddings[1]))
Downloads last month
5
Safetensors
Model size
568M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Datasets used to train fine-tuned/FiQA2018-256-24-gpt-4o-2024-05-13-898550

Spaces using fine-tuned/FiQA2018-256-24-gpt-4o-2024-05-13-898550 2