This model is a fine-tuned version of BAAI/bge-m3 designed for the following use case:

financial sentiment and QA analysis

How to Use

This model can be easily integrated into your NLP pipeline for tasks such as text classification, sentiment analysis, entity recognition, and more. Here's a simple example to get you started:

from sentence_transformers import SentenceTransformer
from sentence_transformers.util import cos_sim

model = SentenceTransformer(
    'fine-tuned/FiQA2018-256-24-gpt-4o-2024-05-13-898550',
    trust_remote_code=True
)

embeddings = model.encode([
    'first text to embed',
    'second text to embed'
])
print(cos_sim(embeddings[0], embeddings[1]))
Downloads last month
5
Safetensors
Model size
568M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Datasets used to train fine-tuned/FiQA2018-256-24-gpt-4o-2024-05-13-898550

Spaces using fine-tuned/FiQA2018-256-24-gpt-4o-2024-05-13-898550 2