This model is a fine-tuned version of BAAI/bge-small-en-v1.5 designed for the following use case:

cloud provider product performance and cost analysis

How to Use

This model can be easily integrated into your NLP pipeline for tasks such as text classification, sentiment analysis, entity recognition, and more. Here's a simple example to get you started:

from sentence_transformers import SentenceTransformer
from sentence_transformers.util import cos_sim

model = SentenceTransformer(
    'fine-tuned/BAAI_bge-small-en-v1_5-5252024-jzfp-webapp',
    trust_remote_code=True
)

embeddings = model.encode([
    'first text to embed',
    'second text to embed'
])
print(cos_sim(embeddings[0], embeddings[1]))
Downloads last month
13
Safetensors
Model size
33.4M params
Tensor type
F32
Β·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Datasets used to train fine-tuned/BAAI_bge-small-en-v1_5-5252024-jzfp-webapp

Spaces using fine-tuned/BAAI_bge-small-en-v1_5-5252024-jzfp-webapp 4