Cross-Encoder for STSB-Multi

This model was trained using SentenceTransformers Cross-Encoder class. The original model is dbmdz/bert-base-italian-uncased.

Training Data

This model was trained on the STS benchmark dataset, in particular the italian translation. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.

Usage and Performance

Pre-trained models can be used like this:

from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name')
scores = model.predict([('Sentence 1', 'Sentence 2'), ('Sentence 3', 'Sentence 4')])

The model will predict scores for the pairs ('Sentence 1', 'Sentence 2') and ('Sentence 3', 'Sentence 4').

You can use this model also without sentence_transformers and by just using Transformers AutoModel class

Downloads last month
488
Safetensors
Model size
110M params
Tensor type
I64
Β·
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Dataset used to train nickprock/cross-encoder-italian-bert-stsb

Spaces using nickprock/cross-encoder-italian-bert-stsb 2

Collection including nickprock/cross-encoder-italian-bert-stsb