Datasets:
mteb
/

Modalities:
Text
Formats:
csv
Size:
< 1K
Libraries:
Datasets
Dask
nouamanetazi HF Staff commited on
Commit
49f48ec
·
1 Parent(s): f9e3310

Update bert-base-uncased_results.csv

Browse files
Files changed (1) hide show
  1. bert-base-uncased_results.csv +57 -57
bert-base-uncased_results.csv CHANGED
@@ -1,57 +1,57 @@
1
- dataset,metric,value
2
- AmazonCounterfactualClassification,accuracy,0.7425373134328358
3
- AmazonPolarityClassification,accuracy,0.7132945
4
- AmazonReviewsClassification,accuracy,0.33564
5
- Banking77Classification,accuracy,0.6340584415584415
6
- EmotionClassification,accuracy,0.3528
7
- ImdbClassification,accuracy,0.653456
8
- MassiveIntentClassification,accuracy,0.5988231338264962
9
- MassiveScenarioClassification,accuracy,0.6427706792199059
10
- MTOPDomainClassification,accuracy,0.8262653898768809
11
- MTOPIntentClassification,accuracy,0.6813725490196079
12
- ToxicConversationsClassification,accuracy,0.6999679999999999
13
- TweetSentimentExtractionClassification,accuracy,0.5180814940577249
14
- ArxivClusteringP2P,v_measure,0.3518932830729758
15
- ArxivClusteringS2S,v_measure,0.2750822951507033
16
- BiorxivClusteringP2P,v_measure,0.3012278640553615
17
- BiorxivClusteringS2S,v_measure,0.2476601053047804
18
- MedrxivClusteringP2P,v_measure,0.26087688306606044
19
- MedrxivClusteringS2S,v_measure,0.23604914608225602
20
- RedditClustering,v_measure,0.27241718557705
21
- RedditClusteringP2P,v_measure,0.43323854460464056
22
- StackExchangeClustering,v_measure,0.4358261870547655
23
- StackExchangeClusteringP2P,v_measure,0.2654851675795123
24
- TwentyNewsgroupsClustering,v_measure,0.23354321403189832
25
- SprintDuplicateQuestions,ap,0.36808703728970593
26
- TwitterSemEval2015,ap,0.5589788752976392
27
- TwitterURLCorpus,ap,0.7628732783437441
28
- AskUbuntuDupQuestions,map,0.4584088706528255
29
- MindSmallReranking,map,0.28366637355845425
30
- SciDocsRR,map,0.6493728987830145
31
- StackOverflowDupQuestions,map,0.34615478798860166
32
- ArguAna,ndcg_at_10,0.28294
33
- ClimateFEVER,ndcg_at_10,0.0541
34
- CQADupstackRetrieval,ndcg_at_10,0.05506583333333333
35
- DBPedia,ndcg_at_10,0.04132
36
- FEVER,ndcg_at_10,0.033
37
- FiQA2018,ndcg_at_10,0.02191
38
- HotpotQA,ndcg_at_10,0.0826
39
- MSMARCO,ndcg_at_10,0.06176
40
- NFCorpus,ndcg_at_10,0.04304
41
- NQ,ndcg_at_10,0.02615
42
- QuoraRetrieval,ndcg_at_10,0.61029
43
- SCIDOCS,ndcg_at_10,0.02815
44
- SciFact,ndcg_at_10,0.13339
45
- Touche2020,ndcg_at_10,0.00967
46
- TRECCOVID,ndcg_at_10,0.14745
47
- BIOSSES,cosine_spearman,0.5469823428818151
48
- SICK-R,cosine_spearman,0.5864506948179484
49
- STS12,cosine_spearman,0.30871788407575457
50
- STS13,cosine_spearman,0.5989485045425808
51
- STS14,cosine_spearman,0.4772791352844941
52
- STS15,cosine_spearman,0.6028567383446486
53
- STS16,cosine_spearman,0.6373272693604403
54
- STS17,cosine_spearman,0.6410023907260639
55
- STS22,cosine_spearman,0.563667996535454
56
- STSBenchmark,cosine_spearman,0.4729108172174081
57
- SummEval,cosine_spearman,0.2981716355664815
 
1
+ model,dataset,metric,value
2
+ bert-base-uncased,AmazonCounterfactualClassification,accuracy,0.7425373134328358
3
+ bert-base-uncased,AmazonPolarityClassification,accuracy,0.7132945
4
+ bert-base-uncased,AmazonReviewsClassification,accuracy,0.33564
5
+ bert-base-uncased,Banking77Classification,accuracy,0.6340584415584415
6
+ bert-base-uncased,EmotionClassification,accuracy,0.3528
7
+ bert-base-uncased,ImdbClassification,accuracy,0.653456
8
+ bert-base-uncased,MassiveIntentClassification,accuracy,0.5988231338264962
9
+ bert-base-uncased,MassiveScenarioClassification,accuracy,0.6427706792199059
10
+ bert-base-uncased,MTOPDomainClassification,accuracy,0.8262653898768809
11
+ bert-base-uncased,MTOPIntentClassification,accuracy,0.6813725490196079
12
+ bert-base-uncased,ToxicConversationsClassification,accuracy,0.6999679999999999
13
+ bert-base-uncased,TweetSentimentExtractionClassification,accuracy,0.5180814940577249
14
+ bert-base-uncased,ArxivClusteringP2P,v_measure,0.3518932830729758
15
+ bert-base-uncased,ArxivClusteringS2S,v_measure,0.2750822951507033
16
+ bert-base-uncased,BiorxivClusteringP2P,v_measure,0.3012278640553615
17
+ bert-base-uncased,BiorxivClusteringS2S,v_measure,0.2476601053047804
18
+ bert-base-uncased,MedrxivClusteringP2P,v_measure,0.26087688306606044
19
+ bert-base-uncased,MedrxivClusteringS2S,v_measure,0.23604914608225602
20
+ bert-base-uncased,RedditClustering,v_measure,0.27241718557705
21
+ bert-base-uncased,RedditClusteringP2P,v_measure,0.43323854460464056
22
+ bert-base-uncased,StackExchangeClustering,v_measure,0.4358261870547655
23
+ bert-base-uncased,StackExchangeClusteringP2P,v_measure,0.2654851675795123
24
+ bert-base-uncased,TwentyNewsgroupsClustering,v_measure,0.23354321403189832
25
+ bert-base-uncased,SprintDuplicateQuestions,ap,0.36808703728970593
26
+ bert-base-uncased,TwitterSemEval2015,ap,0.5589788752976392
27
+ bert-base-uncased,TwitterURLCorpus,ap,0.7628732783437441
28
+ bert-base-uncased,AskUbuntuDupQuestions,map,0.4584088706528255
29
+ bert-base-uncased,MindSmallReranking,map,0.28366637355845425
30
+ bert-base-uncased,SciDocsRR,map,0.6493728987830145
31
+ bert-base-uncased,StackOverflowDupQuestions,map,0.34615478798860166
32
+ bert-base-uncased,ArguAna,ndcg_at_10,0.28294
33
+ bert-base-uncased,ClimateFEVER,ndcg_at_10,0.0541
34
+ bert-base-uncased,CQADupstackRetrieval,ndcg_at_10,0.05506583333333333
35
+ bert-base-uncased,DBPedia,ndcg_at_10,0.04132
36
+ bert-base-uncased,FEVER,ndcg_at_10,0.033
37
+ bert-base-uncased,FiQA2018,ndcg_at_10,0.02191
38
+ bert-base-uncased,HotpotQA,ndcg_at_10,0.0826
39
+ bert-base-uncased,MSMARCO,ndcg_at_10,0.06176
40
+ bert-base-uncased,NFCorpus,ndcg_at_10,0.04304
41
+ bert-base-uncased,NQ,ndcg_at_10,0.02615
42
+ bert-base-uncased,QuoraRetrieval,ndcg_at_10,0.61029
43
+ bert-base-uncased,SCIDOCS,ndcg_at_10,0.02815
44
+ bert-base-uncased,SciFact,ndcg_at_10,0.13339
45
+ bert-base-uncased,Touche2020,ndcg_at_10,0.00967
46
+ bert-base-uncased,TRECCOVID,ndcg_at_10,0.14745
47
+ bert-base-uncased,BIOSSES,cosine_spearman,0.5469823428818151
48
+ bert-base-uncased,SICK-R,cosine_spearman,0.5864506948179484
49
+ bert-base-uncased,STS12,cosine_spearman,0.30871788407575457
50
+ bert-base-uncased,STS13,cosine_spearman,0.5989485045425808
51
+ bert-base-uncased,STS14,cosine_spearman,0.4772791352844941
52
+ bert-base-uncased,STS15,cosine_spearman,0.6028567383446486
53
+ bert-base-uncased,STS16,cosine_spearman,0.6373272693604403
54
+ bert-base-uncased,STS17,cosine_spearman,0.6410023907260639
55
+ bert-base-uncased,STS22,cosine_spearman,0.563667996535454
56
+ bert-base-uncased,STSBenchmark,cosine_spearman,0.4729108172174081
57
+ bert-base-uncased,SummEval,cosine_spearman,0.2981716355664815