giux78 commited on
Commit
2d41043
·
verified ·
1 Parent(s): a39cd44

Upload folder using huggingface_hub

Browse files
Almawave/Velvet-14B (0-shot)_12.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "model_dtype": "bfloat16",
4
+ "model_name": "Almawave/Velvet-14B (0-shot)",
5
+ "model_sha": "main"
6
+ },
7
+ "results": {
8
+ "TextualEntailment": {
9
+ "acc": 67.63
10
+ },
11
+ "TextualEntailment_best": {
12
+ "acc": 78.5
13
+ },
14
+ "Sentiment Analysis": {
15
+ "acc": 60.49
16
+ },
17
+ "Sentiment Analysis_best": {
18
+ "acc": 66.59
19
+ },
20
+ "Hate Speech": {
21
+ "acc": 55.93
22
+ },
23
+ "Hate Speech_best": {
24
+ "acc": 65.64
25
+ },
26
+ "Admission Test": {
27
+ "acc": 42.4
28
+ },
29
+ "Admission Test_best": {
30
+ "acc": 57.4
31
+ },
32
+ "Word in Context": {
33
+ "acc": 37.91
34
+ },
35
+ "Word in Context_best": {
36
+ "acc": 64.31
37
+ },
38
+ "FAQ": {
39
+ "acc": 45.47
40
+ },
41
+ "FAQ_best": {
42
+ "acc": 89.53
43
+ },
44
+ "Lexical Substitution": {
45
+ "acc": 0.07
46
+ },
47
+ "Lexical Substitution_best": {
48
+ "acc": 0.13
49
+ },
50
+ "Summarization": {
51
+ "acc": 31.09
52
+ },
53
+ "Summarization_best": {
54
+ "acc": 31.11
55
+ },
56
+ "NER": {
57
+ "acc": 31.48
58
+ },
59
+ "NER_best": {
60
+ "acc": 43.79
61
+ },
62
+ "REL": {
63
+ "acc": 6.95
64
+ },
65
+ "REL_best": {
66
+ "acc": 9.98
67
+ }
68
+ },
69
+ "versions": null
70
+ }
Almawave/Velvet-14B (5-shot)_13.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "model_dtype": "bfloat16",
4
+ "model_name": "Almawave/Velvet-14B (5-shot)",
5
+ "model_sha": "main"
6
+ },
7
+ "results": {
8
+ "TextualEntailment": {
9
+ "acc": 74.46
10
+ },
11
+ "TextualEntailment_best": {
12
+ "acc": 76.75
13
+ },
14
+ "Sentiment Analysis": {
15
+ "acc": 67.99
16
+ },
17
+ "Sentiment Analysis_best": {
18
+ "acc": 68.93
19
+ },
20
+ "Hate Speech": {
21
+ "acc": 66.35
22
+ },
23
+ "Hate Speech_best": {
24
+ "acc": 68.97
25
+ },
26
+ "Admission Test": {
27
+ "acc": 49.73
28
+ },
29
+ "Admission Test_best": {
30
+ "acc": 61.2
31
+ },
32
+ "Word in Context": {
33
+ "acc": 15.41
34
+ },
35
+ "Word in Context_best": {
36
+ "acc": 37.69
37
+ },
38
+ "FAQ": {
39
+ "acc": 41.73
40
+ },
41
+ "FAQ_best": {
42
+ "acc": 80.55
43
+ },
44
+ "Lexical Substitution": {
45
+ "acc": 9.21
46
+ },
47
+ "Lexical Substitution_best": {
48
+ "acc": 9.45
49
+ },
50
+ "Summarization": {
51
+ "acc": 34.26
52
+ },
53
+ "Summarization_best": {
54
+ "acc": 34.88
55
+ },
56
+ "NER": {
57
+ "acc": 48.95
58
+ },
59
+ "NER_best": {
60
+ "acc": 48.73
61
+ },
62
+ "REL": {
63
+ "acc": 13.07
64
+ },
65
+ "REL_best": {
66
+ "acc": 15.1
67
+ }
68
+ },
69
+ "versions": null
70
+ }
CohereForAI/aya-expanse-8b_10.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "model_dtype": "bfloat16",
4
+ "model_name": "CohereForAI/aya-expanse-8b",
5
+ "model_sha": "main"
6
+ },
7
+ "results": {
8
+ "TextualEntailment": {
9
+ "acc": 60.33
10
+ },
11
+ "TextualEntailment_best": {
12
+ "acc": 75.0
13
+ },
14
+ "Sentiment Analysis": {
15
+ "acc": 60.16
16
+ },
17
+ "Sentiment Analysis_best": {
18
+ "acc": 70.62
19
+ },
20
+ "Hate Speech": {
21
+ "acc": 60.48
22
+ },
23
+ "Hate Speech_best": {
24
+ "acc": 62.97
25
+ },
26
+ "Admission Test": {
27
+ "acc": 38.53
28
+ },
29
+ "Admission Test_best": {
30
+ "acc": 57.8
31
+ },
32
+ "Word in Context": {
33
+ "acc": 45.95
34
+ },
35
+ "Word in Context_best": {
36
+ "acc": 66.19
37
+ },
38
+ "FAQ": {
39
+ "acc": 38.03
40
+ },
41
+ "FAQ_best": {
42
+ "acc": 66.33
43
+ },
44
+ "Lexical Substitution": {
45
+ "acc": 11.62
46
+ },
47
+ "Lexical Substitution_best": {
48
+ "acc": 15.92
49
+ },
50
+ "Summarization": {
51
+ "acc": 18.85
52
+ },
53
+ "Summarization_best": {
54
+ "acc": 19.21
55
+ },
56
+ "NER": {
57
+ "acc": 34.71
58
+ },
59
+ "NER_best": {
60
+ "acc": 39.0
61
+ },
62
+ "REL": {
63
+ "acc": 7.52
64
+ },
65
+ "REL_best": {
66
+ "acc": 9.57
67
+ }
68
+ },
69
+ "versions": null
70
+ }
FairMind/Llama-3-8B-4bit-UltraChat-Ita_8.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "model_dtype": "bfloat16",
4
+ "model_name": "FairMind/Llama-3-8B-4bit-UltraChat-Ita",
5
+ "model_sha": "main"
6
+ },
7
+ "results": {
8
+ "TextualEntailment": {
9
+ "acc": 58.08
10
+ },
11
+ "TextualEntailment_best": {
12
+ "acc": 64.25
13
+ },
14
+ "Sentiment Analysis": {
15
+ "acc": 46.01
16
+ },
17
+ "Sentiment Analysis_best": {
18
+ "acc": 67.3
19
+ },
20
+ "Hate Speech": {
21
+ "acc": 48.72
22
+ },
23
+ "Hate Speech_best": {
24
+ "acc": 63.38
25
+ },
26
+ "Admission Test": {
27
+ "acc": 33.37
28
+ },
29
+ "Admission Test_best": {
30
+ "acc": 49.2
31
+ },
32
+ "Word in Context": {
33
+ "acc": 66.03
34
+ },
35
+ "Word in Context_best": {
36
+ "acc": 66.58
37
+ },
38
+ "FAQ": {
39
+ "acc": 34.54
40
+ },
41
+ "FAQ_best": {
42
+ "acc": 55.61
43
+ },
44
+ "Lexical Substitution": {
45
+ "acc": 0.0
46
+ },
47
+ "Lexical Substitution_best": {
48
+ "acc": 0.0
49
+ },
50
+ "Summarization": {
51
+ "acc": 23.6
52
+ },
53
+ "Summarization_best": {
54
+ "acc": 24.35
55
+ },
56
+ "NER": {
57
+ "acc": 27.44
58
+ },
59
+ "NER_best": {
60
+ "acc": 31.89
61
+ },
62
+ "REL": {
63
+ "acc": 2.62
64
+ },
65
+ "REL_best": {
66
+ "acc": 4.27
67
+ }
68
+ },
69
+ "versions": null
70
+ }
MoxoffSpA/Volare_7.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "model_dtype": "bfloat16",
4
+ "model_name": "MoxoffSpA/Volare",
5
+ "model_sha": "main"
6
+ },
7
+ "results": {
8
+ "TextualEntailment": {
9
+ "acc": 51.25
10
+ },
11
+ "TextualEntailment_best": {
12
+ "acc": 55.5
13
+ },
14
+ "Sentiment Analysis": {
15
+ "acc": 44.15
16
+ },
17
+ "Sentiment Analysis_best": {
18
+ "acc": 58.49
19
+ },
20
+ "Hate Speech": {
21
+ "acc": 43.57
22
+ },
23
+ "Hate Speech_best": {
24
+ "acc": 62.7
25
+ },
26
+ "Admission Test": {
27
+ "acc": 22.77
28
+ },
29
+ "Admission Test_best": {
30
+ "acc": 27.4
31
+ },
32
+ "Word in Context": {
33
+ "acc": 27.8
34
+ },
35
+ "Word in Context_best": {
36
+ "acc": 66.4
37
+ },
38
+ "FAQ": {
39
+ "acc": 27.35
40
+ },
41
+ "FAQ_best": {
42
+ "acc": 28.18
43
+ },
44
+ "Lexical Substitution": {
45
+ "acc": 0.02
46
+ },
47
+ "Lexical Substitution_best": {
48
+ "acc": 0.04
49
+ },
50
+ "Summarization": {
51
+ "acc": 23.15
52
+ },
53
+ "Summarization_best": {
54
+ "acc": 23.32
55
+ },
56
+ "NER": {
57
+ "acc": 21.26
58
+ },
59
+ "NER_best": {
60
+ "acc": 21.57
61
+ },
62
+ "REL": {
63
+ "acc": 15.95
64
+ },
65
+ "REL_best": {
66
+ "acc": 20.37
67
+ }
68
+ },
69
+ "versions": null
70
+ }
arcee-ai/Llama-3.1-SuperNova-Lite_9.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "model_dtype": "bfloat16",
4
+ "model_name": "arcee-ai/Llama-3.1-SuperNova-Lite",
5
+ "model_sha": "main"
6
+ },
7
+ "results": {
8
+ "TextualEntailment": {
9
+ "acc": 68.54
10
+ },
11
+ "TextualEntailment_best": {
12
+ "acc": 74.5
13
+ },
14
+ "Sentiment Analysis": {
15
+ "acc": 63.8
16
+ },
17
+ "Sentiment Analysis_best": {
18
+ "acc": 77.12
19
+ },
20
+ "Hate Speech": {
21
+ "acc": 52.21
22
+ },
23
+ "Hate Speech_best": {
24
+ "acc": 63.8
25
+ },
26
+ "Admission Test": {
27
+ "acc": 43.63
28
+ },
29
+ "Admission Test_best": {
30
+ "acc": 68.6
31
+ },
32
+ "Word in Context": {
33
+ "acc": 42.88
34
+ },
35
+ "Word in Context_best": {
36
+ "acc": 66.58
37
+ },
38
+ "FAQ": {
39
+ "acc": 51.16
40
+ },
41
+ "FAQ_best": {
42
+ "acc": 90.52
43
+ },
44
+ "Lexical Substitution": {
45
+ "acc": 22.05
46
+ },
47
+ "Lexical Substitution_best": {
48
+ "acc": 24.29
49
+ },
50
+ "Summarization": {
51
+ "acc": 22.74
52
+ },
53
+ "Summarization_best": {
54
+ "acc": 22.78
55
+ },
56
+ "NER": {
57
+ "acc": 17.72
58
+ },
59
+ "NER_best": {
60
+ "acc": 20.21
61
+ },
62
+ "REL": {
63
+ "acc": 17.82
64
+ },
65
+ "REL_best": {
66
+ "acc": 20.06
67
+ }
68
+ },
69
+ "versions": null
70
+ }
iGeniusAI/Italia-9B-Instruct-v0.1_3.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "model_dtype": "bfloat16",
4
+ "model_name": "iGeniusAI/Italia-9B-Instruct-v0.1",
5
+ "model_sha": "main"
6
+ },
7
+ "results": {
8
+ "TextualEntailment": {
9
+ "acc": 59.33
10
+ },
11
+ "TextualEntailment_best": {
12
+ "acc": 70.25
13
+ },
14
+ "Sentiment Analysis": {
15
+ "acc": 48.41
16
+ },
17
+ "Sentiment Analysis_best": {
18
+ "acc": 59.26
19
+ },
20
+ "Hate Speech": {
21
+ "acc": 58.9
22
+ },
23
+ "Hate Speech_best": {
24
+ "acc": 63.58
25
+ },
26
+ "Admission Test": {
27
+ "acc": 26.13
28
+ },
29
+ "Admission Test_best": {
30
+ "acc": 30.4
31
+ },
32
+ "Word in Context": {
33
+ "acc": 49.92
34
+ },
35
+ "Word in Context_best": {
36
+ "acc": 66.67
37
+ },
38
+ "FAQ": {
39
+ "acc": 31.26
40
+ },
41
+ "FAQ_best": {
42
+ "acc": 38.9
43
+ },
44
+ "Lexical Substitution": {
45
+ "acc": 0.0
46
+ },
47
+ "Lexical Substitution_best": {
48
+ "acc": 0.0
49
+ },
50
+ "Summarization": {
51
+ "acc": 0.0
52
+ },
53
+ "Summarization_best": {
54
+ "acc": 0.01
55
+ },
56
+ "NER": {
57
+ "acc": 15.07
58
+ },
59
+ "NER_best": {
60
+ "acc": 20.82
61
+ },
62
+ "REL": {
63
+ "acc": 0.0
64
+ },
65
+ "REL_best": {
66
+ "acc": 0.0
67
+ }
68
+ },
69
+ "versions": null
70
+ }
ibm-granite/granite-3.1-8b-instruct_11.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "model_dtype": "bfloat16",
4
+ "model_name": "ibm-granite/granite-3.1-8b-instruct",
5
+ "model_sha": "main"
6
+ },
7
+ "results": {
8
+ "TextualEntailment": {
9
+ "acc": 51.08
10
+ },
11
+ "TextualEntailment_best": {
12
+ "acc": 67.0
13
+ },
14
+ "Sentiment Analysis": {
15
+ "acc": 47.82
16
+ },
17
+ "Sentiment Analysis_best": {
18
+ "acc": 69.45
19
+ },
20
+ "Hate Speech": {
21
+ "acc": 39.51
22
+ },
23
+ "Hate Speech_best": {
24
+ "acc": 62.64
25
+ },
26
+ "Admission Test": {
27
+ "acc": 41.97
28
+ },
29
+ "Admission Test_best": {
30
+ "acc": 59.4
31
+ },
32
+ "Word in Context": {
33
+ "acc": 42.0
34
+ },
35
+ "Word in Context_best": {
36
+ "acc": 66.85
37
+ },
38
+ "FAQ": {
39
+ "acc": 51.29
40
+ },
41
+ "FAQ_best": {
42
+ "acc": 91.52
43
+ },
44
+ "Lexical Substitution": {
45
+ "acc": 0.09
46
+ },
47
+ "Lexical Substitution_best": {
48
+ "acc": 0.17
49
+ },
50
+ "Summarization": {
51
+ "acc": 30.04
52
+ },
53
+ "Summarization_best": {
54
+ "acc": 30.39
55
+ },
56
+ "NER": {
57
+ "acc": 32.05
58
+ },
59
+ "NER_best": {
60
+ "acc": 37.13
61
+ },
62
+ "REL": {
63
+ "acc": 9.18
64
+ },
65
+ "REL_best": {
66
+ "acc": 11.47
67
+ }
68
+ },
69
+ "versions": null
70
+ }
microsoft/Phi-3.5-mini-instruct_1.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "model_dtype": "bfloat16",
4
+ "model_name": "microsoft/Phi-3.5-mini-instruct",
5
+ "model_sha": "main"
6
+ },
7
+ "results": {
8
+ "TextualEntailment": {
9
+ "acc": 70.04
10
+ },
11
+ "TextualEntailment_best": {
12
+ "acc": 81.5
13
+ },
14
+ "Sentiment Analysis": {
15
+ "acc": 43.73
16
+ },
17
+ "Sentiment Analysis_best": {
18
+ "acc": 70.59
19
+ },
20
+ "Hate Speech": {
21
+ "acc": 64.49
22
+ },
23
+ "Hate Speech_best": {
24
+ "acc": 69.17
25
+ },
26
+ "Admission Test": {
27
+ "acc": 40.67
28
+ },
29
+ "Admission Test_best": {
30
+ "acc": 62.8
31
+ },
32
+ "Word in Context": {
33
+ "acc": 56.97
34
+ },
35
+ "Word in Context_best": {
36
+ "acc": 67.41
37
+ },
38
+ "FAQ": {
39
+ "acc": 44.22
40
+ },
41
+ "FAQ_best": {
42
+ "acc": 79.05
43
+ },
44
+ "Lexical Substitution": {
45
+ "acc": 18.29
46
+ },
47
+ "Lexical Substitution_best": {
48
+ "acc": 20.94
49
+ },
50
+ "Summarization": {
51
+ "acc": 22.7
52
+ },
53
+ "Summarization_best": {
54
+ "acc": 23.4
55
+ },
56
+ "NER": {
57
+ "acc": 60.69
58
+ },
59
+ "NER_best": {
60
+ "acc": 61.35
61
+ },
62
+ "REL": {
63
+ "acc": 14.4
64
+ },
65
+ "REL_best": {
66
+ "acc": 20.61
67
+ }
68
+ },
69
+ "versions": null
70
+ }
mistralai/Mistral-7B-Instruct-v0.3_5.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "model_dtype": "bfloat16",
4
+ "model_name": "mistralai/Mistral-7B-Instruct-v0.3",
5
+ "model_sha": "main"
6
+ },
7
+ "results": {
8
+ "TextualEntailment": {
9
+ "acc": 60.83
10
+ },
11
+ "TextualEntailment_best": {
12
+ "acc": 71.25
13
+ },
14
+ "Sentiment Analysis": {
15
+ "acc": 58.71
16
+ },
17
+ "Sentiment Analysis_best": {
18
+ "acc": 66.14
19
+ },
20
+ "Hate Speech": {
21
+ "acc": 55.86
22
+ },
23
+ "Hate Speech_best": {
24
+ "acc": 65.39
25
+ },
26
+ "Admission Test": {
27
+ "acc": 39.23
28
+ },
29
+ "Admission Test_best": {
30
+ "acc": 54.0
31
+ },
32
+ "Word in Context": {
33
+ "acc": 61.07
34
+ },
35
+ "Word in Context_best": {
36
+ "acc": 66.49
37
+ },
38
+ "FAQ": {
39
+ "acc": 47.63
40
+ },
41
+ "FAQ_best": {
42
+ "acc": 84.79
43
+ },
44
+ "Lexical Substitution": {
45
+ "acc": 6.47
46
+ },
47
+ "Lexical Substitution_best": {
48
+ "acc": 10.26
49
+ },
50
+ "Summarization": {
51
+ "acc": 27.28
52
+ },
53
+ "Summarization_best": {
54
+ "acc": 28.43
55
+ },
56
+ "NER": {
57
+ "acc": 19.27
58
+ },
59
+ "NER_best": {
60
+ "acc": 19.46
61
+ },
62
+ "REL": {
63
+ "acc": 20.29
64
+ },
65
+ "REL_best": {
66
+ "acc": 21.83
67
+ }
68
+ },
69
+ "versions": null
70
+ }
occiglot/occiglot-7b-it-en-instruct_2.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "model_dtype": "bfloat16",
4
+ "model_name": "occiglot/occiglot-7b-it-en-instruct",
5
+ "model_sha": "main"
6
+ },
7
+ "results": {
8
+ "TextualEntailment": {
9
+ "acc": 49.92
10
+ },
11
+ "TextualEntailment_best": {
12
+ "acc": 56.0
13
+ },
14
+ "Sentiment Analysis": {
15
+ "acc": 51.87
16
+ },
17
+ "Sentiment Analysis_best": {
18
+ "acc": 61.09
19
+ },
20
+ "Hate Speech": {
21
+ "acc": 48.49
22
+ },
23
+ "Hate Speech_best": {
24
+ "acc": 62.72
25
+ },
26
+ "Admission Test": {
27
+ "acc": 42.8
28
+ },
29
+ "Admission Test_best": {
30
+ "acc": 54.0
31
+ },
32
+ "Word in Context": {
33
+ "acc": 42.86
34
+ },
35
+ "Word in Context_best": {
36
+ "acc": 66.49
37
+ },
38
+ "FAQ": {
39
+ "acc": 42.89
40
+ },
41
+ "FAQ_best": {
42
+ "acc": 83.29
43
+ },
44
+ "Lexical Substitution": {
45
+ "acc": 1.55
46
+ },
47
+ "Lexical Substitution_best": {
48
+ "acc": 2.88
49
+ },
50
+ "Summarization": {
51
+ "acc": 25.5
52
+ },
53
+ "Summarization_best": {
54
+ "acc": 28.56
55
+ },
56
+ "NER": {
57
+ "acc": 10.53
58
+ },
59
+ "NER_best": {
60
+ "acc": 14.15
61
+ },
62
+ "REL": {
63
+ "acc": 32.03
64
+ },
65
+ "REL_best": {
66
+ "acc": 32.45
67
+ }
68
+ },
69
+ "versions": null
70
+ }
sapienzanlp/Minerva-7B-instruct-v1.0_4.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "model_dtype": "bfloat16",
4
+ "model_name": "sapienzanlp/Minerva-7B-instruct-v1.0",
5
+ "model_sha": "main"
6
+ },
7
+ "results": {
8
+ "TextualEntailment": {
9
+ "acc": 54.13
10
+ },
11
+ "TextualEntailment_best": {
12
+ "acc": 56.5
13
+ },
14
+ "Sentiment Analysis": {
15
+ "acc": 44.52
16
+ },
17
+ "Sentiment Analysis_best": {
18
+ "acc": 59.46
19
+ },
20
+ "Hate Speech": {
21
+ "acc": 39.23
22
+ },
23
+ "Hate Speech_best": {
24
+ "acc": 60.48
25
+ },
26
+ "Admission Test": {
27
+ "acc": 28.87
28
+ },
29
+ "Admission Test_best": {
30
+ "acc": 34.0
31
+ },
32
+ "Word in Context": {
33
+ "acc": 53.41
34
+ },
35
+ "Word in Context_best": {
36
+ "acc": 66.04
37
+ },
38
+ "FAQ": {
39
+ "acc": 31.05
40
+ },
41
+ "FAQ_best": {
42
+ "acc": 37.66
43
+ },
44
+ "Lexical Substitution": {
45
+ "acc": 0.0
46
+ },
47
+ "Lexical Substitution_best": {
48
+ "acc": 0.0
49
+ },
50
+ "Summarization": {
51
+ "acc": 16.22
52
+ },
53
+ "Summarization_best": {
54
+ "acc": 16.36
55
+ },
56
+ "NER": {
57
+ "acc": 18.31
58
+ },
59
+ "NER_best": {
60
+ "acc": 19.31
61
+ },
62
+ "REL": {
63
+ "acc": 17.75
64
+ },
65
+ "REL_best": {
66
+ "acc": 21.38
67
+ }
68
+ },
69
+ "versions": null
70
+ }
swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA_6.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "model_dtype": "bfloat16",
4
+ "model_name": "swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA",
5
+ "model_sha": "main"
6
+ },
7
+ "results": {
8
+ "TextualEntailment": {
9
+ "acc": 58.92
10
+ },
11
+ "TextualEntailment_best": {
12
+ "acc": 69.1
13
+ },
14
+ "Sentiment Analysis": {
15
+ "acc": 60.95
16
+ },
17
+ "Sentiment Analysis_best": {
18
+ "acc": 72.04
19
+ },
20
+ "Hate Speech": {
21
+ "acc": 39.59
22
+ },
23
+ "Hate Speech_best": {
24
+ "acc": 66.32
25
+ },
26
+ "Admission Test": {
27
+ "acc": 40.73
28
+ },
29
+ "Admission Test_best": {
30
+ "acc": 62.2
31
+ },
32
+ "Word in Context": {
33
+ "acc": 52.6
34
+ },
35
+ "Word in Context_best": {
36
+ "acc": 66.57
37
+ },
38
+ "FAQ": {
39
+ "acc": 42.85
40
+ },
41
+ "FAQ_best": {
42
+ "acc": 71.82
43
+ },
44
+ "Lexical Substitution": {
45
+ "acc": 19.35
46
+ },
47
+ "Lexical Substitution_best": {
48
+ "acc": 19.37
49
+ },
50
+ "Summarization": {
51
+ "acc": 22.37
52
+ },
53
+ "Summarization_best": {
54
+ "acc": 22.71
55
+ },
56
+ "NER": {
57
+ "acc": 38.04
58
+ },
59
+ "NER_best": {
60
+ "acc": 47.57
61
+ },
62
+ "REL": {
63
+ "acc": 17.81
64
+ },
65
+ "REL_best": {
66
+ "acc": 21.61
67
+ }
68
+ },
69
+ "versions": null
70
+ }