dataset
stringlengths 4
115
| config
stringlengths 1
121
| split
stringlengths 1
228
| num_examples
int64 3
341M
| column_name
stringlengths 1
22.7k
| min
int64 0
1.81M
| max
int64 0
981M
| mean
float64 0
42.2M
| median
float64 0
24M
| std
float64 0
84.2M
| histogram
dict | partial
bool 2
classes |
---|---|---|---|---|---|---|---|---|---|---|---|
ceval/ceval-exam | college_economics | val | 55 | C | 1 | 44 | 10.89091 | 9 | 8.17556 | {
"bin_edges": [
1,
6,
11,
16,
21,
26,
31,
36,
41,
44
],
"hist": [
18,
13,
12,
4,
5,
2,
0,
0,
1
]
} | false |
ceval/ceval-exam | college_economics | val | 55 | D | 1 | 30 | 10.50909 | 8 | 7.20821 | {
"bin_edges": [
1,
4,
7,
10,
13,
16,
19,
22,
25,
28,
30
],
"hist": [
5,
16,
9,
7,
6,
5,
3,
0,
2,
2
]
} | false |
ceval/ceval-exam | college_economics | val | 55 | question | 10 | 168 | 39.29091 | 27 | 34.53108 | {
"bin_edges": [
10,
26,
42,
58,
74,
90,
106,
122,
138,
154,
168
],
"hist": [
25,
17,
3,
4,
1,
1,
1,
1,
0,
2
]
} | false |
ceval/ceval-exam | college_chemistry | test | 224 | A | 1 | 107 | 12.625 | 8 | 14.43782 | {
"bin_edges": [
1,
12,
23,
34,
45,
56,
67,
78,
89,
100,
107
],
"hist": [
139,
53,
21,
5,
3,
0,
0,
0,
1,
2
]
} | false |
ceval/ceval-exam | college_chemistry | test | 224 | B | 1 | 136 | 12.34375 | 8 | 13.86943 | {
"bin_edges": [
1,
15,
29,
43,
57,
71,
85,
99,
113,
127,
136
],
"hist": [
162,
50,
6,
3,
1,
0,
1,
0,
0,
1
]
} | false |
ceval/ceval-exam | college_chemistry | test | 224 | C | 1 | 96 | 12.29464 | 9 | 12.32056 | {
"bin_edges": [
1,
11,
21,
31,
41,
51,
61,
71,
81,
91,
96
],
"hist": [
128,
64,
17,
8,
2,
2,
1,
1,
0,
1
]
} | false |
ceval/ceval-exam | college_chemistry | test | 224 | D | 1 | 113 | 13.39286 | 8 | 14.52529 | {
"bin_edges": [
1,
13,
25,
37,
49,
61,
73,
85,
97,
109,
113
],
"hist": [
144,
46,
21,
8,
2,
0,
0,
2,
0,
1
]
} | false |
ceval/ceval-exam | college_chemistry | test | 224 | question | 7 | 618 | 51.97768 | 29 | 63.57965 | {
"bin_edges": [
7,
69,
131,
193,
255,
317,
379,
441,
503,
565,
618
],
"hist": [
183,
22,
11,
6,
1,
0,
0,
0,
0,
1
]
} | false |
ceval/ceval-exam | college_programming | test | 342 | A | 1 | 64 | 12.52632 | 8 | 12.25382 | {
"bin_edges": [
1,
8,
15,
22,
29,
36,
43,
50,
57,
64,
64
],
"hist": [
161,
77,
45,
20,
19,
7,
5,
5,
2,
1
]
} | false |
ceval/ceval-exam | college_programming | test | 342 | B | 1 | 80 | 12.84795 | 8 | 12.63613 | {
"bin_edges": [
1,
9,
17,
25,
33,
41,
49,
57,
65,
73,
80
],
"hist": [
172,
69,
47,
25,
18,
2,
7,
1,
0,
1
]
} | false |
ceval/ceval-exam | college_programming | test | 342 | C | 1 | 67 | 12.85673 | 9 | 11.85271 | {
"bin_edges": [
1,
8,
15,
22,
29,
36,
43,
50,
57,
64,
67
],
"hist": [
156,
63,
56,
32,
15,
12,
2,
4,
0,
2
]
} | false |
ceval/ceval-exam | college_programming | test | 342 | D | 1 | 70 | 12.83041 | 9 | 11.76731 | {
"bin_edges": [
1,
8,
15,
22,
29,
36,
43,
50,
57,
64,
70
],
"hist": [
150,
83,
42,
28,
23,
9,
1,
4,
1,
1
]
} | false |
ceval/ceval-exam | college_programming | test | 342 | question | 11 | 570 | 60.11111 | 35 | 73.69593 | {
"bin_edges": [
11,
67,
123,
179,
235,
291,
347,
403,
459,
515,
570
],
"hist": [
260,
41,
19,
10,
2,
5,
2,
1,
1,
1
]
} | false |
ceval/ceval-exam | college_programming | val | 37 | A | 1 | 57 | 12.67568 | 7 | 14.44925 | {
"bin_edges": [
1,
7,
13,
19,
25,
31,
37,
43,
49,
55,
57
],
"hist": [
16,
10,
3,
4,
1,
0,
0,
0,
1,
2
]
} | false |
ceval/ceval-exam | college_programming | val | 37 | B | 1 | 55 | 12.54054 | 9 | 13.56178 | {
"bin_edges": [
1,
7,
13,
19,
25,
31,
37,
43,
49,
55,
55
],
"hist": [
17,
8,
5,
3,
1,
0,
0,
0,
2,
1
]
} | false |
ceval/ceval-exam | college_programming | val | 37 | C | 1 | 55 | 12.02703 | 8 | 13.32393 | {
"bin_edges": [
1,
7,
13,
19,
25,
31,
37,
43,
49,
55,
55
],
"hist": [
16,
9,
5,
4,
0,
0,
0,
1,
1,
1
]
} | false |
ceval/ceval-exam | college_programming | val | 37 | D | 1 | 55 | 11.89189 | 7 | 12.75579 | {
"bin_edges": [
1,
7,
13,
19,
25,
31,
37,
43,
49,
55,
55
],
"hist": [
16,
8,
5,
4,
2,
0,
0,
0,
1,
1
]
} | false |
ceval/ceval-exam | college_programming | val | 37 | question | 12 | 376 | 67.89189 | 34 | 93.3416 | {
"bin_edges": [
12,
49,
86,
123,
160,
197,
234,
271,
308,
345,
376
],
"hist": [
24,
8,
1,
0,
0,
1,
0,
0,
1,
2
]
} | false |
ceval/ceval-exam | college_physics | test | 176 | A | 1 | 76 | 13.80682 | 10 | 13.73201 | {
"bin_edges": [
1,
9,
17,
25,
33,
41,
49,
57,
65,
73,
76
],
"hist": [
77,
56,
16,
12,
5,
2,
1,
4,
2,
1
]
} | false |
ceval/ceval-exam | college_physics | test | 176 | B | 1 | 116 | 15.14205 | 11 | 15.8279 | {
"bin_edges": [
1,
13,
25,
37,
49,
61,
73,
85,
97,
109,
116
],
"hist": [
106,
43,
14,
3,
5,
3,
1,
0,
0,
1
]
} | false |
ceval/ceval-exam | college_physics | test | 176 | C | 1 | 83 | 15.96023 | 11 | 15.7289 | {
"bin_edges": [
1,
10,
19,
28,
37,
46,
55,
64,
73,
82,
83
],
"hist": [
81,
48,
18,
15,
3,
4,
0,
4,
2,
1
]
} | false |
ceval/ceval-exam | college_physics | test | 176 | D | 1 | 93 | 15.01705 | 10 | 14.36602 | {
"bin_edges": [
1,
11,
21,
31,
41,
51,
61,
71,
81,
91,
93
],
"hist": [
93,
47,
12,
13,
4,
3,
2,
1,
0,
1
]
} | false |
ceval/ceval-exam | college_physics | test | 176 | question | 13 | 219 | 68.11932 | 65 | 32.88582 | {
"bin_edges": [
13,
34,
55,
76,
97,
118,
139,
160,
181,
202,
219
],
"hist": [
24,
44,
48,
27,
22,
6,
3,
0,
1,
1
]
} | false |
ceval/ceval-exam | computer_network | test | 171 | A | 1 | 36 | 6.74854 | 5 | 6.31254 | {
"bin_edges": [
1,
5,
9,
13,
17,
21,
25,
29,
33,
36
],
"hist": [
81,
49,
20,
9,
4,
3,
1,
3,
1
]
} | false |
ceval/ceval-exam | computer_network | test | 171 | B | 1 | 37 | 7.17544 | 4 | 7.0747 | {
"bin_edges": [
1,
5,
9,
13,
17,
21,
25,
29,
33,
37,
37
],
"hist": [
87,
38,
19,
10,
8,
3,
1,
2,
2,
1
]
} | false |
ceval/ceval-exam | computer_network | test | 171 | C | 1 | 36 | 6.9883 | 5 | 6.16726 | {
"bin_edges": [
1,
5,
9,
13,
17,
21,
25,
29,
33,
36
],
"hist": [
83,
37,
25,
14,
6,
3,
0,
0,
3
]
} | false |
ceval/ceval-exam | computer_network | test | 171 | D | 1 | 36 | 7.36257 | 5 | 6.28421 | {
"bin_edges": [
1,
5,
9,
13,
17,
21,
25,
29,
33,
36
],
"hist": [
78,
42,
20,
17,
7,
4,
0,
0,
3
]
} | false |
ceval/ceval-exam | computer_network | test | 171 | question | 9 | 163 | 42.72515 | 31 | 30.39738 | {
"bin_edges": [
9,
25,
41,
57,
73,
89,
105,
121,
137,
153,
163
],
"hist": [
48,
63,
23,
10,
11,
5,
4,
4,
2,
1
]
} | false |
ceval/ceval-exam | computer_architecture | test | 193 | A | 1 | 35 | 8.93264 | 6 | 7.06301 | {
"bin_edges": [
1,
5,
9,
13,
17,
21,
25,
29,
33,
35
],
"hist": [
67,
51,
28,
19,
10,
10,
2,
3,
3
]
} | false |
ceval/ceval-exam | computer_architecture | test | 193 | B | 1 | 34 | 8.86528 | 6 | 6.94941 | {
"bin_edges": [
1,
5,
9,
13,
17,
21,
25,
29,
33,
34
],
"hist": [
71,
48,
24,
21,
12,
10,
1,
4,
2
]
} | false |
ceval/ceval-exam | computer_architecture | test | 193 | C | 1 | 35 | 9.3057 | 6 | 7.41724 | {
"bin_edges": [
1,
5,
9,
13,
17,
21,
25,
29,
33,
35
],
"hist": [
64,
54,
29,
11,
14,
12,
3,
4,
2
]
} | false |
ceval/ceval-exam | computer_architecture | test | 193 | D | 1 | 32 | 8.94301 | 6 | 6.82 | {
"bin_edges": [
1,
5,
9,
13,
17,
21,
25,
29,
32
],
"hist": [
65,
57,
22,
15,
17,
11,
4,
2
]
} | false |
ceval/ceval-exam | computer_architecture | test | 193 | question | 9 | 127 | 32.72021 | 26 | 20.25191 | {
"bin_edges": [
9,
21,
33,
45,
57,
69,
81,
93,
105,
117,
127
],
"hist": [
49,
84,
29,
9,
10,
4,
3,
1,
2,
2
]
} | false |
ceval/ceval-exam | education_science | test | 270 | A | 2 | 52 | 6.54815 | 4.5 | 5.63607 | {
"bin_edges": [
2,
8,
14,
20,
26,
32,
38,
44,
50,
52
],
"hist": [
207,
36,
18,
7,
0,
0,
1,
0,
1
]
} | false |
ceval/ceval-exam | education_science | test | 270 | B | 2 | 38 | 6.68519 | 5 | 5.37039 | {
"bin_edges": [
2,
6,
10,
14,
18,
22,
26,
30,
34,
38,
38
],
"hist": [
160,
60,
25,
8,
10,
2,
3,
1,
0,
1
]
} | false |
ceval/ceval-exam | education_science | test | 270 | C | 2 | 38 | 6.93333 | 5 | 5.86743 | {
"bin_edges": [
2,
6,
10,
14,
18,
22,
26,
30,
34,
38,
38
],
"hist": [
163,
50,
29,
9,
11,
4,
0,
0,
3,
1
]
} | false |
ceval/ceval-exam | education_science | test | 270 | D | 2 | 38 | 6.95185 | 4.5 | 6.16814 | {
"bin_edges": [
2,
6,
10,
14,
18,
22,
26,
30,
34,
38,
38
],
"hist": [
162,
57,
22,
12,
5,
5,
3,
0,
3,
1
]
} | false |
ceval/ceval-exam | education_science | test | 270 | question | 8 | 267 | 34.2037 | 28 | 23.82946 | {
"bin_edges": [
8,
34,
60,
86,
112,
138,
164,
190,
216,
242,
267
],
"hist": [
168,
80,
16,
2,
2,
0,
1,
0,
0,
1
]
} | false |
ceval/ceval-exam | discrete_mathematics | test | 153 | A | 1 | 112 | 19.80392 | 9 | 23.661 | {
"bin_edges": [
1,
13,
25,
37,
49,
61,
73,
85,
97,
109,
112
],
"hist": [
88,
19,
21,
8,
4,
6,
2,
1,
3,
1
]
} | false |
ceval/ceval-exam | discrete_mathematics | test | 153 | B | 1 | 126 | 21.30065 | 12 | 26.00306 | {
"bin_edges": [
1,
14,
27,
40,
53,
66,
79,
92,
105,
118,
126
],
"hist": [
83,
29,
16,
10,
3,
5,
2,
0,
2,
3
]
} | false |
ceval/ceval-exam | discrete_mathematics | test | 153 | C | 1 | 147 | 22.04575 | 12 | 26.95399 | {
"bin_edges": [
1,
16,
31,
46,
61,
76,
91,
106,
121,
136,
147
],
"hist": [
90,
26,
14,
11,
3,
3,
2,
2,
1,
1
]
} | false |
ceval/ceval-exam | discrete_mathematics | test | 153 | D | 1 | 197 | 22.15686 | 12 | 28.69603 | {
"bin_edges": [
1,
21,
41,
61,
81,
101,
121,
141,
161,
181,
197
],
"hist": [
100,
28,
13,
3,
5,
2,
1,
0,
0,
1
]
} | false |
ceval/ceval-exam | discrete_mathematics | test | 153 | question | 10 | 291 | 47.29412 | 36 | 43.5281 | {
"bin_edges": [
10,
39,
68,
97,
126,
155,
184,
213,
242,
271,
291
],
"hist": [
85,
37,
17,
6,
3,
2,
0,
2,
0,
1
]
} | false |
ceval/ceval-exam | electrical_engineer | test | 339 | A | 1 | 61 | 7.64012 | 5 | 8.53536 | {
"bin_edges": [
1,
8,
15,
22,
29,
36,
43,
50,
57,
61
],
"hist": [
236,
68,
12,
9,
6,
4,
2,
1,
1
]
} | false |
ceval/ceval-exam | electrical_engineer | test | 339 | B | 1 | 55 | 7.71386 | 5 | 7.9932 | {
"bin_edges": [
1,
7,
13,
19,
25,
31,
37,
43,
49,
55,
55
],
"hist": [
212,
82,
19,
7,
7,
6,
4,
1,
0,
1
]
} | false |
ceval/ceval-exam | electrical_engineer | test | 339 | C | 1 | 61 | 7.75516 | 5 | 7.85382 | {
"bin_edges": [
1,
8,
15,
22,
29,
36,
43,
50,
57,
61
],
"hist": [
235,
63,
18,
11,
7,
3,
1,
0,
1
]
} | false |
ceval/ceval-exam | electrical_engineer | test | 339 | D | 1 | 59 | 7.9233 | 5 | 7.89989 | {
"bin_edges": [
1,
7,
13,
19,
25,
31,
37,
43,
49,
55,
59
],
"hist": [
199,
88,
26,
11,
5,
4,
2,
3,
0,
1
]
} | false |
ceval/ceval-exam | electrical_engineer | test | 339 | question | 13 | 137 | 44.87021 | 41 | 21.49008 | {
"bin_edges": [
13,
26,
39,
52,
65,
78,
91,
104,
117,
130,
137
],
"hist": [
54,
97,
89,
47,
23,
14,
9,
2,
3,
1
]
} | false |
ceval/ceval-exam | electrical_engineer | val | 37 | A | 1 | 25 | 6.2973 | 5 | 5.17937 | {
"bin_edges": [
1,
4,
7,
10,
13,
16,
19,
22,
25,
25
],
"hist": [
10,
16,
5,
2,
2,
0,
0,
1,
1
]
} | false |
ceval/ceval-exam | electrical_engineer | val | 37 | B | 1 | 48 | 7.27027 | 5 | 8.2045 | {
"bin_edges": [
1,
6,
11,
16,
21,
26,
31,
36,
41,
46,
48
],
"hist": [
22,
10,
2,
1,
1,
0,
0,
0,
0,
1
]
} | false |
ceval/ceval-exam | electrical_engineer | val | 37 | C | 1 | 52 | 8.18919 | 5 | 10.06876 | {
"bin_edges": [
1,
7,
13,
19,
25,
31,
37,
43,
49,
52
],
"hist": [
22,
11,
1,
0,
1,
0,
1,
0,
1
]
} | false |
ceval/ceval-exam | electrical_engineer | val | 37 | D | 1 | 53 | 8.32432 | 5 | 10.1326 | {
"bin_edges": [
1,
7,
13,
19,
25,
31,
37,
43,
49,
53
],
"hist": [
23,
10,
1,
0,
1,
0,
1,
0,
1
]
} | false |
ceval/ceval-exam | electrical_engineer | val | 37 | question | 14 | 92 | 44.72973 | 36 | 22.21692 | {
"bin_edges": [
14,
22,
30,
38,
46,
54,
62,
70,
78,
86,
92
],
"hist": [
3,
10,
6,
3,
2,
3,
4,
2,
2,
2
]
} | false |
ceval/ceval-exam | fire_engineer | test | 282 | A | 1 | 66 | 11.60638 | 5 | 13.62948 | {
"bin_edges": [
1,
8,
15,
22,
29,
36,
43,
50,
57,
64,
66
],
"hist": [
170,
28,
30,
22,
8,
10,
5,
6,
2,
1
]
} | false |
ceval/ceval-exam | fire_engineer | test | 282 | B | 1 | 65 | 11.53546 | 5 | 13.11408 | {
"bin_edges": [
1,
8,
15,
22,
29,
36,
43,
50,
57,
64,
65
],
"hist": [
170,
32,
23,
18,
13,
16,
6,
3,
0,
1
]
} | false |
ceval/ceval-exam | fire_engineer | test | 282 | C | 1 | 77 | 12.10284 | 5 | 14.68095 | {
"bin_edges": [
1,
9,
17,
25,
33,
41,
49,
57,
65,
73,
77
],
"hist": [
177,
36,
16,
22,
15,
5,
7,
1,
2,
1
]
} | false |
ceval/ceval-exam | fire_engineer | test | 282 | D | 1 | 66 | 11.6383 | 5 | 13.17783 | {
"bin_edges": [
1,
8,
15,
22,
29,
36,
43,
50,
57,
64,
66
],
"hist": [
169,
31,
28,
17,
15,
13,
4,
1,
3,
1
]
} | false |
ceval/ceval-exam | fire_engineer | test | 282 | question | 16 | 127 | 50.85816 | 46.5 | 23.89213 | {
"bin_edges": [
16,
28,
40,
52,
64,
76,
88,
100,
112,
124,
127
],
"hist": [
56,
52,
48,
42,
40,
22,
11,
7,
3,
1
]
} | false |
ceval/ceval-exam | fire_engineer | val | 31 | A | 1 | 71 | 12 | 6 | 14.30618 | {
"bin_edges": [
1,
9,
17,
25,
33,
41,
49,
57,
65,
71
],
"hist": [
17,
5,
7,
0,
1,
0,
0,
0,
1
]
} | false |
ceval/ceval-exam | fire_engineer | val | 31 | B | 1 | 66 | 15.96774 | 8 | 18.96397 | {
"bin_edges": [
1,
8,
15,
22,
29,
36,
43,
50,
57,
64,
66
],
"hist": [
15,
6,
2,
2,
0,
2,
2,
0,
0,
2
]
} | false |
ceval/ceval-exam | fire_engineer | val | 31 | C | 1 | 64 | 15.32258 | 7 | 17.66803 | {
"bin_edges": [
1,
8,
15,
22,
29,
36,
43,
50,
57,
64,
64
],
"hist": [
16,
4,
2,
4,
1,
1,
0,
1,
1,
1
]
} | false |
ceval/ceval-exam | fire_engineer | val | 31 | D | 1 | 74 | 15.09677 | 9 | 17.15101 | {
"bin_edges": [
1,
9,
17,
25,
33,
41,
49,
57,
65,
73,
74
],
"hist": [
15,
7,
1,
4,
2,
1,
0,
0,
0,
1
]
} | false |
ceval/ceval-exam | fire_engineer | val | 31 | question | 23 | 105 | 48.41935 | 41 | 24.033 | {
"bin_edges": [
23,
32,
41,
50,
59,
68,
77,
86,
95,
104,
105
],
"hist": [
10,
5,
6,
1,
2,
2,
2,
1,
1,
1
]
} | false |
ceval/ceval-exam | high_school_chinese | test | 178 | A | 2 | 110 | 40.69101 | 42 | 22.61043 | {
"bin_edges": [
2,
13,
24,
35,
46,
57,
68,
79,
90,
101,
110
],
"hist": [
28,
14,
19,
43,
29,
26,
12,
3,
3,
1
]
} | false |
ceval/ceval-exam | high_school_chinese | test | 178 | B | 2 | 93 | 39.5618 | 41 | 21.04231 | {
"bin_edges": [
2,
12,
22,
32,
42,
52,
62,
72,
82,
92,
93
],
"hist": [
28,
10,
16,
37,
30,
29,
22,
4,
1,
1
]
} | false |
ceval/ceval-exam | high_school_chinese | test | 178 | C | 2 | 89 | 39.27528 | 40 | 21.50568 | {
"bin_edges": [
2,
11,
20,
29,
38,
47,
56,
65,
74,
83,
89
],
"hist": [
26,
11,
16,
26,
19,
35,
24,
13,
5,
3
]
} | false |
ceval/ceval-exam | high_school_chinese | test | 178 | D | 2 | 93 | 39.92697 | 43.5 | 21.388 | {
"bin_edges": [
2,
12,
22,
32,
42,
52,
62,
72,
82,
92,
93
],
"hist": [
25,
13,
24,
23,
33,
36,
14,
7,
2,
1
]
} | false |
ceval/ceval-exam | high_school_chinese | test | 178 | question | 13 | 281 | 50.34831 | 19 | 56.91883 | {
"bin_edges": [
13,
40,
67,
94,
121,
148,
175,
202,
229,
256,
281
],
"hist": [
125,
6,
16,
9,
4,
9,
4,
2,
1,
2
]
} | false |
ceval/ceval-exam | high_school_biology | test | 175 | A | 2 | 49 | 13.22286 | 10 | 9.79746 | {
"bin_edges": [
2,
7,
12,
17,
22,
27,
32,
37,
42,
47,
49
],
"hist": [
60,
31,
27,
17,
20,
8,
9,
2,
0,
1
]
} | false |
ceval/ceval-exam | high_school_biology | test | 175 | B | 2 | 40 | 13.96 | 11 | 10.20613 | {
"bin_edges": [
2,
6,
10,
14,
18,
22,
26,
30,
34,
38,
40
],
"hist": [
49,
30,
20,
14,
18,
13,
15,
9,
4,
3
]
} | false |
ceval/ceval-exam | high_school_biology | test | 175 | C | 2 | 69 | 14.18286 | 11 | 11.18603 | {
"bin_edges": [
2,
9,
16,
23,
30,
37,
44,
51,
58,
65,
69
],
"hist": [
72,
36,
22,
27,
14,
2,
1,
0,
0,
1
]
} | false |
ceval/ceval-exam | high_school_biology | test | 175 | D | 1 | 41 | 14.06857 | 11 | 10.49279 | {
"bin_edges": [
1,
6,
11,
16,
21,
26,
31,
36,
41,
41
],
"hist": [
50,
37,
15,
27,
13,
20,
7,
4,
2
]
} | false |
ceval/ceval-exam | high_school_biology | test | 175 | question | 11 | 215 | 49.73714 | 34 | 39.91629 | {
"bin_edges": [
11,
32,
53,
74,
95,
116,
137,
158,
179,
200,
215
],
"hist": [
83,
38,
17,
15,
9,
4,
3,
3,
1,
2
]
} | false |
ceval/ceval-exam | environmental_impact_assessment_engineer | test | 281 | A | 1 | 55 | 9.879 | 6 | 9.6551 | {
"bin_edges": [
1,
7,
13,
19,
25,
31,
37,
43,
49,
55,
55
],
"hist": [
150,
62,
34,
13,
3,
7,
8,
3,
0,
1
]
} | false |
ceval/ceval-exam | environmental_impact_assessment_engineer | test | 281 | B | 1 | 95 | 10.77936 | 6 | 12.12028 | {
"bin_edges": [
1,
11,
21,
31,
41,
51,
61,
71,
81,
91,
95
],
"hist": [
192,
49,
20,
11,
4,
2,
1,
1,
0,
1
]
} | false |
ceval/ceval-exam | environmental_impact_assessment_engineer | test | 281 | C | 1 | 53 | 10.84342 | 7 | 10.41275 | {
"bin_edges": [
1,
7,
13,
19,
25,
31,
37,
43,
49,
53
],
"hist": [
130,
72,
34,
18,
8,
8,
4,
2,
5
]
} | false |
ceval/ceval-exam | environmental_impact_assessment_engineer | test | 281 | D | 1 | 124 | 11.29181 | 7 | 12.92313 | {
"bin_edges": [
1,
14,
27,
40,
53,
66,
79,
92,
105,
118,
124
],
"hist": [
204,
54,
14,
4,
2,
2,
0,
0,
0,
1
]
} | false |
ceval/ceval-exam | environmental_impact_assessment_engineer | test | 281 | question | 13 | 150 | 54.00356 | 51 | 24.50925 | {
"bin_edges": [
13,
27,
41,
55,
69,
83,
97,
111,
125,
139,
150
],
"hist": [
43,
40,
74,
51,
44,
13,
9,
4,
2,
1
]
} | false |
ceval/ceval-exam | environmental_impact_assessment_engineer | val | 31 | A | 1 | 30 | 8.32258 | 6 | 7.70016 | {
"bin_edges": [
1,
4,
7,
10,
13,
16,
19,
22,
25,
28,
30
],
"hist": [
8,
10,
6,
0,
3,
1,
0,
0,
1,
2
]
} | false |
ceval/ceval-exam | environmental_impact_assessment_engineer | val | 31 | B | 1 | 31 | 8.83871 | 7 | 7.54143 | {
"bin_edges": [
1,
5,
9,
13,
17,
21,
25,
29,
31
],
"hist": [
10,
8,
8,
2,
0,
1,
0,
2
]
} | false |
ceval/ceval-exam | environmental_impact_assessment_engineer | val | 31 | C | 1 | 31 | 9.3871 | 7 | 8.47615 | {
"bin_edges": [
1,
5,
9,
13,
17,
21,
25,
29,
31
],
"hist": [
12,
8,
3,
2,
3,
0,
1,
2
]
} | false |
ceval/ceval-exam | environmental_impact_assessment_engineer | val | 31 | D | 1 | 31 | 9.16129 | 6 | 8.43839 | {
"bin_edges": [
1,
5,
9,
13,
17,
21,
25,
29,
31
],
"hist": [
11,
9,
4,
2,
1,
1,
2,
1
]
} | false |
ceval/ceval-exam | environmental_impact_assessment_engineer | val | 31 | question | 23 | 108 | 58.25806 | 58 | 23.5074 | {
"bin_edges": [
23,
32,
41,
50,
59,
68,
77,
86,
95,
104,
108
],
"hist": [
4,
3,
4,
5,
8,
0,
2,
2,
2,
1
]
} | false |
ceval/ceval-exam | high_school_geography | test | 178 | A | 1 | 31 | 8.91011 | 7 | 6.43725 | {
"bin_edges": [
1,
5,
9,
13,
17,
21,
25,
29,
31
],
"hist": [
54,
48,
31,
22,
15,
2,
1,
5
]
} | false |
ceval/ceval-exam | high_school_geography | test | 178 | B | 2 | 38 | 9.58427 | 8 | 6.78118 | {
"bin_edges": [
2,
6,
10,
14,
18,
22,
26,
30,
34,
38,
38
],
"hist": [
65,
35,
32,
22,
12,
8,
2,
1,
0,
1
]
} | false |
ceval/ceval-exam | high_school_geography | test | 178 | C | 2 | 32 | 9.50562 | 7 | 6.84307 | {
"bin_edges": [
2,
6,
10,
14,
18,
22,
26,
30,
32
],
"hist": [
61,
45,
30,
15,
13,
10,
1,
3
]
} | false |
ceval/ceval-exam | high_school_geography | test | 178 | D | 2 | 36 | 9.9382 | 7 | 7.45166 | {
"bin_edges": [
2,
6,
10,
14,
18,
22,
26,
30,
34,
36
],
"hist": [
59,
47,
26,
18,
11,
9,
4,
2,
2
]
} | false |
ceval/ceval-exam | high_school_geography | test | 178 | question | 11 | 167 | 32.29213 | 23 | 29.01517 | {
"bin_edges": [
11,
27,
43,
59,
75,
91,
107,
123,
139,
155,
167
],
"hist": [
120,
35,
2,
4,
5,
5,
2,
0,
3,
2
]
} | false |
ceval/ceval-exam | high_school_history | test | 182 | A | 2 | 32 | 8.71429 | 9 | 4.49379 | {
"bin_edges": [
2,
6,
10,
14,
18,
22,
26,
30,
32
],
"hist": [
50,
55,
57,
15,
3,
0,
1,
1
]
} | false |
ceval/ceval-exam | high_school_history | test | 182 | B | 2 | 34 | 8.81319 | 8 | 4.73826 | {
"bin_edges": [
2,
6,
10,
14,
18,
22,
26,
30,
34,
34
],
"hist": [
47,
57,
54,
18,
4,
1,
0,
0,
1
]
} | false |
ceval/ceval-exam | high_school_history | test | 182 | C | 2 | 22 | 8.86813 | 9 | 4.36639 | {
"bin_edges": [
2,
5,
8,
11,
14,
17,
20,
22
],
"hist": [
38,
30,
56,
34,
15,
5,
4
]
} | false |
ceval/ceval-exam | high_school_history | test | 182 | D | 2 | 27 | 9.19231 | 9 | 4.74561 | {
"bin_edges": [
2,
5,
8,
11,
14,
17,
20,
23,
26,
27
],
"hist": [
38,
30,
45,
41,
15,
8,
3,
1,
1
]
} | false |
ceval/ceval-exam | high_school_history | test | 182 | question | 14 | 140 | 62.46154 | 60 | 33.81549 | {
"bin_edges": [
14,
27,
40,
53,
66,
79,
92,
105,
118,
131,
140
],
"hist": [
32,
31,
19,
17,
16,
24,
24,
8,
5,
6
]
} | false |
ceval/ceval-exam | high_school_chemistry | test | 172 | A | 1 | 62 | 14.15698 | 11 | 12.16475 | {
"bin_edges": [
1,
8,
15,
22,
29,
36,
43,
50,
57,
62
],
"hist": [
63,
40,
33,
19,
5,
3,
5,
2,
2
]
} | false |
ceval/ceval-exam | high_school_chemistry | test | 172 | B | 1 | 72 | 15.5 | 13 | 12.97524 | {
"bin_edges": [
1,
9,
17,
25,
33,
41,
49,
57,
65,
72
],
"hist": [
64,
44,
29,
14,
13,
5,
0,
1,
2
]
} | false |
ceval/ceval-exam | high_school_chemistry | test | 172 | C | 1 | 65 | 16.15698 | 13 | 13.44922 | {
"bin_edges": [
1,
8,
15,
22,
29,
36,
43,
50,
57,
64,
65
],
"hist": [
62,
34,
24,
24,
11,
10,
3,
1,
2,
1
]
} | false |
ceval/ceval-exam | high_school_chemistry | test | 172 | D | 1 | 110 | 16.68023 | 13 | 15.55549 | {
"bin_edges": [
1,
12,
23,
34,
45,
56,
67,
78,
89,
100,
110
],
"hist": [
81,
44,
31,
8,
3,
2,
2,
0,
0,
1
]
} | false |
ceval/ceval-exam | high_school_chemistry | test | 172 | question | 11 | 267 | 46.6686 | 28.5 | 42.76565 | {
"bin_edges": [
11,
37,
63,
89,
115,
141,
167,
193,
219,
245,
267
],
"hist": [
105,
26,
16,
11,
7,
3,
1,
0,
2,
1
]
} | false |
ceval/ceval-exam | high_school_mathematics | test | 166 | A | 1 | 60 | 11.83735 | 8 | 11.75884 | {
"bin_edges": [
1,
7,
13,
19,
25,
31,
37,
43,
49,
55,
60
],
"hist": [
72,
27,
27,
19,
9,
5,
3,
0,
2,
2
]
} | false |
ceval/ceval-exam | high_school_mathematics | test | 166 | B | 1 | 61 | 13.66867 | 10 | 12.73144 | {
"bin_edges": [
1,
8,
15,
22,
29,
36,
43,
50,
57,
61
],
"hist": [
71,
34,
25,
17,
8,
5,
1,
4,
1
]
} | false |