Dataset Viewer
Auto-converted to Parquet
model
string
quant method
string
truthfulqa_mc2 acc ± stderr
string
arc:challenge acc ± stderr
string
hellaswag acc
string
winogrande acc
string
average
float64
throughput (tok/s)
float64
peak process vram (GB)
float64
calibration/quantization time
string
throughput w/ torch.compile
float64
peak process vram w/ torch.compile
float64
througput w/ marlin
float64
peak process vram w/ marlin
float64
Llama 3.1 8B
baseline (bf16)
0.5456 ± 0.0150
0.5623 ± 0.0145
0.5973 ± 0.0049
0.7372 ± 0.0124
0.6106
38.82
16.77302
null
79.27
17.12954
null
null
Llama 3.1 8B
AWQ (4 bit)
0.5281 ± 0.0150
0.5589 ± .0145
.5869 ± .0049
0.7285 ± 0.0125
0.6006
43.09
8.11808
~10 minutes
null
null
null
null
Llama 3.1 8B
GPTQModel (4 bit)
null
null
null
null
null
36.83
6.31872
~23 minutes
null
null
37.84
6,318.72
Llama 3.1 8B
AutoGPTQ (4 bit)
0.5361 ± 0.0149
0.5589 ± 0.0145
0.5758 ± 0.0049
0.7103 ± 0.0127
0.595275
43.71
6.49278
~30 minutes
null
null
null
null
Llama 3.1 8B
bnb (nf4)
0.5446 ± 0.0149
0.5563 ± 0.0145
0.5769 ± 0.0049
0.7301 ± 0.0125
0.601975
24.35
6.44035
~1 minute
null
null
null
null
Llama 3.1 8B
optimum quanto (w4a16)
0.5162 ± 0.0150
0.5427 ± 0.0146
0.5856 ± 0.0049
0.7419 ± 0.0123
0.5966
31.22
6.57667
~30 seconds
null
null
null
null
Llama 3.1 8B
torchao (int4wo)
0.5166 ±0.0150
0.5418 ± 0.0146
0.5855 ± 0.0049
0.7395 ± 0.0123
0.59585
24.98
6.49907
~20 seconds
85.76
6.8493
null
null
Llama 3.1 8B
HQQ (4 bit)
0.5485 ± 0.0150
0.5495 ± 0.0145
0.5859 ± 0.0049
0.7285 ± 0.0125
0.6031
34.44
6.71718
null
null
null
null
null
Llama 3.1 8B
HIGGS (4 bit)
0.5214 ± 0.0149
0.5486 ± 0.0145
0.5836 ± 0.0049
0.7182 ± 0.0126
0.59295
28.35
6.81994
~5 minutes
null
null
null
null
Llama 3.1 8B
bnb (llm.int8())
0.5446 ± 0.0150
0.5623 ± 0.0145
0.5949 ± 0.0049
0.7324 ± 0.0124
0.60855
20.75
9.70772
~20 seconds
null
null
null
null
Llama 3.1 8B
HQQ (8 bit)
0.5441 ± 0.0150
0.5666 ± 0.0145
0.5979 ± 0.0049
0.7380 ± 0.0124
0.61165
9.07
10.6011
~80 seconds
null
null
null
null
Llama 3.1 8B
optimum quanto (int8wo)
0.5436 ± 0.0150
0.5640 ± 0.0145
0.5992 ± 0.0049
0.7372 ± 0.0124
0.611
15.59
9.81887
~20 seconds
16.01
10.07052
null
null
Llama 3.1 8B
torchao (int8wo)
0.5449 ± 0.0150
0.5640 ± 0.0145
0.5975 ± 0.0049
0.7380 ± 0.0124
0.6111
5.98
13.07155
~30 seconds
43.79
13.66714
null
null
Llama 3.1 8B
fbgemm (fp8)
0.5430 ± 0.0150
0.5580 ± 0.0145
0.5958 ± 0.0049
0.7411 ± 0.0123
0.609475
33.83
10.00551
~30 seconds
null
null
null
null
Llama 3.1 8B
compressed-tensors (fp8)
0.5398 ± 0.0151
0.5589 ± 0.0145
0.5950 ± 0.0049
0.7356 ± 0.0124
0.607325
null
null
null
null
null
null
null
Llama 3.1 8B
VPTQ (2 bit)
0.4543 ± 0.0149
0.4923 ± 0.0146
0.5258 ± 0.0050
0.6930 ± 0.0130
0.54135
32.35
5.28902
~2 hours
31.48
5.28692
null
null
Llama 3.1 8B
AQLM + PV (2 bit)
0.5036 ± 0.0148
0.5230 ± 0.0146
0.5628 ± 0.0050
0.6938 ± 0.0130
0.5708
22.28
4.84023
~1 day
27.27
4.85491
null
null
Llama 3.1 8B
GPTQModel (2 bit)
null
null
null
null
null
19.02
18.45284
~26 minutes
null
null
null
null
Llama 3.1 8B
AutoGPTQ (2 bit)
0.5127 ± 0.0150
0.1988 ± 0.0117
0.2665 ± 0.0044
0.4799 ± 0.0140
0.364475
6.25
11.02473
~26 minutes
null
null
null
null
Llama 3.1 70B
baseline (bf16)
0.6068 ± 0.0147
0.6732 ± 0.0137
0.6666 ± 0.0047
0.8248 ± 0.0107
0.69285
9.73
142.26869
null
10.1
142.81395
null
null
Llama 3.1 70B
AWQ (4 bit)
0.5706 ± 0.0150
0.6681 ± 0.0138
0.6598 ± 0.0047
0.8193 ± 0.0108
0.67945
15.74
43.75288
~1 hour
null
null
null
null
Llama 3.1 70B
GPTQModel (4 bit)
null
null
null
null
null
14.84
40.5757
null
null
null
15.28
40.5757
Llama 3.1 70B
AutoGPTQ (4 bit)
0.5937 ± 0.0147
0.6655 ± 0.0138
0.6568 ± 0.0047
0.8185 ± 0.0108
0.683625
0.46
42.40022
~2 hours
null
null
null
null
Llama 3.1 70B
bnb (nf4)
0.5939 ± 0.0148
0.6724 ± 0.0137
0.6592 ± 0.0047
0.8098 ± 0.0110
0.683825
11.27
44.62949
~2 minutes
null
null
null
null
Llama 3.1 70B
optimum quanto (w4a16)
0.4847 ± 0.0164
0.2082 ± 0.0119
0.2582 ± 0.0044
0.4878 ± 0.0140
0.359725
12.97
80.39013
~2 minutes
null
null
null
null
Llama 3.1 70B
torchao (int4wo)
0.4847 ± 0.0164
0.2108 ± 0.0119
0.2581 ± 0.0044
0.4980 ± 0.0141
0.3629
10.56
41.6054
~2 minutes
18.95
42.26181
null
null
Llama 3.1 70B
HQQ (4 bit)
0.5882 ± 0.0146
0.6706 ± 0.0137
0.6597 ± 0.0047
0.8035 ± 0.0112
0.6805
13.92
44.50366
~10 minutes
null
null
null
null
Llama 3.1 70B
HIGGS (4 bit)
0.4871 ± 0.0163
0.1971 ± 0.0116
0.2575 ± 0.0044
0.4893 ± 0.0140
0.35775
11.61
41.52571
~6 minutes
12.38
41.02868
null
null
Llama 3.1 70B
bnb (llm.int8())
0.5604 ± 0.0169
0.6544 ± 0.0139
0.6382 ± 0.0048
0.7940 ± 0.0114
0.66175
6.87
74.26428
~2 minutes
null
null
null
null
Llama 3.1 70B
HQQ (8 bit)
0.6112 ± 0.0146
0.6732 ± 0.0137
0.6661 ± 0.0047
0.8327 ± 0.0105
0.6958
0.98
80.52435
~10 minutes
0.98
80.39013
null
null
Llama 3.1 70B
optimum quanto (int8wo)
0.5591 ± 0.0150
0.6459 ± 0.0140
0.6413 ± 0.0048
0.7979 ± 0.0113
0.66105
1.79
74.21192
~2 minutes
1.8
74.21401
null
null
Llama 3.1 70B
torchao (int8wo)
0.6094 ± 0.0146
0.6732 ± 0.0137
0.6659 ± 0.0047
0.8240 ± 0.0107
0.693125
0.65
89.85038
~2 minutes
0.65
89.84619
null
null
Llama 3.1 70B
fbgemm (fp8)
0.6075 ± 0.0146
0.6732 ± 0.0137
0.6671 ± 0.0047
0.8216 ± 0.0108
0.69235
13.61
74.04624
~6 minutes
null
null
null
null
Llama 3.1 70B
compressed-tensors (fp8)
0.6062 ± 0.0146
0.6741 ± 0.0137
0.6652 ± 0.0047
0.8216 ± 0.0108
0.691775
null
null
null
null
null
null
null
Llama 3.1 70B
VPTQ (2 bit)
0.5451 ± 0.0150
0.6212 ± 0.0142
0.6073 ± 0.0049
0.7901 ± 0.0114
0.640925
6.29
24.89949
~19 hours
6.18
24.89949
null
null
Llama 3.1 70B
AQLM + PV (2 bit)
0.5706 ± 0.0150
0.6365 ± 0.0141
0.6401 ± 0.0048
0.8066 ± 0.0111
0.66345
6.75
23.12739
10-14 days
7.09
23,607.64
null
null
Llama 3.1 70B
GPTQModel (2 bit)
null
null
null
null
null
null
null
null
null
null
null
null
Llama 3.1 70B
AutoGPTQ (2 bit)
0.4556 ± 0.0147
0.2807 ± 0.0131
0.3642 ± 0.0048
0.5470 ± 0.0140
0.411875
null
null
null
null
null
null
null
README.md exists but content is empty.
Downloads last month
25