Dataset Viewer
model
string | quant method
string | truthfulqa_mc2 acc ± stderr
string | arc:challenge acc ± stderr
string | hellaswag acc
string | winogrande acc
string | average
float64 | throughput (tok/s)
float64 | peak process vram (GB)
float64 | calibration/quantization time
string | throughput w/ torch.compile
float64 | peak process vram w/ torch.compile
float64 | througput w/ marlin
float64 | peak process vram w/ marlin
float64 |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Llama 3.1 8B | baseline (bf16) | 0.5456 ± 0.0150 | 0.5623 ± 0.0145 | 0.5973 ± 0.0049 | 0.7372 ± 0.0124 | 0.6106 | 38.82 | 16.77302 | null | 79.27 | 17.12954 | null | null |
Llama 3.1 8B | AWQ (4 bit) | 0.5281 ± 0.0150 | 0.5589 ± .0145 | .5869 ± .0049 | 0.7285 ± 0.0125 | 0.6006 | 43.09 | 8.11808 | ~10 minutes | null | null | null | null |
Llama 3.1 8B | GPTQModel (4 bit) | null | null | null | null | null | 36.83 | 6.31872 | ~23 minutes | null | null | 37.84 | 6,318.72 |
Llama 3.1 8B | AutoGPTQ (4 bit) | 0.5361 ± 0.0149 | 0.5589 ± 0.0145 | 0.5758 ± 0.0049 | 0.7103 ± 0.0127 | 0.595275 | 43.71 | 6.49278 | ~30 minutes | null | null | null | null |
Llama 3.1 8B | bnb (nf4) | 0.5446 ± 0.0149 | 0.5563 ± 0.0145 | 0.5769 ± 0.0049 | 0.7301 ± 0.0125 | 0.601975 | 24.35 | 6.44035 | ~1 minute | null | null | null | null |
Llama 3.1 8B | optimum quanto (w4a16) | 0.5162 ± 0.0150 | 0.5427 ± 0.0146 | 0.5856 ± 0.0049 | 0.7419 ± 0.0123 | 0.5966 | 31.22 | 6.57667 | ~30 seconds | null | null | null | null |
Llama 3.1 8B | torchao (int4wo) | 0.5166 ±0.0150 | 0.5418 ± 0.0146 | 0.5855 ± 0.0049 | 0.7395 ± 0.0123 | 0.59585 | 24.98 | 6.49907 | ~20 seconds | 85.76 | 6.8493 | null | null |
Llama 3.1 8B | HQQ (4 bit) | 0.5485 ± 0.0150 | 0.5495 ± 0.0145 | 0.5859 ± 0.0049 | 0.7285 ± 0.0125 | 0.6031 | 34.44 | 6.71718 | null | null | null | null | null |
Llama 3.1 8B | HIGGS (4 bit) | 0.5214 ± 0.0149 | 0.5486 ± 0.0145 | 0.5836 ± 0.0049 | 0.7182 ± 0.0126 | 0.59295 | 28.35 | 6.81994 | ~5 minutes | null | null | null | null |
Llama 3.1 8B | bnb (llm.int8()) | 0.5446 ± 0.0150 | 0.5623 ± 0.0145 | 0.5949 ± 0.0049 | 0.7324 ± 0.0124 | 0.60855 | 20.75 | 9.70772 | ~20 seconds | null | null | null | null |
Llama 3.1 8B | HQQ (8 bit) | 0.5441 ± 0.0150 | 0.5666 ± 0.0145 | 0.5979 ± 0.0049 | 0.7380 ± 0.0124 | 0.61165 | 9.07 | 10.6011 | ~80 seconds | null | null | null | null |
Llama 3.1 8B | optimum quanto (int8wo) | 0.5436 ± 0.0150 | 0.5640 ± 0.0145 | 0.5992 ± 0.0049 | 0.7372 ± 0.0124 | 0.611 | 15.59 | 9.81887 | ~20 seconds | 16.01 | 10.07052 | null | null |
Llama 3.1 8B | torchao (int8wo) | 0.5449 ± 0.0150 | 0.5640 ± 0.0145 | 0.5975 ± 0.0049 | 0.7380 ± 0.0124 | 0.6111 | 5.98 | 13.07155 | ~30 seconds | 43.79 | 13.66714 | null | null |
Llama 3.1 8B | fbgemm (fp8) | 0.5430 ± 0.0150 | 0.5580 ± 0.0145 | 0.5958 ± 0.0049 | 0.7411 ± 0.0123 | 0.609475 | 33.83 | 10.00551 | ~30 seconds | null | null | null | null |
Llama 3.1 8B | compressed-tensors (fp8) | 0.5398 ± 0.0151 | 0.5589 ± 0.0145 | 0.5950 ± 0.0049 | 0.7356 ± 0.0124 | 0.607325 | null | null | null | null | null | null | null |
Llama 3.1 8B | VPTQ (2 bit) | 0.4543 ± 0.0149 | 0.4923 ± 0.0146 | 0.5258 ± 0.0050 | 0.6930 ± 0.0130 | 0.54135 | 32.35 | 5.28902 | ~2 hours | 31.48 | 5.28692 | null | null |
Llama 3.1 8B | AQLM + PV (2 bit) | 0.5036 ± 0.0148 | 0.5230 ± 0.0146 | 0.5628 ± 0.0050 | 0.6938 ± 0.0130 | 0.5708 | 22.28 | 4.84023 | ~1 day | 27.27 | 4.85491 | null | null |
Llama 3.1 8B | GPTQModel (2 bit) | null | null | null | null | null | 19.02 | 18.45284 | ~26 minutes | null | null | null | null |
Llama 3.1 8B | AutoGPTQ (2 bit) | 0.5127 ± 0.0150 | 0.1988 ± 0.0117 | 0.2665 ± 0.0044 | 0.4799 ± 0.0140 | 0.364475 | 6.25 | 11.02473 | ~26 minutes | null | null | null | null |
Llama 3.1 70B | baseline (bf16) | 0.6068 ± 0.0147 | 0.6732 ± 0.0137 | 0.6666 ± 0.0047 | 0.8248 ± 0.0107 | 0.69285 | 9.73 | 142.26869 | null | 10.1 | 142.81395 | null | null |
Llama 3.1 70B | AWQ (4 bit) | 0.5706 ± 0.0150 | 0.6681 ± 0.0138 | 0.6598 ± 0.0047 | 0.8193 ± 0.0108 | 0.67945 | 15.74 | 43.75288 | ~1 hour | null | null | null | null |
Llama 3.1 70B | GPTQModel (4 bit) | null | null | null | null | null | 14.84 | 40.5757 | null | null | null | 15.28 | 40.5757 |
Llama 3.1 70B | AutoGPTQ (4 bit) | 0.5937 ± 0.0147 | 0.6655 ± 0.0138 | 0.6568 ± 0.0047 | 0.8185 ± 0.0108 | 0.683625 | 0.46 | 42.40022 | ~2 hours | null | null | null | null |
Llama 3.1 70B | bnb (nf4) | 0.5939 ± 0.0148 | 0.6724 ± 0.0137 | 0.6592 ± 0.0047 | 0.8098 ± 0.0110 | 0.683825 | 11.27 | 44.62949 | ~2 minutes | null | null | null | null |
Llama 3.1 70B | optimum quanto (w4a16) | 0.4847 ± 0.0164 | 0.2082 ± 0.0119 | 0.2582 ± 0.0044 | 0.4878 ± 0.0140 | 0.359725 | 12.97 | 80.39013 | ~2 minutes | null | null | null | null |
Llama 3.1 70B | torchao (int4wo) | 0.4847 ± 0.0164 | 0.2108 ± 0.0119 | 0.2581 ± 0.0044 | 0.4980 ± 0.0141 | 0.3629 | 10.56 | 41.6054 | ~2 minutes | 18.95 | 42.26181 | null | null |
Llama 3.1 70B | HQQ (4 bit) | 0.5882 ± 0.0146 | 0.6706 ± 0.0137 | 0.6597 ± 0.0047 | 0.8035 ± 0.0112 | 0.6805 | 13.92 | 44.50366 | ~10 minutes | null | null | null | null |
Llama 3.1 70B | HIGGS (4 bit) | 0.4871 ± 0.0163 | 0.1971 ± 0.0116 | 0.2575 ± 0.0044 | 0.4893 ± 0.0140 | 0.35775 | 11.61 | 41.52571 | ~6 minutes | 12.38 | 41.02868 | null | null |
Llama 3.1 70B | bnb (llm.int8()) | 0.5604 ± 0.0169 | 0.6544 ± 0.0139 | 0.6382 ± 0.0048 | 0.7940 ± 0.0114 | 0.66175 | 6.87 | 74.26428 | ~2 minutes | null | null | null | null |
Llama 3.1 70B | HQQ (8 bit) | 0.6112 ± 0.0146 | 0.6732 ± 0.0137 | 0.6661 ± 0.0047 | 0.8327 ± 0.0105 | 0.6958 | 0.98 | 80.52435 | ~10 minutes | 0.98 | 80.39013 | null | null |
Llama 3.1 70B | optimum quanto (int8wo) | 0.5591 ± 0.0150 | 0.6459 ± 0.0140 | 0.6413 ± 0.0048 | 0.7979 ± 0.0113 | 0.66105 | 1.79 | 74.21192 | ~2 minutes | 1.8 | 74.21401 | null | null |
Llama 3.1 70B | torchao (int8wo) | 0.6094 ± 0.0146 | 0.6732 ± 0.0137 | 0.6659 ± 0.0047 | 0.8240 ± 0.0107 | 0.693125 | 0.65 | 89.85038 | ~2 minutes | 0.65 | 89.84619 | null | null |
Llama 3.1 70B | fbgemm (fp8) | 0.6075 ± 0.0146 | 0.6732 ± 0.0137 | 0.6671 ± 0.0047 | 0.8216 ± 0.0108 | 0.69235 | 13.61 | 74.04624 | ~6 minutes | null | null | null | null |
Llama 3.1 70B | compressed-tensors (fp8) | 0.6062 ± 0.0146 | 0.6741 ± 0.0137 | 0.6652 ± 0.0047 | 0.8216 ± 0.0108 | 0.691775 | null | null | null | null | null | null | null |
Llama 3.1 70B | VPTQ (2 bit) | 0.5451 ± 0.0150 | 0.6212 ± 0.0142 | 0.6073 ± 0.0049 | 0.7901 ± 0.0114 | 0.640925 | 6.29 | 24.89949 | ~19 hours | 6.18 | 24.89949 | null | null |
Llama 3.1 70B | AQLM + PV (2 bit) | 0.5706 ± 0.0150 | 0.6365 ± 0.0141 | 0.6401 ± 0.0048 | 0.8066 ± 0.0111 | 0.66345 | 6.75 | 23.12739 | 10-14 days | 7.09 | 23,607.64 | null | null |
Llama 3.1 70B | GPTQModel (2 bit) | null | null | null | null | null | null | null | null | null | null | null | null |
Llama 3.1 70B | AutoGPTQ (2 bit) | 0.4556 ± 0.0147 | 0.2807 ± 0.0131 | 0.3642 ± 0.0048 | 0.5470 ± 0.0140 | 0.411875 | null | null | null | null | null | null | null |
README.md exists but content is empty.
- Downloads last month
- 25