Orignal Model by: Qingyun Li
Orignal Model: GRMR-2B-Instruct

For more information about the model, I highly recommend checking out the original model page and the creator while you're at it.

ExLlamaV2 Quantizations:
8.0bpw: 8hb | 6hb
7.75bpw: 8hb | 6hb
7.5bpw: 8hb | 6hb
7.25bpw: 8hb | 6hb
7.0bpw: 8hb | 6hb
6.75bpw: 8hb | 6hb
6.5bpw: 8hb | 6hb
6.25bpw: 8hb | 6hb
6.0bpw: 8hb | 6hb
5.75bpw: 8hb | 6hb
5.5bpw: 8hb | 6hb
5.25bpw: 8hb | 6hb
5.0bpw: 8hb | 6hb
4.75bpw: 8hb | 6hb
4.5bpw: 8hb | 6hb
4.25bpw: 8hb | 6hb
4.0bpw: 8hb | 6hb
3.75bpw: 8hb | 6hb
3.5bpw: 8hb | 6hb
3.25bpw: 8hb | 6hb
3.0bpw: 8hb | 6hb
2.75bpw: 8hb | 6hb
2.5bpw: 8hb | 6hb
2.25bpw: 8hb | 6hb
2.0bpw: 8hb | 6hb

Measurement File (Default/built-in calibration dataset was used)

If you need a specific model quantized or particular bits per weight, please let me know. I’m happy to help.

Your feedback and suggestions are always welcome! They help me improve and make quantizations better for everyone.

Special thanks to turboderp for developing the tools that made these quantizations possible. Your contributions are greatly appreciated!

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.

Model tree for TheMelonGod/GRMR-2B-Instruct-exl2

Base model

google/gemma-2-2b
Quantized
(4)
this model