--- base_model: - google/gemma-3-27b-it base_model_relation: quantized license: mit --- # Model Card - Base model: `google/gemma-3-27b-it` - Quantization method: SqueezeLLM - Target bit-width: 3 - Backend kernel: Any-Precision-LLM kernel (`ap-gemv`) - Calibration data: RedPajama (1024 sentences / 4096 tokens) - Calibration objective: Next-token prediction # How to run - Follow the instruction in https://github.com/snu-mllab/GuidedQuant. # References - [Model Paper](https://arxiv.org/abs/2505.07004)