|
--- |
|
language: |
|
- en |
|
base_model: |
|
- moonshotai/Kimi-K2-Instruct |
|
pipeline_tag: text-generation |
|
tags: |
|
- vllm |
|
- deepseek_v3 |
|
- deepseek |
|
- neuralmagic |
|
- redhat |
|
- llmcompressor |
|
- quantized |
|
- INT4 |
|
- GPTQ |
|
- conversational |
|
- custom_code |
|
- compressed-tensors |
|
- kimi_k2 |
|
license: other |
|
license_name: modified-mit |
|
name: RedHatAI/Kimi-K2-Instruct-quantized.w4a16 |
|
description: This model was obtained by quantizing weights of Kimi-K2-Instruct to INT4 data type. |
|
readme: https://huggingface.co/RedHatAI/Kimi-K2-Instruct-quantized.w4a16/main/README.md |
|
tasks: |
|
- text-to-text |
|
provider: Moonshot AI |
|
license_link: https://huggingface.co/moonshotai/Kimi-K2-Instruct/blob/main/LICENSE |
|
--- |
|
|
|
# Preliminary version of the model |
|
|
|
## Evaluations |
|
|
|
- GSM8k, 5-shot via lm-evaluation-harness |
|
``` |
|
moonshotai/Kimi-K2-Instruct = 94.92 |
|
RedHatAI/Kimi-K2-Instruct-quantized.w4a16 (this model) = 94.84 |
|
``` |
|
|
|
More evals coming very soon... |
|
|