Qwen2_5-32B_instruct_fp8_static / hf_quant_config.json
woodchen7's picture
Upload hf_quant_config.json with huggingface_hub
be2f791 verified
raw
history blame contribute delete
192 Bytes
{
"quantization": {
"quant_algo": "FP8",
"kv_cache_quant_algo": null,
"exclude_modules": [
"lm_head",
"model.embed_tokens"
]
}
}