woodchen7's picture
Upload hf_quant_config.json with huggingface_hub
cba5f50 verified
raw
history blame contribute delete
192 Bytes
{
"quantization": {
"quant_algo": "FP8",
"kv_cache_quant_algo": null,
"exclude_modules": [
"lm_head",
"model.embed_tokens"
]
}
}