{ | |
"bits": 4, | |
"group_size": 128, | |
"sym": true, | |
"data_type": "int", | |
"low_gpu_mem_usage": true, | |
"autoround_version": "0.6.1.dev", | |
"block_name_to_quantize": "model.language_model.layers", | |
"quant_method": "auto-round", | |
"packing_format": "auto_round:auto_gptq" | |
} |