onebitquantized commited on
Commit
95535c9
1 Parent(s): 531288a

Upload of AutoGPTQ quantized model

Browse files
config.json CHANGED
@@ -33,7 +33,7 @@
33
  "model_file_base_name": null,
34
  "model_name_or_path": null,
35
  "quant_method": "gptq",
36
- "shrink": 0.01,
37
  "static_groups": false,
38
  "sym": false,
39
  "true_sequential": true
 
33
  "model_file_base_name": null,
34
  "model_name_or_path": null,
35
  "quant_method": "gptq",
36
+ "shrink": 0.0625,
37
  "static_groups": false,
38
  "sym": false,
39
  "true_sequential": true
gptq_model-4bit--1g.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:64bdbc1b114418d222a7f4fd72557b09fda9f5d66c242ace7b0ff4a7f4c3970d
3
  size 2990657120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c63030276fa9b54a46f6d3f2fea35f7637289d24638dcefcda5ad5ec903b4cd2
3
  size 2990657120
quantize_config.json CHANGED
@@ -9,7 +9,7 @@
9
  "model_name_or_path": null,
10
  "model_file_base_name": null,
11
  "exponent_hinv": 4.0,
12
- "shrink": 0.01,
13
  "quant_method": "gptq",
14
  "checkpoint_format": "gptq"
15
  }
 
9
  "model_name_or_path": null,
10
  "model_file_base_name": null,
11
  "exponent_hinv": 4.0,
12
+ "shrink": 0.0625,
13
  "quant_method": "gptq",
14
  "checkpoint_format": "gptq"
15
  }