OPEA
/

Safetensors
llama
4-bit precision
auto-round
weiweiz1 commited on
Commit
1135608
·
verified ·
1 Parent(s): 423db67

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +1 -2
config.json CHANGED
@@ -26,7 +26,6 @@
26
  "quantization_config": {
27
  "amp": true,
28
  "autoround_version": "0.4.2.dev",
29
- "backend": "auto_round:gptq:exllamav2",
30
  "batch_size": 8,
31
  "bits": 4,
32
  "data_type": "int",
@@ -41,7 +40,7 @@
41
  "lr": 0.005,
42
  "minmax_lr": 0.005,
43
  "nsamples": 128,
44
- "quant_method": "intel/auto-round",
45
  "scale_dtype": "torch.float16",
46
  "seqlen": 2048,
47
  "sym": true,
 
26
  "quantization_config": {
27
  "amp": true,
28
  "autoround_version": "0.4.2.dev",
 
29
  "batch_size": 8,
30
  "bits": 4,
31
  "data_type": "int",
 
40
  "lr": 0.005,
41
  "minmax_lr": 0.005,
42
  "nsamples": 128,
43
+ "quant_method": "auto-round",
44
  "scale_dtype": "torch.float16",
45
  "seqlen": 2048,
46
  "sym": true,