fxmarty commited on
Commit
b5e9b5d
·
verified ·
1 Parent(s): e4a232e

Upload config.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. config.json +2 -3
config.json CHANGED
@@ -67,7 +67,6 @@
67
  "lm_head"
68
  ],
69
  "export": {
70
- "custom_mode": "awq",
71
  "kv_cache_group": [],
72
  "pack_method": "reorder",
73
  "weight_format": "real_quantized",
@@ -94,14 +93,14 @@
94
  "layer_type_quant_config": {},
95
  "pack_method": "reorder",
96
  "quant_method": "quark",
97
- "quant_mode": 1
98
  },
99
  "rms_norm_eps": 1e-06,
100
  "rope_scaling": null,
101
  "rope_theta": 10000.0,
102
  "tie_word_embeddings": false,
103
  "torch_dtype": "float32",
104
- "transformers_version": "4.46.0.dev0",
105
  "use_cache": true,
106
  "vocab_size": 32000
107
  }
 
67
  "lm_head"
68
  ],
69
  "export": {
 
70
  "kv_cache_group": [],
71
  "pack_method": "reorder",
72
  "weight_format": "real_quantized",
 
93
  "layer_type_quant_config": {},
94
  "pack_method": "reorder",
95
  "quant_method": "quark",
96
+ "quant_mode": "eager_mode"
97
  },
98
  "rms_norm_eps": 1e-06,
99
  "rope_scaling": null,
100
  "rope_theta": 10000.0,
101
  "tie_word_embeddings": false,
102
  "torch_dtype": "float32",
103
+ "transformers_version": "4.45.2",
104
  "use_cache": true,
105
  "vocab_size": 32000
106
  }