prapbew commited on
Commit
e5c6f38
·
verified ·
1 Parent(s): 18ca79f

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +13 -1
config.json CHANGED
@@ -11,6 +11,18 @@
11
  "mm_tokens_per_image": 256,
12
  "model_type": "gemma3",
13
  "pad_token_id": 0,
 
 
 
 
 
 
 
 
 
 
 
 
14
  "text_config": {
15
  "attention_bias": false,
16
  "attention_dropout": 0.0,
@@ -60,4 +72,4 @@
60
  "torch_dtype": "float16",
61
  "vision_use_head": false
62
  }
63
- }
 
11
  "mm_tokens_per_image": 256,
12
  "model_type": "gemma3",
13
  "pad_token_id": 0,
14
+ "quantization_config": {
15
+ "bnb_4bit_compute_dtype": "float16",
16
+ "bnb_4bit_quant_type": "nf4",
17
+ "bnb_4bit_use_double_quant": true,
18
+ "llm_int8_enable_fp32_cpu_offload": false,
19
+ "llm_int8_has_fp16_weight": false,
20
+ "llm_int8_skip_modules": null,
21
+ "llm_int8_threshold": 6.0,
22
+ "load_in_4bit": true,
23
+ "load_in_8bit": false,
24
+ "quant_method": "bitsandbytes"
25
+ },
26
  "text_config": {
27
  "attention_bias": false,
28
  "attention_dropout": 0.0,
 
72
  "torch_dtype": "float16",
73
  "vision_use_head": false
74
  }
75
+ }