default_stage: | |
default_modifiers: | |
GPTQModifier: | |
targets: [Linear] | |
ignore: [lm_head, 're:.*mlp.gate$', 're:.*mlp.shared_expert_gate$'] | |
scheme: W4A16 | |
sequential_update: true | |
block_size: 128 | |
dampening_frac: 0.01 | |
offload_hessians: false | |
default_stage: | |
default_modifiers: | |
GPTQModifier: | |
targets: [Linear] | |
ignore: [lm_head, 're:.*mlp.gate$', 're:.*mlp.shared_expert_gate$'] | |
scheme: W4A16 | |
sequential_update: true | |
block_size: 128 | |
dampening_frac: 0.01 | |
offload_hessians: false | |