File size: 1,028 Bytes
836f100 582bb79 836f100 582bb79 836f100 582bb79 836f100 582bb79 836f100 582bb79 836f100 582bb79 836f100 582bb79 836f100 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 |
{
"batcher": null,
"cacher": "fora",
"compiler": "torch_compile",
"factorizer": null,
"pruner": null,
"quantizer": "torchao",
"fora_interval": 3,
"fora_start_step": 2,
"torch_compile_backend": "inductor",
"torch_compile_dynamic": null,
"torch_compile_fullgraph": true,
"torch_compile_make_portable": false,
"torch_compile_max_kv_cache_size": 400,
"torch_compile_mode": "max-autotune-no-cudagraphs",
"torch_compile_seqlen_manual_cuda_graph": 100,
"torch_compile_target": "model",
"torchao_excluded_modules": "norm+embedding",
"torchao_quant_type": "int8dq",
"batch_size": 1,
"device": "cpu",
"device_map": null,
"save_fns": [
"save_before_apply",
"save_before_apply"
],
"load_fns": [
"diffusers"
],
"reapply_after_load": {
"factorizer": null,
"pruner": null,
"quantizer": "torchao",
"cacher": "fora",
"compiler": "torch_compile",
"batcher": null
}
} |