File size: 1,028 Bytes
836f100
 
 
 
582bb79
836f100
582bb79
 
836f100
 
 
 
 
 
582bb79
836f100
 
582bb79
 
836f100
582bb79
 
836f100
 
 
 
 
 
 
 
582bb79
836f100
582bb79
836f100
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
{
    "batcher": null,
    "cacher": "fora",
    "compiler": "torch_compile",
    "factorizer": null,
    "pruner": null,
    "quantizer": "torchao",
    "fora_interval": 3,
    "fora_start_step": 2,
    "torch_compile_backend": "inductor",
    "torch_compile_dynamic": null,
    "torch_compile_fullgraph": true,
    "torch_compile_make_portable": false,
    "torch_compile_max_kv_cache_size": 400,
    "torch_compile_mode": "max-autotune-no-cudagraphs",
    "torch_compile_seqlen_manual_cuda_graph": 100,
    "torch_compile_target": "model",
    "torchao_excluded_modules": "norm+embedding",
    "torchao_quant_type": "int8dq",
    "batch_size": 1,
    "device": "cpu",
    "device_map": null,
    "save_fns": [
        "save_before_apply",
        "save_before_apply"
    ],
    "load_fns": [
        "diffusers"
    ],
    "reapply_after_load": {
        "factorizer": null,
        "pruner": null,
        "quantizer": "torchao",
        "cacher": "fora",
        "compiler": "torch_compile",
        "batcher": null
    }
}