{ | |
"i2v": true, | |
"use_audio": true, | |
"random_prefix_frames": true, | |
"sp_size": 1, | |
"text_encoder_path": "/tmp/pretrained_models/Wan2.1-T2V-14B/models_t5_umt5-xxl-enc-bf16.pth", | |
"image_encoder_path": "None", | |
"dit_path": "/tmp/pretrained_models/Wan2.1-T2V-14B/diffusion_pytorch_model-00001-of-00006.safetensors,/tmp/pretrained_models/Wan2.1-T2V-14B/diffusion_pytorch_model-00002-of-00006.safetensors,/tmp/pretrained_models/Wan2.1-T2V-14B/diffusion_pytorch_model-00003-of-00006.safetensors,/tmp/pretrained_models/Wan2.1-T2V-14B/diffusion_pytorch_model-00004-of-00006.safetensors,/tmp/pretrained_models/Wan2.1-T2V-14B/diffusion_pytorch_model-00005-of-00006.safetensors,/tmp/pretrained_models/Wan2.1-T2V-14B/diffusion_pytorch_model-00006-of-00006.safetensors", | |
"model_config": { | |
"in_dim": 33, | |
"audio_hidden_size": 32 | |
}, | |
"train_architecture": "lora", | |
"lora_target_modules": "q,k,v,o,ffn.0,ffn.2", | |
"init_lora_weights": "kaiming", | |
"lora_rank": 128, | |
"lora_alpha": 64.0 | |
} | |