Upload 3 files
Browse files- args.json +52 -0
- config.yaml +51 -0
- pytorch_model.bin +3 -0
args.json
ADDED
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"guidance_scale_img": 4.5,
|
3 |
+
"guidance_scale_text": 7.5,
|
4 |
+
"config_p": "./configs/VITONHD_agnidmhand_agnv9hand_randdilate_gpt4ov6v12_jointcond_nopose_noip_intermcloth.yaml",
|
5 |
+
"use_sd15": false,
|
6 |
+
"resume_path": null,
|
7 |
+
"pretrained_unet_path": null,
|
8 |
+
"pretrained_unet_before_path": null,
|
9 |
+
"gradient_accumulation_steps": 1,
|
10 |
+
"prediction_type": "epsilon",
|
11 |
+
"timestep_spacing": "leading",
|
12 |
+
"noise_offset": 0.05,
|
13 |
+
"num_inference_steps": 30,
|
14 |
+
"finetuning": false,
|
15 |
+
"save_step": 15000,
|
16 |
+
"validation_step": 30000,
|
17 |
+
"validation_step_lst": [
|
18 |
+
2,
|
19 |
+
30000,
|
20 |
+
60000
|
21 |
+
],
|
22 |
+
"skip_paired": false,
|
23 |
+
"skip_unpaired": false,
|
24 |
+
"n_samples_train": 2,
|
25 |
+
"n_samples_valid": 999999,
|
26 |
+
"save_root_dir": "./logs",
|
27 |
+
"save_name": "VITONHD_agnidmhand_agnv9hand_randdilate_gpt4ov6v12_jointcond_nopose_noip_intermcloth_bs32",
|
28 |
+
"strength": 1.0,
|
29 |
+
"img_h": 1024,
|
30 |
+
"img_w": 768,
|
31 |
+
"use_8bit_adam": false,
|
32 |
+
"no_zero_snr": false,
|
33 |
+
"init_model_path": "./pretrained_models/stable-diffusion-xl-1.0-inpainting-0.1",
|
34 |
+
"init_vae_path": "./pretrained_models/sdxl-vae-fp16-fix",
|
35 |
+
"init_ip_adapter_path": "./pretrained_models/IP-Adapter/sdxl_models/ip-adapter_sdxl.bin",
|
36 |
+
"init_image_encoder_path": "./pretrained_models/IP-Adapter/sdxl_models",
|
37 |
+
"init_cloth_encoder_path": "./pretrained_models/stable-diffusion-xl-base-1.0",
|
38 |
+
"ip_adapter_num_tokens": 4,
|
39 |
+
"logging_dir": "wandb",
|
40 |
+
"learning_rate": 3.2e-05,
|
41 |
+
"lr_scheduler": "constant_with_warmup",
|
42 |
+
"no_lr_scheduler_accelerator": false,
|
43 |
+
"lr_warmup_steps": 500,
|
44 |
+
"lr_num_cycles": 1,
|
45 |
+
"lr_power": 1.0,
|
46 |
+
"weight_decay": 0.01,
|
47 |
+
"guidance_scale": 2.0,
|
48 |
+
"num_workers": 2,
|
49 |
+
"mixed_precision": "fp16",
|
50 |
+
"report_to": "wandb",
|
51 |
+
"total_batch_size": 32
|
52 |
+
}
|
config.yaml
ADDED
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
no_pose: true
|
2 |
+
use_jointcond: true
|
3 |
+
no_ipadapter: true
|
4 |
+
use_interm_cloth_mask: true
|
5 |
+
repaint_start_ratio: 0.5
|
6 |
+
dataset:
|
7 |
+
dataset_name: VITONHDDataset
|
8 |
+
data_root_dir: ./DATA/zalando-hd-resized
|
9 |
+
img_spatial_transform_lst:
|
10 |
+
- randomresizedcrop
|
11 |
+
- randomaffine
|
12 |
+
cloth_spatial_transform_lst:
|
13 |
+
- randomresizedcrop
|
14 |
+
- randomaffine
|
15 |
+
img_cloth_spatial_transform_lst:
|
16 |
+
- hflip
|
17 |
+
color_transform_lst:
|
18 |
+
- colorjitter
|
19 |
+
i_drop_rate: 0.05
|
20 |
+
pose_type: densepose
|
21 |
+
train_folder_name: train_agnidm_sapiens_hand
|
22 |
+
test_folder_name: test_agnidm_sapiens_hand
|
23 |
+
prompt_version: v12
|
24 |
+
text_file_postfix: gpt4o_v6.json
|
25 |
+
train_folder_name_for_interm_cloth_mask: train_agnv9_sapiens_hand
|
26 |
+
test_folder_name_for_interm_cloth_mask: test_agnv9_sapiens_hand
|
27 |
+
use_rand_dilate: true
|
28 |
+
train_min_dilate_folder_name: train_agnidm_sapiens_hand
|
29 |
+
train_max_dilate_folder_name: train_agnv9_sapiens_hand
|
30 |
+
rand_dilate_miniter: 0
|
31 |
+
rand_dilate_maxiter: 200
|
32 |
+
extra_train_dataset:
|
33 |
+
dataset1:
|
34 |
+
dataset_name: VITONHDDataset
|
35 |
+
data_root_dir: ./DATA/zalando-hd-resized
|
36 |
+
img_spatial_transform_lst:
|
37 |
+
- randomresizedcrop
|
38 |
+
- randomaffine
|
39 |
+
cloth_spatial_transform_lst:
|
40 |
+
- randomresizedcrop
|
41 |
+
- randomaffine
|
42 |
+
img_cloth_spatial_transform_lst:
|
43 |
+
- hflip
|
44 |
+
color_transform_lst:
|
45 |
+
- colorjitter
|
46 |
+
i_drop_rate: 0.05
|
47 |
+
pose_type: densepose
|
48 |
+
train_folder_name: train_agnv9_sapiens_hand
|
49 |
+
test_folder_name: test_agnv9_sapiens_hand
|
50 |
+
prompt_version: v12
|
51 |
+
text_file_postfix: gpt4o_v6.json
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6c01a173d0fc1892fe75730c3cf15bf9d81b7e80b41c4cf1fa49a1cb20bfb6dd
|
3 |
+
size 10270149870
|