{ "_name_or_path": "morphablediffusion/config.json", "activation_dropout": 0.0, "activation_function": "gelu", "attention_dropout": 0.0, "attention_window": 512, "bos_token_id": 0, "classifier_dropout": 0.0, "d_model": 1024, "data": { "params": { "batch_size": 70, "data_dir": "/cluster/scratch/xiychen/data/facescape_color_calibrated", "mesh_topology": "flame", "num_workers": 1, "shuffled_expression": true }, "target": "ldm.data.facescape.FaceScapeDataset" }, "decoder_attention_heads": 16, "decoder_ffn_dim": 4096, "decoder_layerdrop": 0.0, "decoder_layers": 12, "decoder_start_token_id": 2, "dropout": 0.1, "encoder_attention_heads": 16, "encoder_ffn_dim": 4096, "encoder_layerdrop": 0.0, "encoder_layers": 12, "eos_token_id": 2, "init_std": 0.02, "is_encoder_decoder": true, "lightning": { "callbacks": {}, "modelcheckpoint": { "params": { "every_n_train_steps": 2000 } }, "trainer": { "accumulate_grad_batches": 1, "benchmark": true, "check_val_every_n_epoch": null, "max_steps": 6000, "num_sanity_val_steps": 0, "precision": 32, "val_check_interval": 250 } }, "max_decoder_position_embeddings": 1024, "max_encoder_position_embeddings": 16384, "model": { "base_learning_rate": "5e-5", "params": { "batch_view_num": 4, "cfg_scale": 2.0, "clip_image_encoder_path": "./ckpt/ViT-L-14.pt", "drop_conditions": false, "finetune_unet": true, "image_size": 256, "output_num": 8, "projection": "perspective", "scheduler_config": { "params": { "cycle_lengths": [ 100000 ], "f_max": [ 1.0 ], "f_min": [ 1.0 ], "f_start": [ 0.02 ], "warm_up_steps": [ 100 ] }, "target": "ldm.lr_scheduler.LambdaLinearScheduler" }, "target_elevation": 0, "unet_config": { "params": { "attention_resolutions": [ 4, 2, 1 ], "channel_mult": [ 1, 2, 4, 4 ], "context_dim": 768, "image_size": 32, "in_channels": 8, "legacy": false, "model_channels": 320, "num_heads": 8, "num_res_blocks": 2, "out_channels": 4, "transformer_depth": 1, "use_checkpoint": true, "use_spatial_transformer": true, "volume_dims": [ 64, 128, 256, 512 ] }, "target": "ldm.models.diffusion.attention.DepthWiseAttention" }, "use_spatial_volume": false, "view_num": 16 }, "target": "ldm.models.diffusion.morphable_diffusion.SyncMultiviewDiffusion" }, "model_type": "led", "num_hidden_layers": 12, "pad_token_id": 1, "transformers_version": "4.42.4", "use_cache": true, "vocab_size": 50265 }