Initial Upload
Browse files- Icon.png +0 -0
- README.md +26 -0
- Sample.png +0 -0
- Sample2.png +0 -0
- Sample3.png +0 -0
- Sample4.png +0 -0
- controlnet/config.json +67 -0
- controlnet/model.onnx +3 -0
- model_index.json +37 -0
- scheduler/scheduler_config.json +15 -0
- text_encoder/config.json +25 -0
- text_encoder/model.onnx +3 -0
- tokenizer/merges.txt +0 -0
- tokenizer/model.onnx +3 -0
- tokenizer/special_tokens_map.json +24 -0
- tokenizer/tokenizer_config.json +30 -0
- tokenizer/vocab.json +0 -0
- unet/config.json +67 -0
- unet/model.onnx +3 -0
- vae_decoder/config.json +36 -0
- vae_decoder/model.onnx +3 -0
- vae_encoder/config.json +36 -0
- vae_encoder/model.onnx +3 -0
    	
        Icon.png
    ADDED
    
    |   | 
    	
        README.md
    ADDED
    
    | @@ -0,0 +1,26 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            # SilversRealmix LCM - Onnx Olive DirectML Optimized 
         | 
| 2 | 
            +
             | 
| 3 | 
            +
            ## Original Model
         | 
| 4 | 
            +
            https://civitai.com/models/226288/silvers-realmix-lcm-edition
         | 
| 5 | 
            +
             | 
| 6 | 
            +
            ## C# Inference Demo
         | 
| 7 | 
            +
            https://github.com/TensorStack-AI/OnnxStack
         | 
| 8 | 
            +
             | 
| 9 | 
            +
            ```csharp
         | 
| 10 | 
            +
            // Create Pipeline
         | 
| 11 | 
            +
            var pipeline = LatentConsistencyPipeline.CreatePipeline("D:\\Models\\SilversRealmix-LCM-onnx");
         | 
| 12 | 
            +
             | 
| 13 | 
            +
            // Prompt
         | 
| 14 | 
            +
            var promptOptions = new PromptOptions
         | 
| 15 | 
            +
            {
         | 
| 16 | 
            +
                Prompt = "Illustrate a fierce barbarian man, with muscles rippling and wielding a massive axe, roaring in triumph"
         | 
| 17 | 
            +
            };
         | 
| 18 | 
            +
             | 
| 19 | 
            +
            // Run pipeline
         | 
| 20 | 
            +
            var result = await pipeline.GenerateImageAsync(promptOptions);
         | 
| 21 | 
            +
             | 
| 22 | 
            +
            // Save Image Result
         | 
| 23 | 
            +
            await result.SaveAsync("Result.png");
         | 
| 24 | 
            +
            ```
         | 
| 25 | 
            +
            ## Inference Result
         | 
| 26 | 
            +
            
         | 
    	
        Sample.png
    ADDED
    
    |   | 
    	
        Sample2.png
    ADDED
    
    |   | 
    	
        Sample3.png
    ADDED
    
    |   | 
    	
        Sample4.png
    ADDED
    
    |   | 
    	
        controlnet/config.json
    ADDED
    
    | @@ -0,0 +1,67 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_class_name": "UNet2DConditionModel",
         | 
| 3 | 
            +
              "_diffusers_version": "0.30.0.dev0",
         | 
| 4 | 
            +
              "act_fn": "silu",
         | 
| 5 | 
            +
              "addition_embed_type": null,
         | 
| 6 | 
            +
              "addition_embed_type_num_heads": 64,
         | 
| 7 | 
            +
              "addition_time_embed_dim": null,
         | 
| 8 | 
            +
              "attention_head_dim": 8,
         | 
| 9 | 
            +
              "attention_type": "default",
         | 
| 10 | 
            +
              "block_out_channels": [
         | 
| 11 | 
            +
                320,
         | 
| 12 | 
            +
                640,
         | 
| 13 | 
            +
                1280,
         | 
| 14 | 
            +
                1280
         | 
| 15 | 
            +
              ],
         | 
| 16 | 
            +
              "center_input_sample": false,
         | 
| 17 | 
            +
              "class_embed_type": null,
         | 
| 18 | 
            +
              "class_embeddings_concat": false,
         | 
| 19 | 
            +
              "conv_in_kernel": 3,
         | 
| 20 | 
            +
              "conv_out_kernel": 3,
         | 
| 21 | 
            +
              "cross_attention_dim": 768,
         | 
| 22 | 
            +
              "cross_attention_norm": null,
         | 
| 23 | 
            +
              "down_block_types": [
         | 
| 24 | 
            +
                "CrossAttnDownBlock2D",
         | 
| 25 | 
            +
                "CrossAttnDownBlock2D",
         | 
| 26 | 
            +
                "CrossAttnDownBlock2D",
         | 
| 27 | 
            +
                "DownBlock2D"
         | 
| 28 | 
            +
              ],
         | 
| 29 | 
            +
              "downsample_padding": 1,
         | 
| 30 | 
            +
              "dropout": 0.0,
         | 
| 31 | 
            +
              "dual_cross_attention": false,
         | 
| 32 | 
            +
              "encoder_hid_dim": null,
         | 
| 33 | 
            +
              "encoder_hid_dim_type": null,
         | 
| 34 | 
            +
              "flip_sin_to_cos": true,
         | 
| 35 | 
            +
              "freq_shift": 0,
         | 
| 36 | 
            +
              "in_channels": 4,
         | 
| 37 | 
            +
              "layers_per_block": 2,
         | 
| 38 | 
            +
              "mid_block_only_cross_attention": null,
         | 
| 39 | 
            +
              "mid_block_scale_factor": 1,
         | 
| 40 | 
            +
              "mid_block_type": "UNetMidBlock2DCrossAttn",
         | 
| 41 | 
            +
              "norm_eps": 1e-05,
         | 
| 42 | 
            +
              "norm_num_groups": 32,
         | 
| 43 | 
            +
              "num_attention_heads": null,
         | 
| 44 | 
            +
              "num_class_embeds": null,
         | 
| 45 | 
            +
              "only_cross_attention": false,
         | 
| 46 | 
            +
              "out_channels": 4,
         | 
| 47 | 
            +
              "projection_class_embeddings_input_dim": null,
         | 
| 48 | 
            +
              "resnet_out_scale_factor": 1.0,
         | 
| 49 | 
            +
              "resnet_skip_time_act": false,
         | 
| 50 | 
            +
              "resnet_time_scale_shift": "default",
         | 
| 51 | 
            +
              "reverse_transformer_layers_per_block": null,
         | 
| 52 | 
            +
              "sample_size": 64,
         | 
| 53 | 
            +
              "time_cond_proj_dim": null,
         | 
| 54 | 
            +
              "time_embedding_act_fn": null,
         | 
| 55 | 
            +
              "time_embedding_dim": null,
         | 
| 56 | 
            +
              "time_embedding_type": "positional",
         | 
| 57 | 
            +
              "timestep_post_act": null,
         | 
| 58 | 
            +
              "transformer_layers_per_block": 1,
         | 
| 59 | 
            +
              "up_block_types": [
         | 
| 60 | 
            +
                "UpBlock2D",
         | 
| 61 | 
            +
                "CrossAttnUpBlock2D",
         | 
| 62 | 
            +
                "CrossAttnUpBlock2D",
         | 
| 63 | 
            +
                "CrossAttnUpBlock2D"
         | 
| 64 | 
            +
              ],
         | 
| 65 | 
            +
              "upcast_attention": false,
         | 
| 66 | 
            +
              "use_linear_projection": false
         | 
| 67 | 
            +
            }
         | 
    	
        controlnet/model.onnx
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:5bdae83a234856c951d6beda7d1fbc3c2ff10ea59adac4df2446be719cd2c494
         | 
| 3 | 
            +
            size 1719495908
         | 
    	
        model_index.json
    ADDED
    
    | @@ -0,0 +1,37 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_class_name": "LatentConsistencyModelPipeline",
         | 
| 3 | 
            +
              "_diffusers_version": "0.30.0.dev0",
         | 
| 4 | 
            +
              "feature_extractor": [
         | 
| 5 | 
            +
                "transformers",
         | 
| 6 | 
            +
                "CLIPImageProcessor"
         | 
| 7 | 
            +
              ],
         | 
| 8 | 
            +
              "image_encoder": [
         | 
| 9 | 
            +
                null,
         | 
| 10 | 
            +
                null
         | 
| 11 | 
            +
              ],
         | 
| 12 | 
            +
              "requires_safety_checker": true,
         | 
| 13 | 
            +
              "safety_checker": [
         | 
| 14 | 
            +
                null,
         | 
| 15 | 
            +
                null
         | 
| 16 | 
            +
              ],
         | 
| 17 | 
            +
              "scheduler": [
         | 
| 18 | 
            +
                "diffusers",
         | 
| 19 | 
            +
                "PNDMScheduler"
         | 
| 20 | 
            +
              ],
         | 
| 21 | 
            +
              "text_encoder": [
         | 
| 22 | 
            +
                "transformers",
         | 
| 23 | 
            +
                "CLIPTextModel"
         | 
| 24 | 
            +
              ],
         | 
| 25 | 
            +
              "tokenizer": [
         | 
| 26 | 
            +
                "transformers",
         | 
| 27 | 
            +
                "CLIPTokenizer"
         | 
| 28 | 
            +
              ],
         | 
| 29 | 
            +
              "unet": [
         | 
| 30 | 
            +
                "diffusers",
         | 
| 31 | 
            +
                "UNet2DConditionModel"
         | 
| 32 | 
            +
              ],
         | 
| 33 | 
            +
              "vae": [
         | 
| 34 | 
            +
                "diffusers",
         | 
| 35 | 
            +
                "AutoencoderKL"
         | 
| 36 | 
            +
              ]
         | 
| 37 | 
            +
            }
         | 
    	
        scheduler/scheduler_config.json
    ADDED
    
    | @@ -0,0 +1,15 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_class_name": "PNDMScheduler",
         | 
| 3 | 
            +
              "_diffusers_version": "0.30.0.dev0",
         | 
| 4 | 
            +
              "beta_end": 0.012,
         | 
| 5 | 
            +
              "beta_schedule": "scaled_linear",
         | 
| 6 | 
            +
              "beta_start": 0.00085,
         | 
| 7 | 
            +
              "clip_sample": false,
         | 
| 8 | 
            +
              "num_train_timesteps": 1000,
         | 
| 9 | 
            +
              "prediction_type": "epsilon",
         | 
| 10 | 
            +
              "set_alpha_to_one": false,
         | 
| 11 | 
            +
              "skip_prk_steps": true,
         | 
| 12 | 
            +
              "steps_offset": 1,
         | 
| 13 | 
            +
              "timestep_spacing": "leading",
         | 
| 14 | 
            +
              "trained_betas": null
         | 
| 15 | 
            +
            }
         | 
    	
        text_encoder/config.json
    ADDED
    
    | @@ -0,0 +1,25 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_name_or_path": "openai/clip-vit-large-patch14",
         | 
| 3 | 
            +
              "architectures": [
         | 
| 4 | 
            +
                "CLIPTextModel"
         | 
| 5 | 
            +
              ],
         | 
| 6 | 
            +
              "attention_dropout": 0.0,
         | 
| 7 | 
            +
              "bos_token_id": 0,
         | 
| 8 | 
            +
              "dropout": 0.0,
         | 
| 9 | 
            +
              "eos_token_id": 2,
         | 
| 10 | 
            +
              "hidden_act": "quick_gelu",
         | 
| 11 | 
            +
              "hidden_size": 768,
         | 
| 12 | 
            +
              "initializer_factor": 1.0,
         | 
| 13 | 
            +
              "initializer_range": 0.02,
         | 
| 14 | 
            +
              "intermediate_size": 3072,
         | 
| 15 | 
            +
              "layer_norm_eps": 1e-05,
         | 
| 16 | 
            +
              "max_position_embeddings": 77,
         | 
| 17 | 
            +
              "model_type": "clip_text_model",
         | 
| 18 | 
            +
              "num_attention_heads": 12,
         | 
| 19 | 
            +
              "num_hidden_layers": 12,
         | 
| 20 | 
            +
              "pad_token_id": 1,
         | 
| 21 | 
            +
              "projection_dim": 768,
         | 
| 22 | 
            +
              "torch_dtype": "float16",
         | 
| 23 | 
            +
              "transformers_version": "4.41.2",
         | 
| 24 | 
            +
              "vocab_size": 49408
         | 
| 25 | 
            +
            }
         | 
    	
        text_encoder/model.onnx
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:a8138e5db16866d3ee0056b7d270b6488c479a37b67bd4439a96c9e77ccc5216
         | 
| 3 | 
            +
            size 246481009
         | 
    	
        tokenizer/merges.txt
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
    	
        tokenizer/model.onnx
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:63b7618c80a5bd8d1ee6fe92b28e7b72dde7aaa522963ff083f284501a9ec7df
         | 
| 3 | 
            +
            size 1683233
         | 
    	
        tokenizer/special_tokens_map.json
    ADDED
    
    | @@ -0,0 +1,24 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "bos_token": {
         | 
| 3 | 
            +
                "content": "<|startoftext|>",
         | 
| 4 | 
            +
                "lstrip": false,
         | 
| 5 | 
            +
                "normalized": true,
         | 
| 6 | 
            +
                "rstrip": false,
         | 
| 7 | 
            +
                "single_word": false
         | 
| 8 | 
            +
              },
         | 
| 9 | 
            +
              "eos_token": {
         | 
| 10 | 
            +
                "content": "<|endoftext|>",
         | 
| 11 | 
            +
                "lstrip": false,
         | 
| 12 | 
            +
                "normalized": true,
         | 
| 13 | 
            +
                "rstrip": false,
         | 
| 14 | 
            +
                "single_word": false
         | 
| 15 | 
            +
              },
         | 
| 16 | 
            +
              "pad_token": "<|endoftext|>",
         | 
| 17 | 
            +
              "unk_token": {
         | 
| 18 | 
            +
                "content": "<|endoftext|>",
         | 
| 19 | 
            +
                "lstrip": false,
         | 
| 20 | 
            +
                "normalized": true,
         | 
| 21 | 
            +
                "rstrip": false,
         | 
| 22 | 
            +
                "single_word": false
         | 
| 23 | 
            +
              }
         | 
| 24 | 
            +
            }
         | 
    	
        tokenizer/tokenizer_config.json
    ADDED
    
    | @@ -0,0 +1,30 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "add_prefix_space": false,
         | 
| 3 | 
            +
              "added_tokens_decoder": {
         | 
| 4 | 
            +
                "49406": {
         | 
| 5 | 
            +
                  "content": "<|startoftext|>",
         | 
| 6 | 
            +
                  "lstrip": false,
         | 
| 7 | 
            +
                  "normalized": true,
         | 
| 8 | 
            +
                  "rstrip": false,
         | 
| 9 | 
            +
                  "single_word": false,
         | 
| 10 | 
            +
                  "special": true
         | 
| 11 | 
            +
                },
         | 
| 12 | 
            +
                "49407": {
         | 
| 13 | 
            +
                  "content": "<|endoftext|>",
         | 
| 14 | 
            +
                  "lstrip": false,
         | 
| 15 | 
            +
                  "normalized": true,
         | 
| 16 | 
            +
                  "rstrip": false,
         | 
| 17 | 
            +
                  "single_word": false,
         | 
| 18 | 
            +
                  "special": true
         | 
| 19 | 
            +
                }
         | 
| 20 | 
            +
              },
         | 
| 21 | 
            +
              "bos_token": "<|startoftext|>",
         | 
| 22 | 
            +
              "clean_up_tokenization_spaces": true,
         | 
| 23 | 
            +
              "do_lower_case": true,
         | 
| 24 | 
            +
              "eos_token": "<|endoftext|>",
         | 
| 25 | 
            +
              "errors": "replace",
         | 
| 26 | 
            +
              "model_max_length": 77,
         | 
| 27 | 
            +
              "pad_token": "<|endoftext|>",
         | 
| 28 | 
            +
              "tokenizer_class": "CLIPTokenizer",
         | 
| 29 | 
            +
              "unk_token": "<|endoftext|>"
         | 
| 30 | 
            +
            }
         | 
    	
        tokenizer/vocab.json
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
    	
        unet/config.json
    ADDED
    
    | @@ -0,0 +1,67 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_class_name": "UNet2DConditionModel",
         | 
| 3 | 
            +
              "_diffusers_version": "0.30.0.dev0",
         | 
| 4 | 
            +
              "act_fn": "silu",
         | 
| 5 | 
            +
              "addition_embed_type": null,
         | 
| 6 | 
            +
              "addition_embed_type_num_heads": 64,
         | 
| 7 | 
            +
              "addition_time_embed_dim": null,
         | 
| 8 | 
            +
              "attention_head_dim": 8,
         | 
| 9 | 
            +
              "attention_type": "default",
         | 
| 10 | 
            +
              "block_out_channels": [
         | 
| 11 | 
            +
                320,
         | 
| 12 | 
            +
                640,
         | 
| 13 | 
            +
                1280,
         | 
| 14 | 
            +
                1280
         | 
| 15 | 
            +
              ],
         | 
| 16 | 
            +
              "center_input_sample": false,
         | 
| 17 | 
            +
              "class_embed_type": null,
         | 
| 18 | 
            +
              "class_embeddings_concat": false,
         | 
| 19 | 
            +
              "conv_in_kernel": 3,
         | 
| 20 | 
            +
              "conv_out_kernel": 3,
         | 
| 21 | 
            +
              "cross_attention_dim": 768,
         | 
| 22 | 
            +
              "cross_attention_norm": null,
         | 
| 23 | 
            +
              "down_block_types": [
         | 
| 24 | 
            +
                "CrossAttnDownBlock2D",
         | 
| 25 | 
            +
                "CrossAttnDownBlock2D",
         | 
| 26 | 
            +
                "CrossAttnDownBlock2D",
         | 
| 27 | 
            +
                "DownBlock2D"
         | 
| 28 | 
            +
              ],
         | 
| 29 | 
            +
              "downsample_padding": 1,
         | 
| 30 | 
            +
              "dropout": 0.0,
         | 
| 31 | 
            +
              "dual_cross_attention": false,
         | 
| 32 | 
            +
              "encoder_hid_dim": null,
         | 
| 33 | 
            +
              "encoder_hid_dim_type": null,
         | 
| 34 | 
            +
              "flip_sin_to_cos": true,
         | 
| 35 | 
            +
              "freq_shift": 0,
         | 
| 36 | 
            +
              "in_channels": 4,
         | 
| 37 | 
            +
              "layers_per_block": 2,
         | 
| 38 | 
            +
              "mid_block_only_cross_attention": null,
         | 
| 39 | 
            +
              "mid_block_scale_factor": 1,
         | 
| 40 | 
            +
              "mid_block_type": "UNetMidBlock2DCrossAttn",
         | 
| 41 | 
            +
              "norm_eps": 1e-05,
         | 
| 42 | 
            +
              "norm_num_groups": 32,
         | 
| 43 | 
            +
              "num_attention_heads": null,
         | 
| 44 | 
            +
              "num_class_embeds": null,
         | 
| 45 | 
            +
              "only_cross_attention": false,
         | 
| 46 | 
            +
              "out_channels": 4,
         | 
| 47 | 
            +
              "projection_class_embeddings_input_dim": null,
         | 
| 48 | 
            +
              "resnet_out_scale_factor": 1.0,
         | 
| 49 | 
            +
              "resnet_skip_time_act": false,
         | 
| 50 | 
            +
              "resnet_time_scale_shift": "default",
         | 
| 51 | 
            +
              "reverse_transformer_layers_per_block": null,
         | 
| 52 | 
            +
              "sample_size": 64,
         | 
| 53 | 
            +
              "time_cond_proj_dim": null,
         | 
| 54 | 
            +
              "time_embedding_act_fn": null,
         | 
| 55 | 
            +
              "time_embedding_dim": null,
         | 
| 56 | 
            +
              "time_embedding_type": "positional",
         | 
| 57 | 
            +
              "timestep_post_act": null,
         | 
| 58 | 
            +
              "transformer_layers_per_block": 1,
         | 
| 59 | 
            +
              "up_block_types": [
         | 
| 60 | 
            +
                "UpBlock2D",
         | 
| 61 | 
            +
                "CrossAttnUpBlock2D",
         | 
| 62 | 
            +
                "CrossAttnUpBlock2D",
         | 
| 63 | 
            +
                "CrossAttnUpBlock2D"
         | 
| 64 | 
            +
              ],
         | 
| 65 | 
            +
              "upcast_attention": false,
         | 
| 66 | 
            +
              "use_linear_projection": false
         | 
| 67 | 
            +
            }
         | 
    	
        unet/model.onnx
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:805a43795f1daf32387a0ae8081b4497752009e547c67e6b0d14e33491e44e2f
         | 
| 3 | 
            +
            size 1719492911
         | 
    	
        vae_decoder/config.json
    ADDED
    
    | @@ -0,0 +1,36 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_class_name": "AutoencoderKL",
         | 
| 3 | 
            +
              "_diffusers_version": "0.30.0.dev0",
         | 
| 4 | 
            +
              "act_fn": "silu",
         | 
| 5 | 
            +
              "block_out_channels": [
         | 
| 6 | 
            +
                128,
         | 
| 7 | 
            +
                256,
         | 
| 8 | 
            +
                512,
         | 
| 9 | 
            +
                512
         | 
| 10 | 
            +
              ],
         | 
| 11 | 
            +
              "down_block_types": [
         | 
| 12 | 
            +
                "DownEncoderBlock2D",
         | 
| 13 | 
            +
                "DownEncoderBlock2D",
         | 
| 14 | 
            +
                "DownEncoderBlock2D",
         | 
| 15 | 
            +
                "DownEncoderBlock2D"
         | 
| 16 | 
            +
              ],
         | 
| 17 | 
            +
              "force_upcast": true,
         | 
| 18 | 
            +
              "in_channels": 3,
         | 
| 19 | 
            +
              "latent_channels": 4,
         | 
| 20 | 
            +
              "latents_mean": null,
         | 
| 21 | 
            +
              "latents_std": null,
         | 
| 22 | 
            +
              "layers_per_block": 2,
         | 
| 23 | 
            +
              "norm_num_groups": 32,
         | 
| 24 | 
            +
              "out_channels": 3,
         | 
| 25 | 
            +
              "sample_size": 512,
         | 
| 26 | 
            +
              "scaling_factor": 0.18215,
         | 
| 27 | 
            +
              "shift_factor": null,
         | 
| 28 | 
            +
              "up_block_types": [
         | 
| 29 | 
            +
                "UpDecoderBlock2D",
         | 
| 30 | 
            +
                "UpDecoderBlock2D",
         | 
| 31 | 
            +
                "UpDecoderBlock2D",
         | 
| 32 | 
            +
                "UpDecoderBlock2D"
         | 
| 33 | 
            +
              ],
         | 
| 34 | 
            +
              "use_post_quant_conv": true,
         | 
| 35 | 
            +
              "use_quant_conv": true
         | 
| 36 | 
            +
            }
         | 
    	
        vae_decoder/model.onnx
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:9a85320a01f6e4b2f219984a5a5488964402dcbe01d629e7f43320ad12f2534a
         | 
| 3 | 
            +
            size 99039207
         | 
    	
        vae_encoder/config.json
    ADDED
    
    | @@ -0,0 +1,36 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_class_name": "AutoencoderKL",
         | 
| 3 | 
            +
              "_diffusers_version": "0.30.0.dev0",
         | 
| 4 | 
            +
              "act_fn": "silu",
         | 
| 5 | 
            +
              "block_out_channels": [
         | 
| 6 | 
            +
                128,
         | 
| 7 | 
            +
                256,
         | 
| 8 | 
            +
                512,
         | 
| 9 | 
            +
                512
         | 
| 10 | 
            +
              ],
         | 
| 11 | 
            +
              "down_block_types": [
         | 
| 12 | 
            +
                "DownEncoderBlock2D",
         | 
| 13 | 
            +
                "DownEncoderBlock2D",
         | 
| 14 | 
            +
                "DownEncoderBlock2D",
         | 
| 15 | 
            +
                "DownEncoderBlock2D"
         | 
| 16 | 
            +
              ],
         | 
| 17 | 
            +
              "force_upcast": true,
         | 
| 18 | 
            +
              "in_channels": 3,
         | 
| 19 | 
            +
              "latent_channels": 4,
         | 
| 20 | 
            +
              "latents_mean": null,
         | 
| 21 | 
            +
              "latents_std": null,
         | 
| 22 | 
            +
              "layers_per_block": 2,
         | 
| 23 | 
            +
              "norm_num_groups": 32,
         | 
| 24 | 
            +
              "out_channels": 3,
         | 
| 25 | 
            +
              "sample_size": 512,
         | 
| 26 | 
            +
              "scaling_factor": 0.18215,
         | 
| 27 | 
            +
              "shift_factor": null,
         | 
| 28 | 
            +
              "up_block_types": [
         | 
| 29 | 
            +
                "UpDecoderBlock2D",
         | 
| 30 | 
            +
                "UpDecoderBlock2D",
         | 
| 31 | 
            +
                "UpDecoderBlock2D",
         | 
| 32 | 
            +
                "UpDecoderBlock2D"
         | 
| 33 | 
            +
              ],
         | 
| 34 | 
            +
              "use_post_quant_conv": true,
         | 
| 35 | 
            +
              "use_quant_conv": true
         | 
| 36 | 
            +
            }
         | 
    	
        vae_encoder/model.onnx
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:5af0eec9da957ab4d2de98a620e088633fe7cb4599833f7f7de4b9a3b7f14736
         | 
| 3 | 
            +
            size 68391734
         | 

