Adding Neuron-optimized model files

#274
by badaoui HF Staff - opened
.gitattributes CHANGED
@@ -29,4 +29,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
29
  *.zip filter=lfs diff=lfs merge=lfs -text
30
  *.zst filter=lfs diff=lfs merge=lfs -text
31
  *tfevents* filter=lfs diff=lfs merge=lfs -text
32
- *.safetensors filter=lfs diff=lfs merge=lfs -text
 
 
 
 
29
  *.zip filter=lfs diff=lfs merge=lfs -text
30
  *.zst filter=lfs diff=lfs merge=lfs -text
31
  *tfevents* filter=lfs diff=lfs merge=lfs -text
32
+ *.safetensors filter=lfs diff=lfs merge=lfs -textvae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
33
+ text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
34
+ unet/model.neuron filter=lfs diff=lfs merge=lfs -text
35
+ vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -4,26 +4,29 @@ tags:
4
  - stable-diffusion
5
  - stable-diffusion-diffusers
6
  - text-to-image
 
7
  widget:
8
- - text: "A high tech solarpunk utopia in the Amazon rainforest"
9
  example_title: Amazon rainforest
10
- - text: "A pikachu fine dining with a view to the Eiffel Tower"
11
  example_title: Pikachu in Paris
12
- - text: "A mecha robot in a favela in expressionist style"
13
  example_title: Expressionist robot
14
- - text: "an insect robot preparing a delicious meal"
15
  example_title: Insect robot
16
- - text: "A small cabin on top of a snowy mountain in the style of Disney, artstation"
17
  example_title: Snowy disney cabin
18
- extra_gated_prompt: |-
19
- This model is open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage.
20
- The CreativeML OpenRAIL License specifies:
21
-
22
- 1. You can't use the model to deliberately produce nor share illegal or harmful outputs or content
23
- 2. The authors claim no rights on the outputs you generate, you are free to use them and are accountable for their use which must not go against the provisions set in the license
24
- 3. You may re-distribute the weights and use the model commercially and/or as a service. If you do, please be aware you have to include the same use restrictions as the ones in the license and share a copy of the CreativeML OpenRAIL-M to all your users (please read the license entirely and carefully)
25
- Please read the full license carefully here: https://huggingface.co/spaces/CompVis/stable-diffusion-license
26
-
 
 
27
  extra_gated_heading: Please read the LICENSE to access this model
28
  ---
29
 
 
4
  - stable-diffusion
5
  - stable-diffusion-diffusers
6
  - text-to-image
7
+ - neuron
8
  widget:
9
+ - text: A high tech solarpunk utopia in the Amazon rainforest
10
  example_title: Amazon rainforest
11
+ - text: A pikachu fine dining with a view to the Eiffel Tower
12
  example_title: Pikachu in Paris
13
+ - text: A mecha robot in a favela in expressionist style
14
  example_title: Expressionist robot
15
+ - text: an insect robot preparing a delicious meal
16
  example_title: Insect robot
17
+ - text: A small cabin on top of a snowy mountain in the style of Disney, artstation
18
  example_title: Snowy disney cabin
19
+ extra_gated_prompt: "This model is open access and available to all, with a CreativeML\
20
+ \ OpenRAIL-M license further specifying rights and usage.\nThe CreativeML OpenRAIL\
21
+ \ License specifies: \n\n1. You can't use the model to deliberately produce nor\
22
+ \ share illegal or harmful outputs or content \n2. The authors claim no rights on\
23
+ \ the outputs you generate, you are free to use them and are accountable for their\
24
+ \ use which must not go against the provisions set in the license\n3. You may re-distribute\
25
+ \ the weights and use the model commercially and/or as a service. If you do, please\
26
+ \ be aware you have to include the same use restrictions as the ones in the license\
27
+ \ and share a copy of the CreativeML OpenRAIL-M to all your users (please read the\
28
+ \ license entirely and carefully)\nPlease read the full license carefully here:\
29
+ \ https://huggingface.co/spaces/CompVis/stable-diffusion-license\n "
30
  extra_gated_heading: Please read the LICENSE to access this model
31
  ---
32
 
feature_extractor/preprocessor_config.json CHANGED
@@ -1,20 +1,27 @@
1
  {
2
- "crop_size": 224,
 
 
 
3
  "do_center_crop": true,
4
  "do_convert_rgb": true,
5
  "do_normalize": true,
 
6
  "do_resize": true,
7
- "feature_extractor_type": "CLIPFeatureExtractor",
8
  "image_mean": [
9
  0.48145466,
10
  0.4578275,
11
  0.40821073
12
  ],
 
13
  "image_std": [
14
  0.26862954,
15
  0.26130258,
16
  0.27577711
17
  ],
18
  "resample": 3,
19
- "size": 224
 
 
 
20
  }
 
1
  {
2
+ "crop_size": {
3
+ "height": 224,
4
+ "width": 224
5
+ },
6
  "do_center_crop": true,
7
  "do_convert_rgb": true,
8
  "do_normalize": true,
9
+ "do_rescale": true,
10
  "do_resize": true,
 
11
  "image_mean": [
12
  0.48145466,
13
  0.4578275,
14
  0.40821073
15
  ],
16
+ "image_processor_type": "CLIPImageProcessor",
17
  "image_std": [
18
  0.26862954,
19
  0.26130258,
20
  0.27577711
21
  ],
22
  "resample": 3,
23
+ "rescale_factor": 0.00392156862745098,
24
+ "size": {
25
+ "shortest_edge": 224
26
+ }
27
  }
model_index.json CHANGED
@@ -1,32 +1,39 @@
1
  {
2
- "_class_name": "StableDiffusionPipeline",
3
- "_diffusers_version": "0.2.2",
4
  "feature_extractor": [
5
  "transformers",
6
  "CLIPImageProcessor"
7
  ],
 
 
 
 
 
 
 
8
  "safety_checker": [
9
- "stable_diffusion",
10
- "StableDiffusionSafetyChecker"
11
  ],
12
  "scheduler": [
13
  "diffusers",
14
  "PNDMScheduler"
15
  ],
16
  "text_encoder": [
17
- "transformers",
18
- "CLIPTextModel"
19
  ],
20
  "tokenizer": [
21
  "transformers",
22
  "CLIPTokenizer"
23
  ],
24
  "unet": [
25
- "diffusers",
26
- "UNet2DConditionModel"
27
  ],
28
  "vae": [
29
- "diffusers",
30
- "AutoencoderKL"
31
  ]
32
  }
 
1
  {
2
+ "_class_name": "NeuronStableDiffusionPipeline",
3
+ "_diffusers_version": "0.34.0",
4
  "feature_extractor": [
5
  "transformers",
6
  "CLIPImageProcessor"
7
  ],
8
+ "force_zeros_for_empty_prompt": true,
9
+ "image_encoder": [
10
+ null,
11
+ null
12
+ ],
13
+ "requires_aesthetics_score": false,
14
+ "requires_safety_checker": true,
15
  "safety_checker": [
16
+ null,
17
+ null
18
  ],
19
  "scheduler": [
20
  "diffusers",
21
  "PNDMScheduler"
22
  ],
23
  "text_encoder": [
24
+ "optimum.neuron.modeling_diffusion",
25
+ "NeuronModelTextEncoder"
26
  ],
27
  "tokenizer": [
28
  "transformers",
29
  "CLIPTokenizer"
30
  ],
31
  "unet": [
32
+ "optimum.neuron.modeling_diffusion",
33
+ "NeuronModelUnet"
34
  ],
35
  "vae": [
36
+ "optimum.neuron.modeling_diffusion",
37
+ "NeuronModelVae"
38
  ]
39
  }
scheduler/scheduler_config.json CHANGED
@@ -1,13 +1,15 @@
1
  {
2
  "_class_name": "PNDMScheduler",
3
- "_diffusers_version": "0.7.0.dev0",
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
 
7
  "num_train_timesteps": 1000,
 
8
  "set_alpha_to_one": false,
9
  "skip_prk_steps": true,
10
  "steps_offset": 1,
11
- "trained_betas": null,
12
- "clip_sample": false
13
  }
 
1
  {
2
  "_class_name": "PNDMScheduler",
3
+ "_diffusers_version": "0.34.0",
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
7
+ "clip_sample": false,
8
  "num_train_timesteps": 1000,
9
+ "prediction_type": "epsilon",
10
  "set_alpha_to_one": false,
11
  "skip_prk_steps": true,
12
  "steps_offset": 1,
13
+ "timestep_spacing": "leading",
14
+ "trained_betas": null
15
  }
text_encoder/config.json CHANGED
@@ -1,5 +1,6 @@
1
  {
2
- "_name_or_path": "openai/clip-vit-large-patch14",
 
3
  "architectures": [
4
  "CLIPTextModel"
5
  ],
@@ -15,10 +16,35 @@
15
  "layer_norm_eps": 1e-05,
16
  "max_position_embeddings": 77,
17
  "model_type": "clip_text_model",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  "num_attention_heads": 12,
19
  "num_hidden_layers": 12,
20
  "pad_token_id": 1,
 
 
21
  "torch_dtype": "float32",
22
- "transformers_version": "4.21.0.dev0",
 
23
  "vocab_size": 49408
24
  }
 
1
  {
2
+ "_attn_implementation_autoset": true,
3
+ "_name_or_path": "/home/user/.cache/huggingface/hub/models--CompVis--stable-diffusion-v1-4/snapshots/133a221b8aa7292a167afc5127cb63fb5005638b/text_encoder",
4
  "architectures": [
5
  "CLIPTextModel"
6
  ],
 
16
  "layer_norm_eps": 1e-05,
17
  "max_position_embeddings": 77,
18
  "model_type": "clip_text_model",
19
+ "neuron": {
20
+ "auto_cast": "matmul",
21
+ "auto_cast_type": "bf16",
22
+ "compiler_type": "neuronx-cc",
23
+ "compiler_version": "2.15.143.0+e39249ad",
24
+ "dynamic_batch_size": false,
25
+ "inline_weights_to_neff": true,
26
+ "input_names": [
27
+ "input_ids"
28
+ ],
29
+ "model_type": "clip-text-model",
30
+ "optlevel": "2",
31
+ "output_attentions": false,
32
+ "output_hidden_states": false,
33
+ "output_names": [
34
+ "last_hidden_state",
35
+ "pooler_output"
36
+ ],
37
+ "static_batch_size": 1,
38
+ "static_sequence_length": 77,
39
+ "tensor_parallel_size": 1
40
+ },
41
  "num_attention_heads": 12,
42
  "num_hidden_layers": 12,
43
  "pad_token_id": 1,
44
+ "projection_dim": 512,
45
+ "task": "feature-extraction",
46
  "torch_dtype": "float32",
47
+ "torchscript": true,
48
+ "transformers_version": "4.48.3",
49
  "vocab_size": 49408
50
  }
text_encoder/model.neuron ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8a5032ffebf9b052ab36054209ecd3d360f2646ae7ea6527c6a2f063b1822f0
3
+ size 375660151
tokenizer/special_tokens_map.json CHANGED
@@ -13,7 +13,13 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": "<|endoftext|>",
 
 
 
 
 
 
17
  "unk_token": {
18
  "content": "<|endoftext|>",
19
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": {
17
+ "content": "<|endoftext|>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
  "unk_token": {
24
  "content": "<|endoftext|>",
25
  "lstrip": false,
tokenizer/tokenizer_config.json CHANGED
@@ -1,34 +1,31 @@
1
  {
2
  "add_prefix_space": false,
3
- "bos_token": {
4
- "__type": "AddedToken",
5
- "content": "<|startoftext|>",
6
- "lstrip": false,
7
- "normalized": true,
8
- "rstrip": false,
9
- "single_word": false
 
 
 
 
 
 
 
 
 
 
10
  },
 
 
11
  "do_lower_case": true,
12
- "eos_token": {
13
- "__type": "AddedToken",
14
- "content": "<|endoftext|>",
15
- "lstrip": false,
16
- "normalized": true,
17
- "rstrip": false,
18
- "single_word": false
19
- },
20
  "errors": "replace",
 
21
  "model_max_length": 77,
22
- "name_or_path": "openai/clip-vit-large-patch14",
23
  "pad_token": "<|endoftext|>",
24
- "special_tokens_map_file": "./special_tokens_map.json",
25
  "tokenizer_class": "CLIPTokenizer",
26
- "unk_token": {
27
- "__type": "AddedToken",
28
- "content": "<|endoftext|>",
29
- "lstrip": false,
30
- "normalized": true,
31
- "rstrip": false,
32
- "single_word": false
33
- }
34
  }
 
1
  {
2
  "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "49406": {
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "49407": {
13
+ "content": "<|endoftext|>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ }
20
  },
21
+ "bos_token": "<|startoftext|>",
22
+ "clean_up_tokenization_spaces": false,
23
  "do_lower_case": true,
24
+ "eos_token": "<|endoftext|>",
 
 
 
 
 
 
 
25
  "errors": "replace",
26
+ "extra_special_tokens": {},
27
  "model_max_length": 77,
 
28
  "pad_token": "<|endoftext|>",
 
29
  "tokenizer_class": "CLIPTokenizer",
30
+ "unk_token": "<|endoftext|>"
 
 
 
 
 
 
 
31
  }
unet/config.json CHANGED
@@ -1,8 +1,47 @@
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
- "_diffusers_version": "0.2.2",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  "act_fn": "silu",
 
 
 
5
  "attention_head_dim": 8,
 
6
  "block_out_channels": [
7
  320,
8
  640,
@@ -10,7 +49,12 @@
10
  1280
11
  ],
12
  "center_input_sample": false,
 
 
 
 
13
  "cross_attention_dim": 768,
 
14
  "down_block_types": [
15
  "CrossAttnDownBlock2D",
16
  "CrossAttnDownBlock2D",
@@ -18,19 +62,70 @@
18
  "DownBlock2D"
19
  ],
20
  "downsample_padding": 1,
 
 
 
 
21
  "flip_sin_to_cos": true,
22
  "freq_shift": 0,
23
  "in_channels": 4,
24
  "layers_per_block": 2,
 
25
  "mid_block_scale_factor": 1,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  "norm_eps": 1e-05,
27
  "norm_num_groups": 32,
 
 
 
28
  "out_channels": 4,
 
 
 
 
 
29
  "sample_size": 64,
 
 
 
 
 
 
 
 
30
  "up_block_types": [
31
  "UpBlock2D",
32
  "CrossAttnUpBlock2D",
33
  "CrossAttnUpBlock2D",
34
  "CrossAttnUpBlock2D"
35
- ]
 
 
36
  }
 
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
+ "_commit_hash": null,
4
+ "_diffusers_version": "0.34.0",
5
+ "_name_or_path": "/home/user/.cache/huggingface/hub/models--CompVis--stable-diffusion-v1-4/snapshots/133a221b8aa7292a167afc5127cb63fb5005638b/unet",
6
+ "_use_default_values": [
7
+ "cross_attention_norm",
8
+ "dropout",
9
+ "timestep_post_act",
10
+ "time_cond_proj_dim",
11
+ "mid_block_type",
12
+ "resnet_time_scale_shift",
13
+ "class_embeddings_concat",
14
+ "time_embedding_act_fn",
15
+ "mid_block_only_cross_attention",
16
+ "conv_in_kernel",
17
+ "conv_out_kernel",
18
+ "reverse_transformer_layers_per_block",
19
+ "time_embedding_type",
20
+ "num_attention_heads",
21
+ "addition_time_embed_dim",
22
+ "upcast_attention",
23
+ "time_embedding_dim",
24
+ "addition_embed_type",
25
+ "addition_embed_type_num_heads",
26
+ "resnet_skip_time_act",
27
+ "only_cross_attention",
28
+ "dual_cross_attention",
29
+ "attention_type",
30
+ "transformer_layers_per_block",
31
+ "encoder_hid_dim_type",
32
+ "encoder_hid_dim",
33
+ "class_embed_type",
34
+ "projection_class_embeddings_input_dim",
35
+ "use_linear_projection",
36
+ "num_class_embeds",
37
+ "resnet_out_scale_factor"
38
+ ],
39
  "act_fn": "silu",
40
+ "addition_embed_type": null,
41
+ "addition_embed_type_num_heads": 64,
42
+ "addition_time_embed_dim": null,
43
  "attention_head_dim": 8,
44
+ "attention_type": "default",
45
  "block_out_channels": [
46
  320,
47
  640,
 
49
  1280
50
  ],
51
  "center_input_sample": false,
52
+ "class_embed_type": null,
53
+ "class_embeddings_concat": false,
54
+ "conv_in_kernel": 3,
55
+ "conv_out_kernel": 3,
56
  "cross_attention_dim": 768,
57
+ "cross_attention_norm": null,
58
  "down_block_types": [
59
  "CrossAttnDownBlock2D",
60
  "CrossAttnDownBlock2D",
 
62
  "DownBlock2D"
63
  ],
64
  "downsample_padding": 1,
65
+ "dropout": 0.0,
66
+ "dual_cross_attention": false,
67
+ "encoder_hid_dim": null,
68
+ "encoder_hid_dim_type": null,
69
  "flip_sin_to_cos": true,
70
  "freq_shift": 0,
71
  "in_channels": 4,
72
  "layers_per_block": 2,
73
+ "mid_block_only_cross_attention": null,
74
  "mid_block_scale_factor": 1,
75
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
76
+ "neuron": {
77
+ "auto_cast": "matmul",
78
+ "auto_cast_type": "bf16",
79
+ "compiler_type": "neuronx-cc",
80
+ "compiler_version": "2.15.143.0+e39249ad",
81
+ "dynamic_batch_size": false,
82
+ "inline_weights_to_neff": true,
83
+ "input_names": [
84
+ "sample",
85
+ "timestep",
86
+ "encoder_hidden_states"
87
+ ],
88
+ "model_type": "unet",
89
+ "optlevel": "2",
90
+ "output_attentions": false,
91
+ "output_hidden_states": false,
92
+ "output_names": [
93
+ "sample"
94
+ ],
95
+ "static_batch_size": 1,
96
+ "static_height": 64,
97
+ "static_num_channels": 4,
98
+ "static_sequence_length": 77,
99
+ "static_vae_scale_factor": 8,
100
+ "static_width": 64,
101
+ "tensor_parallel_size": 1
102
+ },
103
  "norm_eps": 1e-05,
104
  "norm_num_groups": 32,
105
+ "num_attention_heads": null,
106
+ "num_class_embeds": null,
107
+ "only_cross_attention": false,
108
  "out_channels": 4,
109
+ "projection_class_embeddings_input_dim": null,
110
+ "resnet_out_scale_factor": 1.0,
111
+ "resnet_skip_time_act": false,
112
+ "resnet_time_scale_shift": "default",
113
+ "reverse_transformer_layers_per_block": null,
114
  "sample_size": 64,
115
+ "task": "semantic-segmentation",
116
+ "time_cond_proj_dim": null,
117
+ "time_embedding_act_fn": null,
118
+ "time_embedding_dim": null,
119
+ "time_embedding_type": "positional",
120
+ "timestep_post_act": null,
121
+ "transformer_layers_per_block": 1,
122
+ "transformers_version": null,
123
  "up_block_types": [
124
  "UpBlock2D",
125
  "CrossAttnUpBlock2D",
126
  "CrossAttnUpBlock2D",
127
  "CrossAttnUpBlock2D"
128
+ ],
129
+ "upcast_attention": false,
130
+ "use_linear_projection": false
131
  }
unet/model.neuron ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d3c33c39c4dbf4fd193e6bb4c8a09da946bc299b5061d30437c086ef0bd9569
3
+ size 1900028915
vae_decoder/config.json ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_commit_hash": null,
4
+ "_diffusers_version": "0.34.0",
5
+ "_name_or_path": "/home/user/.cache/huggingface/hub/models--CompVis--stable-diffusion-v1-4/snapshots/133a221b8aa7292a167afc5127cb63fb5005638b/vae",
6
+ "_use_default_values": [
7
+ "norm_num_groups",
8
+ "force_upcast",
9
+ "use_post_quant_conv",
10
+ "latents_std",
11
+ "use_quant_conv",
12
+ "mid_block_add_attention",
13
+ "latents_mean",
14
+ "shift_factor"
15
+ ],
16
+ "act_fn": "silu",
17
+ "block_out_channels": [
18
+ 128,
19
+ 256,
20
+ 512,
21
+ 512
22
+ ],
23
+ "down_block_types": [
24
+ "DownEncoderBlock2D",
25
+ "DownEncoderBlock2D",
26
+ "DownEncoderBlock2D",
27
+ "DownEncoderBlock2D"
28
+ ],
29
+ "force_upcast": true,
30
+ "in_channels": 3,
31
+ "latent_channels": 4,
32
+ "latents_mean": null,
33
+ "latents_std": null,
34
+ "layers_per_block": 2,
35
+ "mid_block_add_attention": true,
36
+ "neuron": {
37
+ "auto_cast": "matmul",
38
+ "auto_cast_type": "bf16",
39
+ "compiler_type": "neuronx-cc",
40
+ "compiler_version": "2.15.143.0+e39249ad",
41
+ "dynamic_batch_size": false,
42
+ "inline_weights_to_neff": true,
43
+ "input_names": [
44
+ "latent_sample"
45
+ ],
46
+ "model_type": "vae-decoder",
47
+ "optlevel": "2",
48
+ "output_attentions": false,
49
+ "output_hidden_states": false,
50
+ "output_names": [
51
+ "sample"
52
+ ],
53
+ "static_batch_size": 1,
54
+ "static_height": 64,
55
+ "static_num_channels": 4,
56
+ "static_width": 64,
57
+ "tensor_parallel_size": 1
58
+ },
59
+ "norm_num_groups": 32,
60
+ "out_channels": 3,
61
+ "sample_size": 512,
62
+ "scaling_factor": 0.18215,
63
+ "shift_factor": null,
64
+ "task": "semantic-segmentation",
65
+ "transformers_version": null,
66
+ "up_block_types": [
67
+ "UpDecoderBlock2D",
68
+ "UpDecoderBlock2D",
69
+ "UpDecoderBlock2D",
70
+ "UpDecoderBlock2D"
71
+ ],
72
+ "use_post_quant_conv": true,
73
+ "use_quant_conv": true
74
+ }
vae_decoder/model.neuron ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4157adfc4fb887890bcc46dce7b0a6cb352f3cf96f3fd528d4fb48c75605c5c1
3
+ size 247896819
vae_encoder/config.json ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_commit_hash": null,
4
+ "_diffusers_version": "0.34.0",
5
+ "_name_or_path": "/home/user/.cache/huggingface/hub/models--CompVis--stable-diffusion-v1-4/snapshots/133a221b8aa7292a167afc5127cb63fb5005638b/vae",
6
+ "_use_default_values": [
7
+ "norm_num_groups",
8
+ "force_upcast",
9
+ "use_post_quant_conv",
10
+ "latents_std",
11
+ "use_quant_conv",
12
+ "mid_block_add_attention",
13
+ "latents_mean",
14
+ "shift_factor"
15
+ ],
16
+ "act_fn": "silu",
17
+ "block_out_channels": [
18
+ 128,
19
+ 256,
20
+ 512,
21
+ 512
22
+ ],
23
+ "down_block_types": [
24
+ "DownEncoderBlock2D",
25
+ "DownEncoderBlock2D",
26
+ "DownEncoderBlock2D",
27
+ "DownEncoderBlock2D"
28
+ ],
29
+ "force_upcast": true,
30
+ "in_channels": 3,
31
+ "latent_channels": 4,
32
+ "latents_mean": null,
33
+ "latents_std": null,
34
+ "layers_per_block": 2,
35
+ "mid_block_add_attention": true,
36
+ "neuron": {
37
+ "auto_cast": "matmul",
38
+ "auto_cast_type": "bf16",
39
+ "compiler_type": "neuronx-cc",
40
+ "compiler_version": "2.15.143.0+e39249ad",
41
+ "dynamic_batch_size": false,
42
+ "inline_weights_to_neff": true,
43
+ "input_names": [
44
+ "sample"
45
+ ],
46
+ "model_type": "vae-encoder",
47
+ "optlevel": "2",
48
+ "output_attentions": false,
49
+ "output_hidden_states": false,
50
+ "output_names": [
51
+ "latent_parameters"
52
+ ],
53
+ "static_batch_size": 1,
54
+ "static_height": 512,
55
+ "static_num_channels": 3,
56
+ "static_width": 512,
57
+ "tensor_parallel_size": 1
58
+ },
59
+ "norm_num_groups": 32,
60
+ "out_channels": 3,
61
+ "sample_size": 512,
62
+ "scaling_factor": 0.18215,
63
+ "shift_factor": null,
64
+ "task": "semantic-segmentation",
65
+ "transformers_version": null,
66
+ "up_block_types": [
67
+ "UpDecoderBlock2D",
68
+ "UpDecoderBlock2D",
69
+ "UpDecoderBlock2D",
70
+ "UpDecoderBlock2D"
71
+ ],
72
+ "use_post_quant_conv": true,
73
+ "use_quant_conv": true
74
+ }
vae_encoder/model.neuron ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:702c82996a50c462cdbe251483af14f8fa55541c2e89f1735b3387773c162892
3
+ size 155910897