badaoui HF Staff commited on
Commit
6bb235f
·
verified ·
1 Parent(s): 133a221

Adding Neuron-optimized model files

Browse files

🤖 Neuron Export Bot: On behalf of [badaoui](https://huggingface.co/badaoui), adding AWS Neuron-optimized model files.
Neuron-optimized models can achieve high-performance inference on AWS Inferentia and Trainium chips. Learn more:
- [AWS Neuron Documentation](https://awsdocs-neuron.readthedocs-hosted.com)
- [🤗 Optimum Neuron Guide](https://huggingface.co/docs/optimum-neuron/index)

.gitattributes CHANGED
@@ -29,4 +29,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
29
  *.zip filter=lfs diff=lfs merge=lfs -text
30
  *.zst filter=lfs diff=lfs merge=lfs -text
31
  *tfevents* filter=lfs diff=lfs merge=lfs -text
32
- *.safetensors filter=lfs diff=lfs merge=lfs -text
 
 
 
 
29
  *.zip filter=lfs diff=lfs merge=lfs -text
30
  *.zst filter=lfs diff=lfs merge=lfs -text
31
  *tfevents* filter=lfs diff=lfs merge=lfs -text
32
+ *.safetensors filter=lfs diff=lfs merge=lfs -textneuron/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
33
+ neuron/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
34
+ neuron/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
35
+ neuron/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -4,26 +4,29 @@ tags:
4
  - stable-diffusion
5
  - stable-diffusion-diffusers
6
  - text-to-image
 
7
  widget:
8
- - text: "A high tech solarpunk utopia in the Amazon rainforest"
9
  example_title: Amazon rainforest
10
- - text: "A pikachu fine dining with a view to the Eiffel Tower"
11
  example_title: Pikachu in Paris
12
- - text: "A mecha robot in a favela in expressionist style"
13
  example_title: Expressionist robot
14
- - text: "an insect robot preparing a delicious meal"
15
  example_title: Insect robot
16
- - text: "A small cabin on top of a snowy mountain in the style of Disney, artstation"
17
  example_title: Snowy disney cabin
18
- extra_gated_prompt: |-
19
- This model is open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage.
20
- The CreativeML OpenRAIL License specifies:
21
-
22
- 1. You can't use the model to deliberately produce nor share illegal or harmful outputs or content
23
- 2. The authors claim no rights on the outputs you generate, you are free to use them and are accountable for their use which must not go against the provisions set in the license
24
- 3. You may re-distribute the weights and use the model commercially and/or as a service. If you do, please be aware you have to include the same use restrictions as the ones in the license and share a copy of the CreativeML OpenRAIL-M to all your users (please read the license entirely and carefully)
25
- Please read the full license carefully here: https://huggingface.co/spaces/CompVis/stable-diffusion-license
26
-
 
 
27
  extra_gated_heading: Please read the LICENSE to access this model
28
  ---
29
 
 
4
  - stable-diffusion
5
  - stable-diffusion-diffusers
6
  - text-to-image
7
+ - neuron
8
  widget:
9
+ - text: A high tech solarpunk utopia in the Amazon rainforest
10
  example_title: Amazon rainforest
11
+ - text: A pikachu fine dining with a view to the Eiffel Tower
12
  example_title: Pikachu in Paris
13
+ - text: A mecha robot in a favela in expressionist style
14
  example_title: Expressionist robot
15
+ - text: an insect robot preparing a delicious meal
16
  example_title: Insect robot
17
+ - text: A small cabin on top of a snowy mountain in the style of Disney, artstation
18
  example_title: Snowy disney cabin
19
+ extra_gated_prompt: "This model is open access and available to all, with a CreativeML\
20
+ \ OpenRAIL-M license further specifying rights and usage.\nThe CreativeML OpenRAIL\
21
+ \ License specifies: \n\n1. You can't use the model to deliberately produce nor\
22
+ \ share illegal or harmful outputs or content \n2. The authors claim no rights on\
23
+ \ the outputs you generate, you are free to use them and are accountable for their\
24
+ \ use which must not go against the provisions set in the license\n3. You may re-distribute\
25
+ \ the weights and use the model commercially and/or as a service. If you do, please\
26
+ \ be aware you have to include the same use restrictions as the ones in the license\
27
+ \ and share a copy of the CreativeML OpenRAIL-M to all your users (please read the\
28
+ \ license entirely and carefully)\nPlease read the full license carefully here:\
29
+ \ https://huggingface.co/spaces/CompVis/stable-diffusion-license\n "
30
  extra_gated_heading: Please read the LICENSE to access this model
31
  ---
32
 
neuron/feature_extractor/preprocessor_config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": {
3
+ "height": 224,
4
+ "width": 224
5
+ },
6
+ "do_center_crop": true,
7
+ "do_convert_rgb": true,
8
+ "do_normalize": true,
9
+ "do_rescale": true,
10
+ "do_resize": true,
11
+ "image_mean": [
12
+ 0.48145466,
13
+ 0.4578275,
14
+ 0.40821073
15
+ ],
16
+ "image_processor_type": "CLIPImageProcessor",
17
+ "image_std": [
18
+ 0.26862954,
19
+ 0.26130258,
20
+ 0.27577711
21
+ ],
22
+ "resample": 3,
23
+ "rescale_factor": 0.00392156862745098,
24
+ "size": {
25
+ "shortest_edge": 224
26
+ }
27
+ }
neuron/model_index.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "NeuronStableDiffusionPipeline",
3
+ "_diffusers_version": "0.34.0",
4
+ "feature_extractor": [
5
+ "transformers",
6
+ "CLIPImageProcessor"
7
+ ],
8
+ "force_zeros_for_empty_prompt": true,
9
+ "image_encoder": [
10
+ null,
11
+ null
12
+ ],
13
+ "requires_aesthetics_score": false,
14
+ "requires_safety_checker": true,
15
+ "safety_checker": [
16
+ null,
17
+ null
18
+ ],
19
+ "scheduler": [
20
+ "diffusers",
21
+ "PNDMScheduler"
22
+ ],
23
+ "text_encoder": [
24
+ "optimum.neuron.modeling_diffusion",
25
+ "NeuronModelTextEncoder"
26
+ ],
27
+ "tokenizer": [
28
+ "transformers",
29
+ "CLIPTokenizer"
30
+ ],
31
+ "unet": [
32
+ "optimum.neuron.modeling_diffusion",
33
+ "NeuronModelUnet"
34
+ ],
35
+ "vae": [
36
+ "optimum.neuron.modeling_diffusion",
37
+ "NeuronModelVae"
38
+ ]
39
+ }
neuron/scheduler/scheduler_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "PNDMScheduler",
3
+ "_diffusers_version": "0.34.0",
4
+ "beta_end": 0.012,
5
+ "beta_schedule": "scaled_linear",
6
+ "beta_start": 0.00085,
7
+ "clip_sample": false,
8
+ "num_train_timesteps": 1000,
9
+ "prediction_type": "epsilon",
10
+ "set_alpha_to_one": false,
11
+ "skip_prk_steps": true,
12
+ "steps_offset": 1,
13
+ "timestep_spacing": "leading",
14
+ "trained_betas": null
15
+ }
neuron/text_encoder/config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "_name_or_path": "/home/user/.cache/huggingface/hub/models--CompVis--stable-diffusion-v1-4/snapshots/133a221b8aa7292a167afc5127cb63fb5005638b/text_encoder",
4
+ "architectures": [
5
+ "CLIPTextModel"
6
+ ],
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 0,
9
+ "dropout": 0.0,
10
+ "eos_token_id": 2,
11
+ "hidden_act": "quick_gelu",
12
+ "hidden_size": 768,
13
+ "initializer_factor": 1.0,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
+ "layer_norm_eps": 1e-05,
17
+ "max_position_embeddings": 77,
18
+ "model_type": "clip_text_model",
19
+ "neuron": {
20
+ "auto_cast": "matmul",
21
+ "auto_cast_type": "bf16",
22
+ "compiler_type": "neuronx-cc",
23
+ "compiler_version": "2.15.143.0+e39249ad",
24
+ "dynamic_batch_size": false,
25
+ "inline_weights_to_neff": true,
26
+ "input_names": [
27
+ "input_ids"
28
+ ],
29
+ "model_type": "clip-text-model",
30
+ "optlevel": "2",
31
+ "output_attentions": false,
32
+ "output_hidden_states": false,
33
+ "output_names": [
34
+ "last_hidden_state",
35
+ "pooler_output"
36
+ ],
37
+ "static_batch_size": 1,
38
+ "static_sequence_length": 77,
39
+ "tensor_parallel_size": 1
40
+ },
41
+ "num_attention_heads": 12,
42
+ "num_hidden_layers": 12,
43
+ "pad_token_id": 1,
44
+ "projection_dim": 512,
45
+ "task": "feature-extraction",
46
+ "torch_dtype": "float32",
47
+ "torchscript": true,
48
+ "transformers_version": "4.48.3",
49
+ "vocab_size": 49408
50
+ }
neuron/text_encoder/model.neuron ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0317993ca77ba8658bb9cb3e6acf8de57b98b862fbe51fa11d499bd53194d896
3
+ size 375660151
neuron/tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
neuron/tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|endoftext|>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<|endoftext|>",
25
+ "lstrip": false,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
neuron/tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "49406": {
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "49407": {
13
+ "content": "<|endoftext|>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ }
20
+ },
21
+ "bos_token": "<|startoftext|>",
22
+ "clean_up_tokenization_spaces": false,
23
+ "do_lower_case": true,
24
+ "eos_token": "<|endoftext|>",
25
+ "errors": "replace",
26
+ "extra_special_tokens": {},
27
+ "model_max_length": 77,
28
+ "pad_token": "<|endoftext|>",
29
+ "tokenizer_class": "CLIPTokenizer",
30
+ "unk_token": "<|endoftext|>"
31
+ }
neuron/tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
neuron/unet/config.json ADDED
@@ -0,0 +1,131 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DConditionModel",
3
+ "_commit_hash": null,
4
+ "_diffusers_version": "0.34.0",
5
+ "_name_or_path": "/home/user/.cache/huggingface/hub/models--CompVis--stable-diffusion-v1-4/snapshots/133a221b8aa7292a167afc5127cb63fb5005638b/unet",
6
+ "_use_default_values": [
7
+ "num_attention_heads",
8
+ "num_class_embeds",
9
+ "time_embedding_type",
10
+ "only_cross_attention",
11
+ "class_embeddings_concat",
12
+ "resnet_out_scale_factor",
13
+ "dual_cross_attention",
14
+ "time_cond_proj_dim",
15
+ "conv_in_kernel",
16
+ "time_embedding_act_fn",
17
+ "reverse_transformer_layers_per_block",
18
+ "resnet_skip_time_act",
19
+ "cross_attention_norm",
20
+ "resnet_time_scale_shift",
21
+ "conv_out_kernel",
22
+ "addition_time_embed_dim",
23
+ "addition_embed_type",
24
+ "mid_block_only_cross_attention",
25
+ "mid_block_type",
26
+ "projection_class_embeddings_input_dim",
27
+ "encoder_hid_dim",
28
+ "transformer_layers_per_block",
29
+ "upcast_attention",
30
+ "attention_type",
31
+ "use_linear_projection",
32
+ "dropout",
33
+ "addition_embed_type_num_heads",
34
+ "time_embedding_dim",
35
+ "timestep_post_act",
36
+ "encoder_hid_dim_type",
37
+ "class_embed_type"
38
+ ],
39
+ "act_fn": "silu",
40
+ "addition_embed_type": null,
41
+ "addition_embed_type_num_heads": 64,
42
+ "addition_time_embed_dim": null,
43
+ "attention_head_dim": 8,
44
+ "attention_type": "default",
45
+ "block_out_channels": [
46
+ 320,
47
+ 640,
48
+ 1280,
49
+ 1280
50
+ ],
51
+ "center_input_sample": false,
52
+ "class_embed_type": null,
53
+ "class_embeddings_concat": false,
54
+ "conv_in_kernel": 3,
55
+ "conv_out_kernel": 3,
56
+ "cross_attention_dim": 768,
57
+ "cross_attention_norm": null,
58
+ "down_block_types": [
59
+ "CrossAttnDownBlock2D",
60
+ "CrossAttnDownBlock2D",
61
+ "CrossAttnDownBlock2D",
62
+ "DownBlock2D"
63
+ ],
64
+ "downsample_padding": 1,
65
+ "dropout": 0.0,
66
+ "dual_cross_attention": false,
67
+ "encoder_hid_dim": null,
68
+ "encoder_hid_dim_type": null,
69
+ "flip_sin_to_cos": true,
70
+ "freq_shift": 0,
71
+ "in_channels": 4,
72
+ "layers_per_block": 2,
73
+ "mid_block_only_cross_attention": null,
74
+ "mid_block_scale_factor": 1,
75
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
76
+ "neuron": {
77
+ "auto_cast": "matmul",
78
+ "auto_cast_type": "bf16",
79
+ "compiler_type": "neuronx-cc",
80
+ "compiler_version": "2.15.143.0+e39249ad",
81
+ "dynamic_batch_size": false,
82
+ "inline_weights_to_neff": true,
83
+ "input_names": [
84
+ "sample",
85
+ "timestep",
86
+ "encoder_hidden_states"
87
+ ],
88
+ "model_type": "unet",
89
+ "optlevel": "2",
90
+ "output_attentions": false,
91
+ "output_hidden_states": false,
92
+ "output_names": [
93
+ "sample"
94
+ ],
95
+ "static_batch_size": 1,
96
+ "static_height": 64,
97
+ "static_num_channels": 4,
98
+ "static_sequence_length": 77,
99
+ "static_vae_scale_factor": 8,
100
+ "static_width": 64,
101
+ "tensor_parallel_size": 1
102
+ },
103
+ "norm_eps": 1e-05,
104
+ "norm_num_groups": 32,
105
+ "num_attention_heads": null,
106
+ "num_class_embeds": null,
107
+ "only_cross_attention": false,
108
+ "out_channels": 4,
109
+ "projection_class_embeddings_input_dim": null,
110
+ "resnet_out_scale_factor": 1.0,
111
+ "resnet_skip_time_act": false,
112
+ "resnet_time_scale_shift": "default",
113
+ "reverse_transformer_layers_per_block": null,
114
+ "sample_size": 64,
115
+ "task": "semantic-segmentation",
116
+ "time_cond_proj_dim": null,
117
+ "time_embedding_act_fn": null,
118
+ "time_embedding_dim": null,
119
+ "time_embedding_type": "positional",
120
+ "timestep_post_act": null,
121
+ "transformer_layers_per_block": 1,
122
+ "transformers_version": null,
123
+ "up_block_types": [
124
+ "UpBlock2D",
125
+ "CrossAttnUpBlock2D",
126
+ "CrossAttnUpBlock2D",
127
+ "CrossAttnUpBlock2D"
128
+ ],
129
+ "upcast_attention": false,
130
+ "use_linear_projection": false
131
+ }
neuron/unet/model.neuron ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5d2ea7e8f9e69a2c4625cee3286e40c07168ece969951a9ea71b28c5e5d9fd6
3
+ size 1900028915
neuron/vae_decoder/config.json ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_commit_hash": null,
4
+ "_diffusers_version": "0.34.0",
5
+ "_name_or_path": "/home/user/.cache/huggingface/hub/models--CompVis--stable-diffusion-v1-4/snapshots/133a221b8aa7292a167afc5127cb63fb5005638b/vae",
6
+ "_use_default_values": [
7
+ "use_post_quant_conv",
8
+ "mid_block_add_attention",
9
+ "shift_factor",
10
+ "latents_std",
11
+ "use_quant_conv",
12
+ "latents_mean",
13
+ "norm_num_groups",
14
+ "force_upcast"
15
+ ],
16
+ "act_fn": "silu",
17
+ "block_out_channels": [
18
+ 128,
19
+ 256,
20
+ 512,
21
+ 512
22
+ ],
23
+ "down_block_types": [
24
+ "DownEncoderBlock2D",
25
+ "DownEncoderBlock2D",
26
+ "DownEncoderBlock2D",
27
+ "DownEncoderBlock2D"
28
+ ],
29
+ "force_upcast": true,
30
+ "in_channels": 3,
31
+ "latent_channels": 4,
32
+ "latents_mean": null,
33
+ "latents_std": null,
34
+ "layers_per_block": 2,
35
+ "mid_block_add_attention": true,
36
+ "neuron": {
37
+ "auto_cast": "matmul",
38
+ "auto_cast_type": "bf16",
39
+ "compiler_type": "neuronx-cc",
40
+ "compiler_version": "2.15.143.0+e39249ad",
41
+ "dynamic_batch_size": false,
42
+ "inline_weights_to_neff": true,
43
+ "input_names": [
44
+ "latent_sample"
45
+ ],
46
+ "model_type": "vae-decoder",
47
+ "optlevel": "2",
48
+ "output_attentions": false,
49
+ "output_hidden_states": false,
50
+ "output_names": [
51
+ "sample"
52
+ ],
53
+ "static_batch_size": 1,
54
+ "static_height": 64,
55
+ "static_num_channels": 4,
56
+ "static_width": 64,
57
+ "tensor_parallel_size": 1
58
+ },
59
+ "norm_num_groups": 32,
60
+ "out_channels": 3,
61
+ "sample_size": 512,
62
+ "scaling_factor": 0.18215,
63
+ "shift_factor": null,
64
+ "task": "semantic-segmentation",
65
+ "transformers_version": null,
66
+ "up_block_types": [
67
+ "UpDecoderBlock2D",
68
+ "UpDecoderBlock2D",
69
+ "UpDecoderBlock2D",
70
+ "UpDecoderBlock2D"
71
+ ],
72
+ "use_post_quant_conv": true,
73
+ "use_quant_conv": true
74
+ }
neuron/vae_decoder/model.neuron ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e34b73737a49dbadfae2e5bc5d21fdbdc35441fe5935d81926366d5905d5585
3
+ size 247896819
neuron/vae_encoder/config.json ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_commit_hash": null,
4
+ "_diffusers_version": "0.34.0",
5
+ "_name_or_path": "/home/user/.cache/huggingface/hub/models--CompVis--stable-diffusion-v1-4/snapshots/133a221b8aa7292a167afc5127cb63fb5005638b/vae",
6
+ "_use_default_values": [
7
+ "use_post_quant_conv",
8
+ "mid_block_add_attention",
9
+ "shift_factor",
10
+ "latents_std",
11
+ "use_quant_conv",
12
+ "latents_mean",
13
+ "norm_num_groups",
14
+ "force_upcast"
15
+ ],
16
+ "act_fn": "silu",
17
+ "block_out_channels": [
18
+ 128,
19
+ 256,
20
+ 512,
21
+ 512
22
+ ],
23
+ "down_block_types": [
24
+ "DownEncoderBlock2D",
25
+ "DownEncoderBlock2D",
26
+ "DownEncoderBlock2D",
27
+ "DownEncoderBlock2D"
28
+ ],
29
+ "force_upcast": true,
30
+ "in_channels": 3,
31
+ "latent_channels": 4,
32
+ "latents_mean": null,
33
+ "latents_std": null,
34
+ "layers_per_block": 2,
35
+ "mid_block_add_attention": true,
36
+ "neuron": {
37
+ "auto_cast": "matmul",
38
+ "auto_cast_type": "bf16",
39
+ "compiler_type": "neuronx-cc",
40
+ "compiler_version": "2.15.143.0+e39249ad",
41
+ "dynamic_batch_size": false,
42
+ "inline_weights_to_neff": true,
43
+ "input_names": [
44
+ "sample"
45
+ ],
46
+ "model_type": "vae-encoder",
47
+ "optlevel": "2",
48
+ "output_attentions": false,
49
+ "output_hidden_states": false,
50
+ "output_names": [
51
+ "latent_parameters"
52
+ ],
53
+ "static_batch_size": 1,
54
+ "static_height": 512,
55
+ "static_num_channels": 3,
56
+ "static_width": 512,
57
+ "tensor_parallel_size": 1
58
+ },
59
+ "norm_num_groups": 32,
60
+ "out_channels": 3,
61
+ "sample_size": 512,
62
+ "scaling_factor": 0.18215,
63
+ "shift_factor": null,
64
+ "task": "semantic-segmentation",
65
+ "transformers_version": null,
66
+ "up_block_types": [
67
+ "UpDecoderBlock2D",
68
+ "UpDecoderBlock2D",
69
+ "UpDecoderBlock2D",
70
+ "UpDecoderBlock2D"
71
+ ],
72
+ "use_post_quant_conv": true,
73
+ "use_quant_conv": true
74
+ }
neuron/vae_encoder/model.neuron ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3851ce0e1548466616908a111352321f080a5fd7672e09f91a28cab0eb7103c
3
+ size 155910897