rjeeva commited on
Commit
4be4621
·
verified ·
1 Parent(s): 4b17e41
config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "EfficientNetEncoder"
4
+ ],
5
+ "model_type": "efficient_net",
6
+ "hidden_size": 1280,
7
+ "num_attention_heads": null,
8
+ "image_size": 224,
9
+ "patch_size": null,
10
+ "num_channels": 3,
11
+ "initializer_range": 0.02,
12
+ "layer_norm_eps": 1e-12,
13
+ "dropout": 0.0,
14
+ "attention_dropout": 0.0,
15
+ "hidden_dropout_prob": 0.0,
16
+ "use_cache": true,
17
+ "pad_token_id": null,
18
+ "bos_token_id": null,
19
+ "eos_token_id": null,
20
+ "tie_word_embeddings": false,
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.28.0",
23
+ "framework": "onnx"
24
+ }
model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c53d30fb79700944bd107b1fba6a242f77b1e0f08aaeee06f2b03939591d138
3
+ size 16028204
preprocessor_config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_resize": true,
3
+ "size": {
4
+ "height": 224,
5
+ "width": 224
6
+ },
7
+ "do_center_crop": true,
8
+ "crop_size": {
9
+ "height": 224,
10
+ "width": 224
11
+ },
12
+ "do_normalize": true,
13
+ "image_mean": [
14
+ 0.485,
15
+ 0.456,
16
+ 0.406
17
+ ],
18
+ "image_std": [
19
+ 0.229,
20
+ 0.224,
21
+ 0.225
22
+ ],
23
+ "do_convert_rgb": true,
24
+ "preprocessor_class": "AutoImageProcessor"
25
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "unk_token": null,
3
+ "sep_token": null,
4
+ "pad_token": null,
5
+ "cls_token": null,
6
+ "mask_token": null,
7
+ "additional_special_tokens": []
8
+ }
tokenizer.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
5
+ "added_tokens": [],
6
+ "normalizer": null,
7
+ "pre_tokenizer": null,
8
+ "post_processor": null,
9
+ "decoder": null,
10
+ "model": {
11
+ "type": "unigram",
12
+ "unk_id": null,
13
+ "bos_id": null,
14
+ "eos_id": null,
15
+ "vocab": {}
16
+ }
17
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_max_length": 224,
3
+ "padding_side": "right",
4
+ "truncation_side": "right",
5
+ "model_input_names": [
6
+ "input"
7
+ ],
8
+ "special_tokens_map_file": "special_tokens_map.json",
9
+ "tokenizer_class": "PreTrainedTokenizer",
10
+ "is_vision_model": true
11
+ }