nielsr HF staff commited on
Commit
f2c3905
·
1 Parent(s): c7707a1

Upload MaskRCNNForObjectDetection

Browse files
Files changed (2) hide show
  1. config.json +223 -0
  2. pytorch_model.bin +3 -0
config.json ADDED
@@ -0,0 +1,223 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_commit_hash": null,
3
+ "anchor_generator_ratios": [
4
+ 0.5,
5
+ 1.0,
6
+ 2.0
7
+ ],
8
+ "anchor_generator_scales": [
9
+ 8
10
+ ],
11
+ "anchor_generator_strides": [
12
+ 4,
13
+ 8,
14
+ 16,
15
+ 32,
16
+ 64
17
+ ],
18
+ "architectures": [
19
+ "MaskRCNNForObjectDetection"
20
+ ],
21
+ "backbone_config": {
22
+ "_out_features": [
23
+ "stage4",
24
+ "stage6",
25
+ "stage8",
26
+ "stage12"
27
+ ],
28
+ "_out_indices": [
29
+ 4,
30
+ 6,
31
+ 8,
32
+ 12
33
+ ],
34
+ "add_fpn": true,
35
+ "attention_probs_dropout_prob": 0.0,
36
+ "auxiliary_channels": 256,
37
+ "auxiliary_concat_input": false,
38
+ "auxiliary_loss_weight": 0.4,
39
+ "auxiliary_num_convs": 1,
40
+ "drop_path_rate": 0.1,
41
+ "hidden_dropout_prob": 0.0,
42
+ "hidden_size": 768,
43
+ "intermediate_size": 3072,
44
+ "layer_scale_init_value": 0.1,
45
+ "model_type": "beit",
46
+ "num_attention_heads": 12,
47
+ "num_hidden_layers": 12,
48
+ "patch_size": 16,
49
+ "pool_scales": [
50
+ 1,
51
+ 2,
52
+ 3,
53
+ 6
54
+ ],
55
+ "reshape_hidden_states": true,
56
+ "semantic_loss_ignore_index": 255,
57
+ "stage_names": [
58
+ "stem",
59
+ "stage1",
60
+ "stage2",
61
+ "stage3",
62
+ "stage4",
63
+ "stage5",
64
+ "stage6",
65
+ "stage7",
66
+ "stage8",
67
+ "stage9",
68
+ "stage10",
69
+ "stage11",
70
+ "stage12"
71
+ ],
72
+ "use_absolute_position_embeddings": true,
73
+ "use_auxiliary_head": true,
74
+ "use_mask_token": false,
75
+ "use_mean_pooling": true,
76
+ "use_relative_position_bias": false,
77
+ "use_shared_relative_position_bias": false,
78
+ "vocab_size": 8192
79
+ },
80
+ "bbox_head_bbox_coder_target_means": [
81
+ 0.0,
82
+ 0.0,
83
+ 0.0,
84
+ 0.0
85
+ ],
86
+ "bbox_head_bbox_coder_target_stds": [
87
+ 0.1,
88
+ 0.1,
89
+ 0.2,
90
+ 0.2
91
+ ],
92
+ "bbox_head_fc_out_channels": 1024,
93
+ "bbox_head_in_channels": 256,
94
+ "bbox_head_num_shared_fcs": 2,
95
+ "bbox_head_reg_class_agnostic": false,
96
+ "bbox_head_reg_decoded_bbox": false,
97
+ "bbox_head_roi_feat_size": 7,
98
+ "bbox_roi_extractor_featmap_strides": [
99
+ 4,
100
+ 8,
101
+ 16,
102
+ 32
103
+ ],
104
+ "bbox_roi_extractor_out_channels": 256,
105
+ "bbox_roi_extractor_roi_layer": {
106
+ "output_size": 7,
107
+ "sampling_ratio": 0,
108
+ "type": "RoIAlign"
109
+ },
110
+ "fpn_num_outputs": 5,
111
+ "fpn_out_channels": 256,
112
+ "id2label": {
113
+ "0": "LABEL_0",
114
+ "1": "LABEL_1",
115
+ "2": "LABEL_2",
116
+ "3": "LABEL_3",
117
+ "4": "LABEL_4"
118
+ },
119
+ "initializer_range": 0.02,
120
+ "label2id": {
121
+ "LABEL_0": 0,
122
+ "LABEL_1": 1,
123
+ "LABEL_2": 2,
124
+ "LABEL_3": 3,
125
+ "LABEL_4": 4
126
+ },
127
+ "mask_head_conv_out_channels": 256,
128
+ "mask_head_in_channels": 256,
129
+ "mask_head_num_convs": 4,
130
+ "mask_roi_extractor_featmap_strides": [
131
+ 4,
132
+ 8,
133
+ 16,
134
+ 32
135
+ ],
136
+ "mask_roi_extractor_out_channels": 256,
137
+ "mask_roi_extractor_roi_layer": {
138
+ "output_size": 14,
139
+ "sampling_ratio": 0,
140
+ "type": "RoIAlign"
141
+ },
142
+ "model_type": "maskrcnn",
143
+ "rcnn_assigner_ignore_iof_thr": -1,
144
+ "rcnn_assigner_match_low_quality": true,
145
+ "rcnn_assigner_min_pos_iou": 0.5,
146
+ "rcnn_assigner_neg_iou_thr": 0.5,
147
+ "rcnn_assigner_pos_iou_thr": 0.5,
148
+ "rcnn_sampler_add_gt_as_proposals": true,
149
+ "rcnn_sampler_num": 512,
150
+ "rcnn_sampler_num_samples_upper_bound": -1,
151
+ "rcnn_sampler_pos_fraction": 0.25,
152
+ "rcnn_test_cfg": {
153
+ "mask_thr_binary": 0.5,
154
+ "max_per_img": 100,
155
+ "nms": {
156
+ "iou_threshold": 0.5,
157
+ "type": "nms"
158
+ },
159
+ "score_thr": 0.05
160
+ },
161
+ "rcnn_train_cfg": {
162
+ "debug": false,
163
+ "mask_size": 28,
164
+ "pos_weight": -1
165
+ },
166
+ "rpn_assigner_ignore_iof_thr": -1,
167
+ "rpn_assigner_match_low_quality": true,
168
+ "rpn_assigner_min_pos_iou": 0.3,
169
+ "rpn_assigner_neg_iou_thr": 0.3,
170
+ "rpn_assigner_pos_iou_thr": 0.7,
171
+ "rpn_bbox_coder_target_means": [
172
+ 0.0,
173
+ 0.0,
174
+ 0.0,
175
+ 0.0
176
+ ],
177
+ "rpn_bbox_coder_target_stds": [
178
+ 1.0,
179
+ 1.0,
180
+ 1.0,
181
+ 1.0
182
+ ],
183
+ "rpn_feat_channels": 256,
184
+ "rpn_in_channels": 256,
185
+ "rpn_loss_bbox": {
186
+ "loss_weight": 1.0,
187
+ "type": "L1Loss"
188
+ },
189
+ "rpn_loss_cls": {
190
+ "loss_weight": 1.0,
191
+ "type": "CrossEntropyLoss",
192
+ "use_sigmoid": true
193
+ },
194
+ "rpn_proposal": {
195
+ "max_per_img": 1000,
196
+ "min_bbox_size": 0,
197
+ "nms": {
198
+ "iou_threshold": 0.7,
199
+ "type": "nms"
200
+ },
201
+ "nms_pre": 2000
202
+ },
203
+ "rpn_sampler_add_gt_as_proposals": false,
204
+ "rpn_sampler_num": 256,
205
+ "rpn_sampler_num_samples_upper_bound": -1,
206
+ "rpn_sampler_pos_fraction": 0.5,
207
+ "rpn_test_cfg": {
208
+ "max_per_img": 1000,
209
+ "min_bbox_size": 0,
210
+ "nms": {
211
+ "iou_threshold": 0.7,
212
+ "type": "nms"
213
+ },
214
+ "nms_pre": 1000
215
+ },
216
+ "rpn_train_cfg": {
217
+ "allowed_border": -1,
218
+ "debug": false,
219
+ "pos_weight": -1
220
+ },
221
+ "torch_dtype": "float32",
222
+ "transformers_version": null
223
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46a3b5ede3345fbadff18aa41991d5ecba874ab3b5195a14188a1bbdc574f030
3
+ size 452795617