Delores-Lin commited on
Commit
4da7452
·
verified ·
1 Parent(s): 0e05a65

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -1,35 +1,37 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Recognition/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ Structure/PP-DocLayout_plus-L/inference.pdiparams filter=lfs diff=lfs merge=lfs -text
Recognition/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
Recognition/chat_template.jinja ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0]['role'] == 'system' %}
4
+ {{- messages[0]['content'] }}
5
+ {%- else %}
6
+ {{- 'You are a helpful assistant.' }}
7
+ {%- endif %}
8
+ {{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
9
+ {%- for tool in tools %}
10
+ {{- "\n" }}
11
+ {{- tool | tojson }}
12
+ {%- endfor %}
13
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
14
+ {%- else %}
15
+ {%- if messages[0]['role'] == 'system' %}
16
+ {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
17
+ {%- else %}
18
+ {{- '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}
19
+ {%- endif %}
20
+ {%- endif %}
21
+ {%- for message in messages %}
22
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) or (message.role == "assistant" and not message.tool_calls) %}
23
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
24
+ {%- elif message.role == "assistant" %}
25
+ {{- '<|im_start|>' + message.role }}
26
+ {%- if message.content %}
27
+ {{- '\n' + message.content }}
28
+ {%- endif %}
29
+ {%- for tool_call in message.tool_calls %}
30
+ {%- if tool_call.function is defined %}
31
+ {%- set tool_call = tool_call.function %}
32
+ {%- endif %}
33
+ {{- '\n<tool_call>\n{"name": "' }}
34
+ {{- tool_call.name }}
35
+ {{- '", "arguments": ' }}
36
+ {{- tool_call.arguments | tojson }}
37
+ {{- '}\n</tool_call>' }}
38
+ {%- endfor %}
39
+ {{- '<|im_end|>\n' }}
40
+ {%- elif message.role == "tool" %}
41
+ {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != "tool") %}
42
+ {{- '<|im_start|>user' }}
43
+ {%- endif %}
44
+ {{- '\n<tool_response>\n' }}
45
+ {{- message.content }}
46
+ {{- '\n</tool_response>' }}
47
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
48
+ {{- '<|im_end|>\n' }}
49
+ {%- endif %}
50
+ {%- endif %}
51
+ {%- endfor %}
52
+ {%- if add_generation_prompt %}
53
+ {{- '<|im_start|>assistant\n' }}
54
+ {%- endif %}
Recognition/config.json ADDED
@@ -0,0 +1,113 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2_5_VLForConditionalGeneration"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 2048,
10
+ "image_token_id": 151655,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 11008,
13
+ "max_position_embeddings": 128000,
14
+ "max_window_layers": 70,
15
+ "model_type": "qwen2_5_vl",
16
+ "num_attention_heads": 16,
17
+ "num_hidden_layers": 12,
18
+ "num_key_value_heads": 2,
19
+ "quantization_config": {
20
+ "bits": 4,
21
+ "group_size": 128,
22
+ "quant_method": "awq",
23
+ "version": "gemm",
24
+ "zero_point": true
25
+ },
26
+ "rms_norm_eps": 1e-06,
27
+ "rope_scaling": {
28
+ "mrope_section": [
29
+ 16,
30
+ 24,
31
+ 24
32
+ ],
33
+ "rope_type": "default",
34
+ "type": "default"
35
+ },
36
+ "rope_theta": 1000000.0,
37
+ "sliding_window": 32768,
38
+ "text_config": {
39
+ "_name_or_path": "MonkeyOCR-1.2B-0709",
40
+ "architectures": [
41
+ "Qwen2_5_VLForConditionalGeneration"
42
+ ],
43
+ "attention_dropout": 0.0,
44
+ "bos_token_id": 151643,
45
+ "eos_token_id": 151645,
46
+ "hidden_act": "silu",
47
+ "hidden_size": 2048,
48
+ "image_token_id": null,
49
+ "initializer_range": 0.02,
50
+ "intermediate_size": 11008,
51
+ "max_position_embeddings": 128000,
52
+ "max_window_layers": 70,
53
+ "model_type": "qwen2_5_vl_text",
54
+ "num_attention_heads": 16,
55
+ "num_hidden_layers": 12,
56
+ "num_key_value_heads": 2,
57
+ "rms_norm_eps": 1e-06,
58
+ "rope_scaling": {
59
+ "mrope_section": [
60
+ 16,
61
+ 24,
62
+ 24
63
+ ],
64
+ "rope_type": "default",
65
+ "type": "default"
66
+ },
67
+ "rope_theta": 1000000.0,
68
+ "sliding_window": 32768,
69
+ "tie_word_embeddings": true,
70
+ "torch_dtype": "float32",
71
+ "use_cache": false,
72
+ "use_sliding_window": false,
73
+ "video_token_id": null,
74
+ "vision_end_token_id": 151653,
75
+ "vision_start_token_id": 151652,
76
+ "vision_token_id": 151654,
77
+ "vocab_size": 151936
78
+ },
79
+ "torch_dtype": "float32",
80
+ "transformers_version": "4.52.4",
81
+ "use_cache": false,
82
+ "use_sliding_window": false,
83
+ "video_token_id": 151656,
84
+ "vision_config": {
85
+ "depth": 32,
86
+ "fullatt_block_indexes": [
87
+ 7,
88
+ 15,
89
+ 23,
90
+ 31
91
+ ],
92
+ "hidden_act": "silu",
93
+ "hidden_size": 1280,
94
+ "in_channels": 3,
95
+ "in_chans": 3,
96
+ "initializer_range": 0.02,
97
+ "intermediate_size": 3420,
98
+ "model_type": "qwen2_5_vl",
99
+ "num_heads": 16,
100
+ "out_hidden_size": 2048,
101
+ "patch_size": 14,
102
+ "spatial_merge_size": 2,
103
+ "spatial_patch_size": 14,
104
+ "temporal_patch_size": 2,
105
+ "tokens_per_second": 2,
106
+ "torch_dtype": "float32",
107
+ "window_size": 112
108
+ },
109
+ "vision_end_token_id": 151653,
110
+ "vision_start_token_id": 151652,
111
+ "vision_token_id": 151654,
112
+ "vocab_size": 151936
113
+ }
Recognition/generation_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.05,
10
+ "temperature": 0.1,
11
+ "top_k": 1,
12
+ "top_p": 0.001,
13
+ "transformers_version": "4.50.0.dev0"
14
+ }
Recognition/inputs_stats.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c22316a388263bafc234f1def98ffbdacaa3730fa3688f251756c8bfc74a5c90
3
+ size 5710002
Recognition/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
Recognition/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0669f5e982744527eb71177f7861f0b3515aec4aa7e7abf66d27604ae50e841f
3
+ size 4414737960
Recognition/outputs_stats.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00325e1830bdf0ddaaa002010ea4e686259f2ba19ef35f2d5868898bb7b22d11
3
+ size 8985570
Recognition/preprocessor_config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": true,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.48145466,
8
+ 0.4578275,
9
+ 0.40821073
10
+ ],
11
+ "image_processor_type": "Qwen2VLImageProcessor",
12
+ "image_std": [
13
+ 0.26862954,
14
+ 0.26130258,
15
+ 0.27577711
16
+ ],
17
+ "max_pixels": 12845056,
18
+ "merge_size": 2,
19
+ "min_pixels": 3136,
20
+ "patch_size": 14,
21
+ "processor_class": "Qwen2_5_VLProcessor",
22
+ "resample": 3,
23
+ "rescale_factor": 0.00392156862745098,
24
+ "size": {
25
+ "longest_edge": 12845056,
26
+ "shortest_edge": 3136
27
+ },
28
+ "temporal_patch_size": 2
29
+ }
Recognition/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
Recognition/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
Recognition/tokenizer_config.json ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "clean_up_tokenization_spaces": false,
199
+ "eos_token": "<|im_end|>",
200
+ "errors": "replace",
201
+ "extra_special_tokens": {},
202
+ "model_max_length": 8196,
203
+ "pad_token": "<|endoftext|>",
204
+ "padding_side": "right",
205
+ "processor_class": "Qwen2_5_VLProcessor",
206
+ "split_special_tokens": false,
207
+ "tokenizer_class": "Qwen2Tokenizer",
208
+ "unk_token": null
209
+ }
Recognition/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
Relation/config.json ADDED
@@ -0,0 +1,1063 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LayoutLMv3ForTokenClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
+ "classifier_dropout": null,
8
+ "coordinate_size": 171,
9
+ "eos_token_id": 2,
10
+ "has_relative_attention_bias": true,
11
+ "has_spatial_attention_bias": true,
12
+ "hidden_act": "gelu",
13
+ "hidden_dropout_prob": 0.1,
14
+ "hidden_size": 1024,
15
+ "id2label": {
16
+ "0": "LABEL_0",
17
+ "1": "LABEL_1",
18
+ "2": "LABEL_2",
19
+ "3": "LABEL_3",
20
+ "4": "LABEL_4",
21
+ "5": "LABEL_5",
22
+ "6": "LABEL_6",
23
+ "7": "LABEL_7",
24
+ "8": "LABEL_8",
25
+ "9": "LABEL_9",
26
+ "10": "LABEL_10",
27
+ "11": "LABEL_11",
28
+ "12": "LABEL_12",
29
+ "13": "LABEL_13",
30
+ "14": "LABEL_14",
31
+ "15": "LABEL_15",
32
+ "16": "LABEL_16",
33
+ "17": "LABEL_17",
34
+ "18": "LABEL_18",
35
+ "19": "LABEL_19",
36
+ "20": "LABEL_20",
37
+ "21": "LABEL_21",
38
+ "22": "LABEL_22",
39
+ "23": "LABEL_23",
40
+ "24": "LABEL_24",
41
+ "25": "LABEL_25",
42
+ "26": "LABEL_26",
43
+ "27": "LABEL_27",
44
+ "28": "LABEL_28",
45
+ "29": "LABEL_29",
46
+ "30": "LABEL_30",
47
+ "31": "LABEL_31",
48
+ "32": "LABEL_32",
49
+ "33": "LABEL_33",
50
+ "34": "LABEL_34",
51
+ "35": "LABEL_35",
52
+ "36": "LABEL_36",
53
+ "37": "LABEL_37",
54
+ "38": "LABEL_38",
55
+ "39": "LABEL_39",
56
+ "40": "LABEL_40",
57
+ "41": "LABEL_41",
58
+ "42": "LABEL_42",
59
+ "43": "LABEL_43",
60
+ "44": "LABEL_44",
61
+ "45": "LABEL_45",
62
+ "46": "LABEL_46",
63
+ "47": "LABEL_47",
64
+ "48": "LABEL_48",
65
+ "49": "LABEL_49",
66
+ "50": "LABEL_50",
67
+ "51": "LABEL_51",
68
+ "52": "LABEL_52",
69
+ "53": "LABEL_53",
70
+ "54": "LABEL_54",
71
+ "55": "LABEL_55",
72
+ "56": "LABEL_56",
73
+ "57": "LABEL_57",
74
+ "58": "LABEL_58",
75
+ "59": "LABEL_59",
76
+ "60": "LABEL_60",
77
+ "61": "LABEL_61",
78
+ "62": "LABEL_62",
79
+ "63": "LABEL_63",
80
+ "64": "LABEL_64",
81
+ "65": "LABEL_65",
82
+ "66": "LABEL_66",
83
+ "67": "LABEL_67",
84
+ "68": "LABEL_68",
85
+ "69": "LABEL_69",
86
+ "70": "LABEL_70",
87
+ "71": "LABEL_71",
88
+ "72": "LABEL_72",
89
+ "73": "LABEL_73",
90
+ "74": "LABEL_74",
91
+ "75": "LABEL_75",
92
+ "76": "LABEL_76",
93
+ "77": "LABEL_77",
94
+ "78": "LABEL_78",
95
+ "79": "LABEL_79",
96
+ "80": "LABEL_80",
97
+ "81": "LABEL_81",
98
+ "82": "LABEL_82",
99
+ "83": "LABEL_83",
100
+ "84": "LABEL_84",
101
+ "85": "LABEL_85",
102
+ "86": "LABEL_86",
103
+ "87": "LABEL_87",
104
+ "88": "LABEL_88",
105
+ "89": "LABEL_89",
106
+ "90": "LABEL_90",
107
+ "91": "LABEL_91",
108
+ "92": "LABEL_92",
109
+ "93": "LABEL_93",
110
+ "94": "LABEL_94",
111
+ "95": "LABEL_95",
112
+ "96": "LABEL_96",
113
+ "97": "LABEL_97",
114
+ "98": "LABEL_98",
115
+ "99": "LABEL_99",
116
+ "100": "LABEL_100",
117
+ "101": "LABEL_101",
118
+ "102": "LABEL_102",
119
+ "103": "LABEL_103",
120
+ "104": "LABEL_104",
121
+ "105": "LABEL_105",
122
+ "106": "LABEL_106",
123
+ "107": "LABEL_107",
124
+ "108": "LABEL_108",
125
+ "109": "LABEL_109",
126
+ "110": "LABEL_110",
127
+ "111": "LABEL_111",
128
+ "112": "LABEL_112",
129
+ "113": "LABEL_113",
130
+ "114": "LABEL_114",
131
+ "115": "LABEL_115",
132
+ "116": "LABEL_116",
133
+ "117": "LABEL_117",
134
+ "118": "LABEL_118",
135
+ "119": "LABEL_119",
136
+ "120": "LABEL_120",
137
+ "121": "LABEL_121",
138
+ "122": "LABEL_122",
139
+ "123": "LABEL_123",
140
+ "124": "LABEL_124",
141
+ "125": "LABEL_125",
142
+ "126": "LABEL_126",
143
+ "127": "LABEL_127",
144
+ "128": "LABEL_128",
145
+ "129": "LABEL_129",
146
+ "130": "LABEL_130",
147
+ "131": "LABEL_131",
148
+ "132": "LABEL_132",
149
+ "133": "LABEL_133",
150
+ "134": "LABEL_134",
151
+ "135": "LABEL_135",
152
+ "136": "LABEL_136",
153
+ "137": "LABEL_137",
154
+ "138": "LABEL_138",
155
+ "139": "LABEL_139",
156
+ "140": "LABEL_140",
157
+ "141": "LABEL_141",
158
+ "142": "LABEL_142",
159
+ "143": "LABEL_143",
160
+ "144": "LABEL_144",
161
+ "145": "LABEL_145",
162
+ "146": "LABEL_146",
163
+ "147": "LABEL_147",
164
+ "148": "LABEL_148",
165
+ "149": "LABEL_149",
166
+ "150": "LABEL_150",
167
+ "151": "LABEL_151",
168
+ "152": "LABEL_152",
169
+ "153": "LABEL_153",
170
+ "154": "LABEL_154",
171
+ "155": "LABEL_155",
172
+ "156": "LABEL_156",
173
+ "157": "LABEL_157",
174
+ "158": "LABEL_158",
175
+ "159": "LABEL_159",
176
+ "160": "LABEL_160",
177
+ "161": "LABEL_161",
178
+ "162": "LABEL_162",
179
+ "163": "LABEL_163",
180
+ "164": "LABEL_164",
181
+ "165": "LABEL_165",
182
+ "166": "LABEL_166",
183
+ "167": "LABEL_167",
184
+ "168": "LABEL_168",
185
+ "169": "LABEL_169",
186
+ "170": "LABEL_170",
187
+ "171": "LABEL_171",
188
+ "172": "LABEL_172",
189
+ "173": "LABEL_173",
190
+ "174": "LABEL_174",
191
+ "175": "LABEL_175",
192
+ "176": "LABEL_176",
193
+ "177": "LABEL_177",
194
+ "178": "LABEL_178",
195
+ "179": "LABEL_179",
196
+ "180": "LABEL_180",
197
+ "181": "LABEL_181",
198
+ "182": "LABEL_182",
199
+ "183": "LABEL_183",
200
+ "184": "LABEL_184",
201
+ "185": "LABEL_185",
202
+ "186": "LABEL_186",
203
+ "187": "LABEL_187",
204
+ "188": "LABEL_188",
205
+ "189": "LABEL_189",
206
+ "190": "LABEL_190",
207
+ "191": "LABEL_191",
208
+ "192": "LABEL_192",
209
+ "193": "LABEL_193",
210
+ "194": "LABEL_194",
211
+ "195": "LABEL_195",
212
+ "196": "LABEL_196",
213
+ "197": "LABEL_197",
214
+ "198": "LABEL_198",
215
+ "199": "LABEL_199",
216
+ "200": "LABEL_200",
217
+ "201": "LABEL_201",
218
+ "202": "LABEL_202",
219
+ "203": "LABEL_203",
220
+ "204": "LABEL_204",
221
+ "205": "LABEL_205",
222
+ "206": "LABEL_206",
223
+ "207": "LABEL_207",
224
+ "208": "LABEL_208",
225
+ "209": "LABEL_209",
226
+ "210": "LABEL_210",
227
+ "211": "LABEL_211",
228
+ "212": "LABEL_212",
229
+ "213": "LABEL_213",
230
+ "214": "LABEL_214",
231
+ "215": "LABEL_215",
232
+ "216": "LABEL_216",
233
+ "217": "LABEL_217",
234
+ "218": "LABEL_218",
235
+ "219": "LABEL_219",
236
+ "220": "LABEL_220",
237
+ "221": "LABEL_221",
238
+ "222": "LABEL_222",
239
+ "223": "LABEL_223",
240
+ "224": "LABEL_224",
241
+ "225": "LABEL_225",
242
+ "226": "LABEL_226",
243
+ "227": "LABEL_227",
244
+ "228": "LABEL_228",
245
+ "229": "LABEL_229",
246
+ "230": "LABEL_230",
247
+ "231": "LABEL_231",
248
+ "232": "LABEL_232",
249
+ "233": "LABEL_233",
250
+ "234": "LABEL_234",
251
+ "235": "LABEL_235",
252
+ "236": "LABEL_236",
253
+ "237": "LABEL_237",
254
+ "238": "LABEL_238",
255
+ "239": "LABEL_239",
256
+ "240": "LABEL_240",
257
+ "241": "LABEL_241",
258
+ "242": "LABEL_242",
259
+ "243": "LABEL_243",
260
+ "244": "LABEL_244",
261
+ "245": "LABEL_245",
262
+ "246": "LABEL_246",
263
+ "247": "LABEL_247",
264
+ "248": "LABEL_248",
265
+ "249": "LABEL_249",
266
+ "250": "LABEL_250",
267
+ "251": "LABEL_251",
268
+ "252": "LABEL_252",
269
+ "253": "LABEL_253",
270
+ "254": "LABEL_254",
271
+ "255": "LABEL_255",
272
+ "256": "LABEL_256",
273
+ "257": "LABEL_257",
274
+ "258": "LABEL_258",
275
+ "259": "LABEL_259",
276
+ "260": "LABEL_260",
277
+ "261": "LABEL_261",
278
+ "262": "LABEL_262",
279
+ "263": "LABEL_263",
280
+ "264": "LABEL_264",
281
+ "265": "LABEL_265",
282
+ "266": "LABEL_266",
283
+ "267": "LABEL_267",
284
+ "268": "LABEL_268",
285
+ "269": "LABEL_269",
286
+ "270": "LABEL_270",
287
+ "271": "LABEL_271",
288
+ "272": "LABEL_272",
289
+ "273": "LABEL_273",
290
+ "274": "LABEL_274",
291
+ "275": "LABEL_275",
292
+ "276": "LABEL_276",
293
+ "277": "LABEL_277",
294
+ "278": "LABEL_278",
295
+ "279": "LABEL_279",
296
+ "280": "LABEL_280",
297
+ "281": "LABEL_281",
298
+ "282": "LABEL_282",
299
+ "283": "LABEL_283",
300
+ "284": "LABEL_284",
301
+ "285": "LABEL_285",
302
+ "286": "LABEL_286",
303
+ "287": "LABEL_287",
304
+ "288": "LABEL_288",
305
+ "289": "LABEL_289",
306
+ "290": "LABEL_290",
307
+ "291": "LABEL_291",
308
+ "292": "LABEL_292",
309
+ "293": "LABEL_293",
310
+ "294": "LABEL_294",
311
+ "295": "LABEL_295",
312
+ "296": "LABEL_296",
313
+ "297": "LABEL_297",
314
+ "298": "LABEL_298",
315
+ "299": "LABEL_299",
316
+ "300": "LABEL_300",
317
+ "301": "LABEL_301",
318
+ "302": "LABEL_302",
319
+ "303": "LABEL_303",
320
+ "304": "LABEL_304",
321
+ "305": "LABEL_305",
322
+ "306": "LABEL_306",
323
+ "307": "LABEL_307",
324
+ "308": "LABEL_308",
325
+ "309": "LABEL_309",
326
+ "310": "LABEL_310",
327
+ "311": "LABEL_311",
328
+ "312": "LABEL_312",
329
+ "313": "LABEL_313",
330
+ "314": "LABEL_314",
331
+ "315": "LABEL_315",
332
+ "316": "LABEL_316",
333
+ "317": "LABEL_317",
334
+ "318": "LABEL_318",
335
+ "319": "LABEL_319",
336
+ "320": "LABEL_320",
337
+ "321": "LABEL_321",
338
+ "322": "LABEL_322",
339
+ "323": "LABEL_323",
340
+ "324": "LABEL_324",
341
+ "325": "LABEL_325",
342
+ "326": "LABEL_326",
343
+ "327": "LABEL_327",
344
+ "328": "LABEL_328",
345
+ "329": "LABEL_329",
346
+ "330": "LABEL_330",
347
+ "331": "LABEL_331",
348
+ "332": "LABEL_332",
349
+ "333": "LABEL_333",
350
+ "334": "LABEL_334",
351
+ "335": "LABEL_335",
352
+ "336": "LABEL_336",
353
+ "337": "LABEL_337",
354
+ "338": "LABEL_338",
355
+ "339": "LABEL_339",
356
+ "340": "LABEL_340",
357
+ "341": "LABEL_341",
358
+ "342": "LABEL_342",
359
+ "343": "LABEL_343",
360
+ "344": "LABEL_344",
361
+ "345": "LABEL_345",
362
+ "346": "LABEL_346",
363
+ "347": "LABEL_347",
364
+ "348": "LABEL_348",
365
+ "349": "LABEL_349",
366
+ "350": "LABEL_350",
367
+ "351": "LABEL_351",
368
+ "352": "LABEL_352",
369
+ "353": "LABEL_353",
370
+ "354": "LABEL_354",
371
+ "355": "LABEL_355",
372
+ "356": "LABEL_356",
373
+ "357": "LABEL_357",
374
+ "358": "LABEL_358",
375
+ "359": "LABEL_359",
376
+ "360": "LABEL_360",
377
+ "361": "LABEL_361",
378
+ "362": "LABEL_362",
379
+ "363": "LABEL_363",
380
+ "364": "LABEL_364",
381
+ "365": "LABEL_365",
382
+ "366": "LABEL_366",
383
+ "367": "LABEL_367",
384
+ "368": "LABEL_368",
385
+ "369": "LABEL_369",
386
+ "370": "LABEL_370",
387
+ "371": "LABEL_371",
388
+ "372": "LABEL_372",
389
+ "373": "LABEL_373",
390
+ "374": "LABEL_374",
391
+ "375": "LABEL_375",
392
+ "376": "LABEL_376",
393
+ "377": "LABEL_377",
394
+ "378": "LABEL_378",
395
+ "379": "LABEL_379",
396
+ "380": "LABEL_380",
397
+ "381": "LABEL_381",
398
+ "382": "LABEL_382",
399
+ "383": "LABEL_383",
400
+ "384": "LABEL_384",
401
+ "385": "LABEL_385",
402
+ "386": "LABEL_386",
403
+ "387": "LABEL_387",
404
+ "388": "LABEL_388",
405
+ "389": "LABEL_389",
406
+ "390": "LABEL_390",
407
+ "391": "LABEL_391",
408
+ "392": "LABEL_392",
409
+ "393": "LABEL_393",
410
+ "394": "LABEL_394",
411
+ "395": "LABEL_395",
412
+ "396": "LABEL_396",
413
+ "397": "LABEL_397",
414
+ "398": "LABEL_398",
415
+ "399": "LABEL_399",
416
+ "400": "LABEL_400",
417
+ "401": "LABEL_401",
418
+ "402": "LABEL_402",
419
+ "403": "LABEL_403",
420
+ "404": "LABEL_404",
421
+ "405": "LABEL_405",
422
+ "406": "LABEL_406",
423
+ "407": "LABEL_407",
424
+ "408": "LABEL_408",
425
+ "409": "LABEL_409",
426
+ "410": "LABEL_410",
427
+ "411": "LABEL_411",
428
+ "412": "LABEL_412",
429
+ "413": "LABEL_413",
430
+ "414": "LABEL_414",
431
+ "415": "LABEL_415",
432
+ "416": "LABEL_416",
433
+ "417": "LABEL_417",
434
+ "418": "LABEL_418",
435
+ "419": "LABEL_419",
436
+ "420": "LABEL_420",
437
+ "421": "LABEL_421",
438
+ "422": "LABEL_422",
439
+ "423": "LABEL_423",
440
+ "424": "LABEL_424",
441
+ "425": "LABEL_425",
442
+ "426": "LABEL_426",
443
+ "427": "LABEL_427",
444
+ "428": "LABEL_428",
445
+ "429": "LABEL_429",
446
+ "430": "LABEL_430",
447
+ "431": "LABEL_431",
448
+ "432": "LABEL_432",
449
+ "433": "LABEL_433",
450
+ "434": "LABEL_434",
451
+ "435": "LABEL_435",
452
+ "436": "LABEL_436",
453
+ "437": "LABEL_437",
454
+ "438": "LABEL_438",
455
+ "439": "LABEL_439",
456
+ "440": "LABEL_440",
457
+ "441": "LABEL_441",
458
+ "442": "LABEL_442",
459
+ "443": "LABEL_443",
460
+ "444": "LABEL_444",
461
+ "445": "LABEL_445",
462
+ "446": "LABEL_446",
463
+ "447": "LABEL_447",
464
+ "448": "LABEL_448",
465
+ "449": "LABEL_449",
466
+ "450": "LABEL_450",
467
+ "451": "LABEL_451",
468
+ "452": "LABEL_452",
469
+ "453": "LABEL_453",
470
+ "454": "LABEL_454",
471
+ "455": "LABEL_455",
472
+ "456": "LABEL_456",
473
+ "457": "LABEL_457",
474
+ "458": "LABEL_458",
475
+ "459": "LABEL_459",
476
+ "460": "LABEL_460",
477
+ "461": "LABEL_461",
478
+ "462": "LABEL_462",
479
+ "463": "LABEL_463",
480
+ "464": "LABEL_464",
481
+ "465": "LABEL_465",
482
+ "466": "LABEL_466",
483
+ "467": "LABEL_467",
484
+ "468": "LABEL_468",
485
+ "469": "LABEL_469",
486
+ "470": "LABEL_470",
487
+ "471": "LABEL_471",
488
+ "472": "LABEL_472",
489
+ "473": "LABEL_473",
490
+ "474": "LABEL_474",
491
+ "475": "LABEL_475",
492
+ "476": "LABEL_476",
493
+ "477": "LABEL_477",
494
+ "478": "LABEL_478",
495
+ "479": "LABEL_479",
496
+ "480": "LABEL_480",
497
+ "481": "LABEL_481",
498
+ "482": "LABEL_482",
499
+ "483": "LABEL_483",
500
+ "484": "LABEL_484",
501
+ "485": "LABEL_485",
502
+ "486": "LABEL_486",
503
+ "487": "LABEL_487",
504
+ "488": "LABEL_488",
505
+ "489": "LABEL_489",
506
+ "490": "LABEL_490",
507
+ "491": "LABEL_491",
508
+ "492": "LABEL_492",
509
+ "493": "LABEL_493",
510
+ "494": "LABEL_494",
511
+ "495": "LABEL_495",
512
+ "496": "LABEL_496",
513
+ "497": "LABEL_497",
514
+ "498": "LABEL_498",
515
+ "499": "LABEL_499",
516
+ "500": "LABEL_500",
517
+ "501": "LABEL_501",
518
+ "502": "LABEL_502",
519
+ "503": "LABEL_503",
520
+ "504": "LABEL_504",
521
+ "505": "LABEL_505",
522
+ "506": "LABEL_506",
523
+ "507": "LABEL_507",
524
+ "508": "LABEL_508",
525
+ "509": "LABEL_509"
526
+ },
527
+ "initializer_range": 0.02,
528
+ "input_size": 224,
529
+ "intermediate_size": 4096,
530
+ "label2id": {
531
+ "LABEL_0": 0,
532
+ "LABEL_1": 1,
533
+ "LABEL_10": 10,
534
+ "LABEL_100": 100,
535
+ "LABEL_101": 101,
536
+ "LABEL_102": 102,
537
+ "LABEL_103": 103,
538
+ "LABEL_104": 104,
539
+ "LABEL_105": 105,
540
+ "LABEL_106": 106,
541
+ "LABEL_107": 107,
542
+ "LABEL_108": 108,
543
+ "LABEL_109": 109,
544
+ "LABEL_11": 11,
545
+ "LABEL_110": 110,
546
+ "LABEL_111": 111,
547
+ "LABEL_112": 112,
548
+ "LABEL_113": 113,
549
+ "LABEL_114": 114,
550
+ "LABEL_115": 115,
551
+ "LABEL_116": 116,
552
+ "LABEL_117": 117,
553
+ "LABEL_118": 118,
554
+ "LABEL_119": 119,
555
+ "LABEL_12": 12,
556
+ "LABEL_120": 120,
557
+ "LABEL_121": 121,
558
+ "LABEL_122": 122,
559
+ "LABEL_123": 123,
560
+ "LABEL_124": 124,
561
+ "LABEL_125": 125,
562
+ "LABEL_126": 126,
563
+ "LABEL_127": 127,
564
+ "LABEL_128": 128,
565
+ "LABEL_129": 129,
566
+ "LABEL_13": 13,
567
+ "LABEL_130": 130,
568
+ "LABEL_131": 131,
569
+ "LABEL_132": 132,
570
+ "LABEL_133": 133,
571
+ "LABEL_134": 134,
572
+ "LABEL_135": 135,
573
+ "LABEL_136": 136,
574
+ "LABEL_137": 137,
575
+ "LABEL_138": 138,
576
+ "LABEL_139": 139,
577
+ "LABEL_14": 14,
578
+ "LABEL_140": 140,
579
+ "LABEL_141": 141,
580
+ "LABEL_142": 142,
581
+ "LABEL_143": 143,
582
+ "LABEL_144": 144,
583
+ "LABEL_145": 145,
584
+ "LABEL_146": 146,
585
+ "LABEL_147": 147,
586
+ "LABEL_148": 148,
587
+ "LABEL_149": 149,
588
+ "LABEL_15": 15,
589
+ "LABEL_150": 150,
590
+ "LABEL_151": 151,
591
+ "LABEL_152": 152,
592
+ "LABEL_153": 153,
593
+ "LABEL_154": 154,
594
+ "LABEL_155": 155,
595
+ "LABEL_156": 156,
596
+ "LABEL_157": 157,
597
+ "LABEL_158": 158,
598
+ "LABEL_159": 159,
599
+ "LABEL_16": 16,
600
+ "LABEL_160": 160,
601
+ "LABEL_161": 161,
602
+ "LABEL_162": 162,
603
+ "LABEL_163": 163,
604
+ "LABEL_164": 164,
605
+ "LABEL_165": 165,
606
+ "LABEL_166": 166,
607
+ "LABEL_167": 167,
608
+ "LABEL_168": 168,
609
+ "LABEL_169": 169,
610
+ "LABEL_17": 17,
611
+ "LABEL_170": 170,
612
+ "LABEL_171": 171,
613
+ "LABEL_172": 172,
614
+ "LABEL_173": 173,
615
+ "LABEL_174": 174,
616
+ "LABEL_175": 175,
617
+ "LABEL_176": 176,
618
+ "LABEL_177": 177,
619
+ "LABEL_178": 178,
620
+ "LABEL_179": 179,
621
+ "LABEL_18": 18,
622
+ "LABEL_180": 180,
623
+ "LABEL_181": 181,
624
+ "LABEL_182": 182,
625
+ "LABEL_183": 183,
626
+ "LABEL_184": 184,
627
+ "LABEL_185": 185,
628
+ "LABEL_186": 186,
629
+ "LABEL_187": 187,
630
+ "LABEL_188": 188,
631
+ "LABEL_189": 189,
632
+ "LABEL_19": 19,
633
+ "LABEL_190": 190,
634
+ "LABEL_191": 191,
635
+ "LABEL_192": 192,
636
+ "LABEL_193": 193,
637
+ "LABEL_194": 194,
638
+ "LABEL_195": 195,
639
+ "LABEL_196": 196,
640
+ "LABEL_197": 197,
641
+ "LABEL_198": 198,
642
+ "LABEL_199": 199,
643
+ "LABEL_2": 2,
644
+ "LABEL_20": 20,
645
+ "LABEL_200": 200,
646
+ "LABEL_201": 201,
647
+ "LABEL_202": 202,
648
+ "LABEL_203": 203,
649
+ "LABEL_204": 204,
650
+ "LABEL_205": 205,
651
+ "LABEL_206": 206,
652
+ "LABEL_207": 207,
653
+ "LABEL_208": 208,
654
+ "LABEL_209": 209,
655
+ "LABEL_21": 21,
656
+ "LABEL_210": 210,
657
+ "LABEL_211": 211,
658
+ "LABEL_212": 212,
659
+ "LABEL_213": 213,
660
+ "LABEL_214": 214,
661
+ "LABEL_215": 215,
662
+ "LABEL_216": 216,
663
+ "LABEL_217": 217,
664
+ "LABEL_218": 218,
665
+ "LABEL_219": 219,
666
+ "LABEL_22": 22,
667
+ "LABEL_220": 220,
668
+ "LABEL_221": 221,
669
+ "LABEL_222": 222,
670
+ "LABEL_223": 223,
671
+ "LABEL_224": 224,
672
+ "LABEL_225": 225,
673
+ "LABEL_226": 226,
674
+ "LABEL_227": 227,
675
+ "LABEL_228": 228,
676
+ "LABEL_229": 229,
677
+ "LABEL_23": 23,
678
+ "LABEL_230": 230,
679
+ "LABEL_231": 231,
680
+ "LABEL_232": 232,
681
+ "LABEL_233": 233,
682
+ "LABEL_234": 234,
683
+ "LABEL_235": 235,
684
+ "LABEL_236": 236,
685
+ "LABEL_237": 237,
686
+ "LABEL_238": 238,
687
+ "LABEL_239": 239,
688
+ "LABEL_24": 24,
689
+ "LABEL_240": 240,
690
+ "LABEL_241": 241,
691
+ "LABEL_242": 242,
692
+ "LABEL_243": 243,
693
+ "LABEL_244": 244,
694
+ "LABEL_245": 245,
695
+ "LABEL_246": 246,
696
+ "LABEL_247": 247,
697
+ "LABEL_248": 248,
698
+ "LABEL_249": 249,
699
+ "LABEL_25": 25,
700
+ "LABEL_250": 250,
701
+ "LABEL_251": 251,
702
+ "LABEL_252": 252,
703
+ "LABEL_253": 253,
704
+ "LABEL_254": 254,
705
+ "LABEL_255": 255,
706
+ "LABEL_256": 256,
707
+ "LABEL_257": 257,
708
+ "LABEL_258": 258,
709
+ "LABEL_259": 259,
710
+ "LABEL_26": 26,
711
+ "LABEL_260": 260,
712
+ "LABEL_261": 261,
713
+ "LABEL_262": 262,
714
+ "LABEL_263": 263,
715
+ "LABEL_264": 264,
716
+ "LABEL_265": 265,
717
+ "LABEL_266": 266,
718
+ "LABEL_267": 267,
719
+ "LABEL_268": 268,
720
+ "LABEL_269": 269,
721
+ "LABEL_27": 27,
722
+ "LABEL_270": 270,
723
+ "LABEL_271": 271,
724
+ "LABEL_272": 272,
725
+ "LABEL_273": 273,
726
+ "LABEL_274": 274,
727
+ "LABEL_275": 275,
728
+ "LABEL_276": 276,
729
+ "LABEL_277": 277,
730
+ "LABEL_278": 278,
731
+ "LABEL_279": 279,
732
+ "LABEL_28": 28,
733
+ "LABEL_280": 280,
734
+ "LABEL_281": 281,
735
+ "LABEL_282": 282,
736
+ "LABEL_283": 283,
737
+ "LABEL_284": 284,
738
+ "LABEL_285": 285,
739
+ "LABEL_286": 286,
740
+ "LABEL_287": 287,
741
+ "LABEL_288": 288,
742
+ "LABEL_289": 289,
743
+ "LABEL_29": 29,
744
+ "LABEL_290": 290,
745
+ "LABEL_291": 291,
746
+ "LABEL_292": 292,
747
+ "LABEL_293": 293,
748
+ "LABEL_294": 294,
749
+ "LABEL_295": 295,
750
+ "LABEL_296": 296,
751
+ "LABEL_297": 297,
752
+ "LABEL_298": 298,
753
+ "LABEL_299": 299,
754
+ "LABEL_3": 3,
755
+ "LABEL_30": 30,
756
+ "LABEL_300": 300,
757
+ "LABEL_301": 301,
758
+ "LABEL_302": 302,
759
+ "LABEL_303": 303,
760
+ "LABEL_304": 304,
761
+ "LABEL_305": 305,
762
+ "LABEL_306": 306,
763
+ "LABEL_307": 307,
764
+ "LABEL_308": 308,
765
+ "LABEL_309": 309,
766
+ "LABEL_31": 31,
767
+ "LABEL_310": 310,
768
+ "LABEL_311": 311,
769
+ "LABEL_312": 312,
770
+ "LABEL_313": 313,
771
+ "LABEL_314": 314,
772
+ "LABEL_315": 315,
773
+ "LABEL_316": 316,
774
+ "LABEL_317": 317,
775
+ "LABEL_318": 318,
776
+ "LABEL_319": 319,
777
+ "LABEL_32": 32,
778
+ "LABEL_320": 320,
779
+ "LABEL_321": 321,
780
+ "LABEL_322": 322,
781
+ "LABEL_323": 323,
782
+ "LABEL_324": 324,
783
+ "LABEL_325": 325,
784
+ "LABEL_326": 326,
785
+ "LABEL_327": 327,
786
+ "LABEL_328": 328,
787
+ "LABEL_329": 329,
788
+ "LABEL_33": 33,
789
+ "LABEL_330": 330,
790
+ "LABEL_331": 331,
791
+ "LABEL_332": 332,
792
+ "LABEL_333": 333,
793
+ "LABEL_334": 334,
794
+ "LABEL_335": 335,
795
+ "LABEL_336": 336,
796
+ "LABEL_337": 337,
797
+ "LABEL_338": 338,
798
+ "LABEL_339": 339,
799
+ "LABEL_34": 34,
800
+ "LABEL_340": 340,
801
+ "LABEL_341": 341,
802
+ "LABEL_342": 342,
803
+ "LABEL_343": 343,
804
+ "LABEL_344": 344,
805
+ "LABEL_345": 345,
806
+ "LABEL_346": 346,
807
+ "LABEL_347": 347,
808
+ "LABEL_348": 348,
809
+ "LABEL_349": 349,
810
+ "LABEL_35": 35,
811
+ "LABEL_350": 350,
812
+ "LABEL_351": 351,
813
+ "LABEL_352": 352,
814
+ "LABEL_353": 353,
815
+ "LABEL_354": 354,
816
+ "LABEL_355": 355,
817
+ "LABEL_356": 356,
818
+ "LABEL_357": 357,
819
+ "LABEL_358": 358,
820
+ "LABEL_359": 359,
821
+ "LABEL_36": 36,
822
+ "LABEL_360": 360,
823
+ "LABEL_361": 361,
824
+ "LABEL_362": 362,
825
+ "LABEL_363": 363,
826
+ "LABEL_364": 364,
827
+ "LABEL_365": 365,
828
+ "LABEL_366": 366,
829
+ "LABEL_367": 367,
830
+ "LABEL_368": 368,
831
+ "LABEL_369": 369,
832
+ "LABEL_37": 37,
833
+ "LABEL_370": 370,
834
+ "LABEL_371": 371,
835
+ "LABEL_372": 372,
836
+ "LABEL_373": 373,
837
+ "LABEL_374": 374,
838
+ "LABEL_375": 375,
839
+ "LABEL_376": 376,
840
+ "LABEL_377": 377,
841
+ "LABEL_378": 378,
842
+ "LABEL_379": 379,
843
+ "LABEL_38": 38,
844
+ "LABEL_380": 380,
845
+ "LABEL_381": 381,
846
+ "LABEL_382": 382,
847
+ "LABEL_383": 383,
848
+ "LABEL_384": 384,
849
+ "LABEL_385": 385,
850
+ "LABEL_386": 386,
851
+ "LABEL_387": 387,
852
+ "LABEL_388": 388,
853
+ "LABEL_389": 389,
854
+ "LABEL_39": 39,
855
+ "LABEL_390": 390,
856
+ "LABEL_391": 391,
857
+ "LABEL_392": 392,
858
+ "LABEL_393": 393,
859
+ "LABEL_394": 394,
860
+ "LABEL_395": 395,
861
+ "LABEL_396": 396,
862
+ "LABEL_397": 397,
863
+ "LABEL_398": 398,
864
+ "LABEL_399": 399,
865
+ "LABEL_4": 4,
866
+ "LABEL_40": 40,
867
+ "LABEL_400": 400,
868
+ "LABEL_401": 401,
869
+ "LABEL_402": 402,
870
+ "LABEL_403": 403,
871
+ "LABEL_404": 404,
872
+ "LABEL_405": 405,
873
+ "LABEL_406": 406,
874
+ "LABEL_407": 407,
875
+ "LABEL_408": 408,
876
+ "LABEL_409": 409,
877
+ "LABEL_41": 41,
878
+ "LABEL_410": 410,
879
+ "LABEL_411": 411,
880
+ "LABEL_412": 412,
881
+ "LABEL_413": 413,
882
+ "LABEL_414": 414,
883
+ "LABEL_415": 415,
884
+ "LABEL_416": 416,
885
+ "LABEL_417": 417,
886
+ "LABEL_418": 418,
887
+ "LABEL_419": 419,
888
+ "LABEL_42": 42,
889
+ "LABEL_420": 420,
890
+ "LABEL_421": 421,
891
+ "LABEL_422": 422,
892
+ "LABEL_423": 423,
893
+ "LABEL_424": 424,
894
+ "LABEL_425": 425,
895
+ "LABEL_426": 426,
896
+ "LABEL_427": 427,
897
+ "LABEL_428": 428,
898
+ "LABEL_429": 429,
899
+ "LABEL_43": 43,
900
+ "LABEL_430": 430,
901
+ "LABEL_431": 431,
902
+ "LABEL_432": 432,
903
+ "LABEL_433": 433,
904
+ "LABEL_434": 434,
905
+ "LABEL_435": 435,
906
+ "LABEL_436": 436,
907
+ "LABEL_437": 437,
908
+ "LABEL_438": 438,
909
+ "LABEL_439": 439,
910
+ "LABEL_44": 44,
911
+ "LABEL_440": 440,
912
+ "LABEL_441": 441,
913
+ "LABEL_442": 442,
914
+ "LABEL_443": 443,
915
+ "LABEL_444": 444,
916
+ "LABEL_445": 445,
917
+ "LABEL_446": 446,
918
+ "LABEL_447": 447,
919
+ "LABEL_448": 448,
920
+ "LABEL_449": 449,
921
+ "LABEL_45": 45,
922
+ "LABEL_450": 450,
923
+ "LABEL_451": 451,
924
+ "LABEL_452": 452,
925
+ "LABEL_453": 453,
926
+ "LABEL_454": 454,
927
+ "LABEL_455": 455,
928
+ "LABEL_456": 456,
929
+ "LABEL_457": 457,
930
+ "LABEL_458": 458,
931
+ "LABEL_459": 459,
932
+ "LABEL_46": 46,
933
+ "LABEL_460": 460,
934
+ "LABEL_461": 461,
935
+ "LABEL_462": 462,
936
+ "LABEL_463": 463,
937
+ "LABEL_464": 464,
938
+ "LABEL_465": 465,
939
+ "LABEL_466": 466,
940
+ "LABEL_467": 467,
941
+ "LABEL_468": 468,
942
+ "LABEL_469": 469,
943
+ "LABEL_47": 47,
944
+ "LABEL_470": 470,
945
+ "LABEL_471": 471,
946
+ "LABEL_472": 472,
947
+ "LABEL_473": 473,
948
+ "LABEL_474": 474,
949
+ "LABEL_475": 475,
950
+ "LABEL_476": 476,
951
+ "LABEL_477": 477,
952
+ "LABEL_478": 478,
953
+ "LABEL_479": 479,
954
+ "LABEL_48": 48,
955
+ "LABEL_480": 480,
956
+ "LABEL_481": 481,
957
+ "LABEL_482": 482,
958
+ "LABEL_483": 483,
959
+ "LABEL_484": 484,
960
+ "LABEL_485": 485,
961
+ "LABEL_486": 486,
962
+ "LABEL_487": 487,
963
+ "LABEL_488": 488,
964
+ "LABEL_489": 489,
965
+ "LABEL_49": 49,
966
+ "LABEL_490": 490,
967
+ "LABEL_491": 491,
968
+ "LABEL_492": 492,
969
+ "LABEL_493": 493,
970
+ "LABEL_494": 494,
971
+ "LABEL_495": 495,
972
+ "LABEL_496": 496,
973
+ "LABEL_497": 497,
974
+ "LABEL_498": 498,
975
+ "LABEL_499": 499,
976
+ "LABEL_5": 5,
977
+ "LABEL_50": 50,
978
+ "LABEL_500": 500,
979
+ "LABEL_501": 501,
980
+ "LABEL_502": 502,
981
+ "LABEL_503": 503,
982
+ "LABEL_504": 504,
983
+ "LABEL_505": 505,
984
+ "LABEL_506": 506,
985
+ "LABEL_507": 507,
986
+ "LABEL_508": 508,
987
+ "LABEL_509": 509,
988
+ "LABEL_51": 51,
989
+ "LABEL_52": 52,
990
+ "LABEL_53": 53,
991
+ "LABEL_54": 54,
992
+ "LABEL_55": 55,
993
+ "LABEL_56": 56,
994
+ "LABEL_57": 57,
995
+ "LABEL_58": 58,
996
+ "LABEL_59": 59,
997
+ "LABEL_6": 6,
998
+ "LABEL_60": 60,
999
+ "LABEL_61": 61,
1000
+ "LABEL_62": 62,
1001
+ "LABEL_63": 63,
1002
+ "LABEL_64": 64,
1003
+ "LABEL_65": 65,
1004
+ "LABEL_66": 66,
1005
+ "LABEL_67": 67,
1006
+ "LABEL_68": 68,
1007
+ "LABEL_69": 69,
1008
+ "LABEL_7": 7,
1009
+ "LABEL_70": 70,
1010
+ "LABEL_71": 71,
1011
+ "LABEL_72": 72,
1012
+ "LABEL_73": 73,
1013
+ "LABEL_74": 74,
1014
+ "LABEL_75": 75,
1015
+ "LABEL_76": 76,
1016
+ "LABEL_77": 77,
1017
+ "LABEL_78": 78,
1018
+ "LABEL_79": 79,
1019
+ "LABEL_8": 8,
1020
+ "LABEL_80": 80,
1021
+ "LABEL_81": 81,
1022
+ "LABEL_82": 82,
1023
+ "LABEL_83": 83,
1024
+ "LABEL_84": 84,
1025
+ "LABEL_85": 85,
1026
+ "LABEL_86": 86,
1027
+ "LABEL_87": 87,
1028
+ "LABEL_88": 88,
1029
+ "LABEL_89": 89,
1030
+ "LABEL_9": 9,
1031
+ "LABEL_90": 90,
1032
+ "LABEL_91": 91,
1033
+ "LABEL_92": 92,
1034
+ "LABEL_93": 93,
1035
+ "LABEL_94": 94,
1036
+ "LABEL_95": 95,
1037
+ "LABEL_96": 96,
1038
+ "LABEL_97": 97,
1039
+ "LABEL_98": 98,
1040
+ "LABEL_99": 99
1041
+ },
1042
+ "layer_norm_eps": 1e-05,
1043
+ "max_2d_position_embeddings": 1024,
1044
+ "max_position_embeddings": 514,
1045
+ "max_rel_2d_pos": 256,
1046
+ "max_rel_pos": 128,
1047
+ "model_type": "layoutlmv3",
1048
+ "num_attention_heads": 16,
1049
+ "num_channels": 3,
1050
+ "num_hidden_layers": 24,
1051
+ "pad_token_id": 1,
1052
+ "patch_size": 16,
1053
+ "rel_2d_pos_bins": 64,
1054
+ "rel_pos_bins": 32,
1055
+ "second_input_size": 112,
1056
+ "shape_size": 170,
1057
+ "text_embed": true,
1058
+ "torch_dtype": "bfloat16",
1059
+ "transformers_version": "4.50.0",
1060
+ "type_vocab_size": 1,
1061
+ "visual_embed": false,
1062
+ "vocab_size": 50265
1063
+ }
Relation/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78a7ca1cb2ba8162b2672641f9d94ebde8b953fdf35c9417c0c8383e82751265
3
+ size 713217212
Structure/PP-DocLayout_plus-L/.msc ADDED
Binary file (366 Bytes). View file
 
Structure/PP-DocLayout_plus-L/.mv ADDED
@@ -0,0 +1 @@
 
 
1
+ Revision:master,CreatedAt:1751519173
Structure/PP-DocLayout_plus-L/README.md ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: PaddleOCR
4
+ language:
5
+ - en
6
+ - zh
7
+ pipeline_tag: image-to-text
8
+ tags:
9
+ - OCR
10
+ - PaddlePaddle
11
+ - PaddleOCR
12
+ ---
13
+
14
+ # PP-DocLayout_plus-L
15
+
16
+ ## Introduction
17
+
18
+ A higher-precision layout area localization model trained on a self-built dataset containing Chinese and English papers, PPT, multi-layout magazines, contracts, books, exams, ancient books and research reports using RT-DETR-L. The layout detection model includes 20 common categories: document title, paragraph title, text, page number, abstract, table, references, footnotes, header, footer, algorithm, formula, formula number, image, table, seal, figure_table title, chart, and sidebar text and lists of references. The key metrics are as follow:
19
+
20
+ | Model| mAP(0.5) (%) |
21
+ | --- | --- |
22
+ |PP-DocLayout_plus-L | 83.2 |
23
+
24
+ **Note**: the evaluation set of the above precision indicators is the self built version sub area detection data set, including Chinese and English papers, magazines, newspapers, research reports PPT、 1000 document type pictures such as test papers and textbooks.
25
+
26
+
27
+ ## Quick Start
28
+
29
+ ### Installation
30
+
31
+ 1. PaddlePaddle
32
+
33
+ Please refer to the following commands to install PaddlePaddle using pip:
34
+
35
+ ```bash
36
+ # for CUDA11.8
37
+ python -m pip install paddlepaddle-gpu==3.0.0 -i https://www.paddlepaddle.org.cn/packages/stable/cu118/
38
+
39
+ # for CUDA12.6
40
+ python -m pip install paddlepaddle-gpu==3.0.0 -i https://www.paddlepaddle.org.cn/packages/stable/cu126/
41
+
42
+ # for CPU
43
+ python -m pip install paddlepaddle==3.0.0 -i https://www.paddlepaddle.org.cn/packages/stable/cpu/
44
+ ```
45
+
46
+ For details about PaddlePaddle installation, please refer to the [PaddlePaddle official website](https://www.paddlepaddle.org.cn/en/install/quick).
47
+
48
+ 2. PaddleOCR
49
+
50
+ Install the latest version of the PaddleOCR inference package from PyPI:
51
+
52
+ ```bash
53
+ python -m pip install paddleocr
54
+ ```
55
+
56
+
57
+ ### Model Usage
58
+
59
+ You can quickly experience the functionality with a single command:
60
+
61
+ ```bash
62
+ paddleocr layout_detection \
63
+ --model_name PP-DocLayout_plus-L \
64
+ -i https://cdn-uploads.huggingface.co/production/uploads/63d7b8ee07cd1aa3c49a2026/N5C68HPVAI-xQAWTxpbA6.jpeg
65
+ ```
66
+
67
+ You can also integrate the model inference of the layout detection module into your project. Before running the following code, please download the sample image to your local machine.
68
+
69
+ ```python
70
+ from paddleocr import LayoutDetection
71
+
72
+ model = LayoutDetection(model_name="PP-DocLayout_plus-L")
73
+ output = model.predict("N5C68HPVAI-xQAWTxpbA6.jpeg", batch_size=1, layout_nms=True)
74
+ for res in output:
75
+ res.print()
76
+ res.save_to_img(save_path="./output/")
77
+ res.save_to_json(save_path="./output/res.json")
78
+ ```
79
+
80
+ After running, the obtained result is as follows:
81
+
82
+ ```json
83
+ {'res': {'input_path': '/root/.paddlex/predict_input/N5C68HPVAI-xQAWTxpbA6.jpeg', 'page_index': None, 'boxes': [{'cls_id': 2, 'label': 'text', 'score': 0.9870168566703796, 'coordinate': [34.101395, 349.85275, 358.5929, 611.0788]}, {'cls_id': 2, 'label': 'text', 'score': 0.986599326133728, 'coordinate': [34.500305, 647.15753, 358.29437, 848.66925]}, {'cls_id': 2, 'label': 'text', 'score': 0.984662652015686, 'coordinate': [385.71417, 497.41037, 711.22656, 697.8426]}, {'cls_id': 8, 'label': 'table', 'score': 0.9841272234916687, 'coordinate': [73.76732, 105.94854, 321.95355, 298.85074]}, {'cls_id': 8, 'label': 'table', 'score': 0.983431875705719, 'coordinate': [436.95523, 105.81446, 662.71814, 313.4865]}, {'cls_id': 2, 'label': 'text', 'score': 0.9832285642623901, 'coordinate': [385.62766, 346.22888, 710.10205, 458.772]}, {'cls_id': 2, 'label': 'text', 'score': 0.9816107749938965, 'coordinate': [385.78085, 735.19293, 710.5613, 849.97656]}, {'cls_id': 6, 'label': 'figure_title', 'score': 0.9577467441558838, 'coordinate': [34.421764, 20.055021, 358.7124, 76.53721]}, {'cls_id': 6, 'label': 'figure_title', 'score': 0.9505674839019775, 'coordinate': [385.7235, 20.054104, 711.2928, 74.92819]}, {'cls_id': 0, 'label': 'paragraph_title', 'score': 0.9001894593238831, 'coordinate': [386.46353, 477.035, 699.4023, 490.07495]}, {'cls_id': 0, 'label': 'paragraph_title', 'score': 0.8846081495285034, 'coordinate': [35.413055, 627.7365, 185.58315, 640.522]}, {'cls_id': 0, 'label': 'paragraph_title', 'score': 0.8837621808052063, 'coordinate': [387.1759, 716.34235, 524.78345, 729.2588]}, {'cls_id': 0, 'label': 'paragraph_title', 'score': 0.8509567975997925, 'coordinate': [35.50049, 331.18472, 141.64497, 344.81168]}]}}
84
+ ```
85
+
86
+ The visualized image is as follows:
87
+
88
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/63d7b8ee07cd1aa3c49a2026/5gAq1cFy1IX_wX26C2XmM.jpeg)
89
+
90
+ For details about usage command and descriptions of parameters, please refer to the [Document](https://paddlepaddle.github.io/PaddleOCR/latest/en/version3.x/module_usage/layout_detection.html#iii-quick-integration).
91
+
92
+ ### Pipeline Usage
93
+
94
+ The ability of a single model is limited. But the pipeline consists of several models can provide more capacity to resolve difficult problems in real-world scenarios.
95
+
96
+ #### PP-StructureV3
97
+
98
+ Layout analysis is a technique used to extract structured information from document images. PP-StructureV3 includes the following six modules:
99
+ * Layout Detection Module
100
+ * General OCR Sub-pipeline
101
+ * Document Image Preprocessing Sub-pipeline (Optional)
102
+ * Table Recognition Sub-pipeline (Optional)
103
+ * Seal Recognition Sub-pipeline (Optional)
104
+ * Formula Recognition Sub-pipeline (Optional)
105
+
106
+ You can quickly experience the PP-StructureV3 pipeline with a single command.
107
+
108
+ ```bash
109
+ paddleocr pp_structurev3 -i https://cdn-uploads.huggingface.co/production/uploads/63d7b8ee07cd1aa3c49a2026/KP10tiSZfAjMuwZUSLtRp.png
110
+ ```
111
+
112
+ You can experience the inference of the pipeline with just a few lines of code. Taking the PP-StructureV3 pipeline as an example:
113
+
114
+ ```python
115
+ from paddleocr import PPStructureV3
116
+
117
+ pipeline = PPStructureV3()
118
+ # ocr = PPStructureV3(use_doc_orientation_classify=True) # Use use_doc_orientation_classify to enable/disable document orientation classification model
119
+ # ocr = PPStructureV3(use_doc_unwarping=True) # Use use_doc_unwarping to enable/disable document unwarping module
120
+ # ocr = PPStructureV3(use_textline_orientation=True) # Use use_textline_orientation to enable/disable textline orientation classification model
121
+ # ocr = PPStructureV3(device="gpu") # Use device to specify GPU for model inference
122
+ output = pipeline.predict("./KP10tiSZfAjMuwZUSLtRp.png")
123
+ for res in output:
124
+ res.print() ## Print the structured prediction output
125
+ res.save_to_json(save_path="output") ## Save the current image's structured result in JSON format
126
+ res.save_to_markdown(save_path="output") ## Save the current image's result in Markdown format
127
+ ```
128
+
129
+ The default model used in pipeline is `PP-DocLayout_plus-L`.
130
+ For details about usage command and descriptions of parameters, please refer to the [Document](https://paddlepaddle.github.io/PaddleOCR/latest/en/version3.x/pipeline_usage/PP-StructureV3.html#2-quick-start).
131
+
132
+ ## Links
133
+
134
+ [PaddleOCR Repo](https://github.com/paddlepaddle/paddleocr)
135
+
136
+ [PaddleOCR Documentation](https://paddlepaddle.github.io/PaddleOCR/latest/en/index.html)
137
+
Structure/PP-DocLayout_plus-L/config.json ADDED
@@ -0,0 +1,227 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "mode": "paddle",
3
+ "draw_threshold": 0.5,
4
+ "metric": "COCO",
5
+ "use_dynamic_shape": false,
6
+ "Global": {
7
+ "model_name": "PP-DocLayout_plus-L"
8
+ },
9
+ "arch": "DETR",
10
+ "min_subgraph_size": 3,
11
+ "Preprocess": [
12
+ {
13
+ "interp": 2,
14
+ "keep_ratio": false,
15
+ "target_size": [
16
+ 800,
17
+ 800
18
+ ],
19
+ "type": "Resize"
20
+ },
21
+ {
22
+ "mean": [
23
+ 0.0,
24
+ 0.0,
25
+ 0.0
26
+ ],
27
+ "norm_type": "none",
28
+ "std": [
29
+ 1.0,
30
+ 1.0,
31
+ 1.0
32
+ ],
33
+ "type": "NormalizeImage"
34
+ },
35
+ {
36
+ "type": "Permute"
37
+ }
38
+ ],
39
+ "label_list": [
40
+ "paragraph_title",
41
+ "image",
42
+ "text",
43
+ "number",
44
+ "abstract",
45
+ "content",
46
+ "figure_title",
47
+ "formula",
48
+ "table",
49
+ "reference",
50
+ "doc_title",
51
+ "footnote",
52
+ "header",
53
+ "algorithm",
54
+ "footer",
55
+ "seal",
56
+ "chart",
57
+ "formula_number",
58
+ "aside_text",
59
+ "reference_content"
60
+ ],
61
+ "Hpi": {
62
+ "backend_configs": {
63
+ "paddle_infer": {
64
+ "trt_dynamic_shapes": {
65
+ "im_shape": [
66
+ [
67
+ 1,
68
+ 2
69
+ ],
70
+ [
71
+ 1,
72
+ 2
73
+ ],
74
+ [
75
+ 8,
76
+ 2
77
+ ]
78
+ ],
79
+ "image": [
80
+ [
81
+ 1,
82
+ 3,
83
+ 800,
84
+ 800
85
+ ],
86
+ [
87
+ 1,
88
+ 3,
89
+ 800,
90
+ 800
91
+ ],
92
+ [
93
+ 8,
94
+ 3,
95
+ 800,
96
+ 800
97
+ ]
98
+ ],
99
+ "scale_factor": [
100
+ [
101
+ 1,
102
+ 2
103
+ ],
104
+ [
105
+ 1,
106
+ 2
107
+ ],
108
+ [
109
+ 8,
110
+ 2
111
+ ]
112
+ ]
113
+ },
114
+ "trt_dynamic_shape_input_data": {
115
+ "im_shape": [
116
+ [
117
+ 800,
118
+ 800
119
+ ],
120
+ [
121
+ 800,
122
+ 800
123
+ ],
124
+ [
125
+ 800,
126
+ 800,
127
+ 800,
128
+ 800,
129
+ 800,
130
+ 800,
131
+ 800,
132
+ 800,
133
+ 800,
134
+ 800,
135
+ 800,
136
+ 800,
137
+ 800,
138
+ 800,
139
+ 800,
140
+ 800
141
+ ]
142
+ ],
143
+ "scale_factor": [
144
+ [
145
+ 2,
146
+ 2
147
+ ],
148
+ [
149
+ 1,
150
+ 1
151
+ ],
152
+ [
153
+ 0.67,
154
+ 0.67,
155
+ 0.67,
156
+ 0.67,
157
+ 0.67,
158
+ 0.67,
159
+ 0.67,
160
+ 0.67,
161
+ 0.67,
162
+ 0.67,
163
+ 0.67,
164
+ 0.67,
165
+ 0.67,
166
+ 0.67,
167
+ 0.67,
168
+ 0.67
169
+ ]
170
+ ]
171
+ }
172
+ },
173
+ "tensorrt": {
174
+ "dynamic_shapes": {
175
+ "im_shape": [
176
+ [
177
+ 1,
178
+ 2
179
+ ],
180
+ [
181
+ 1,
182
+ 2
183
+ ],
184
+ [
185
+ 8,
186
+ 2
187
+ ]
188
+ ],
189
+ "image": [
190
+ [
191
+ 1,
192
+ 3,
193
+ 800,
194
+ 800
195
+ ],
196
+ [
197
+ 1,
198
+ 3,
199
+ 800,
200
+ 800
201
+ ],
202
+ [
203
+ 8,
204
+ 3,
205
+ 800,
206
+ 800
207
+ ]
208
+ ],
209
+ "scale_factor": [
210
+ [
211
+ 1,
212
+ 2
213
+ ],
214
+ [
215
+ 1,
216
+ 2
217
+ ],
218
+ [
219
+ 8,
220
+ 2
221
+ ]
222
+ ]
223
+ }
224
+ }
225
+ }
226
+ }
227
+ }
Structure/PP-DocLayout_plus-L/inference.json ADDED
The diff for this file is too large to render. See raw diff
 
Structure/PP-DocLayout_plus-L/inference.pdiparams ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24ca3e2e442164505e250deef59f7ee9a54ea12dd32875c9cd6155d959dc97da
3
+ size 129307978
Structure/PP-DocLayout_plus-L/inference.yml ADDED
@@ -0,0 +1,123 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ mode: paddle
2
+ draw_threshold: 0.5
3
+ metric: COCO
4
+ use_dynamic_shape: false
5
+ Global:
6
+ model_name: PP-DocLayout_plus-L
7
+ arch: DETR
8
+ min_subgraph_size: 3
9
+ Preprocess:
10
+ - interp: 2
11
+ keep_ratio: false
12
+ target_size:
13
+ - 800
14
+ - 800
15
+ type: Resize
16
+ - mean:
17
+ - 0.0
18
+ - 0.0
19
+ - 0.0
20
+ norm_type: none
21
+ std:
22
+ - 1.0
23
+ - 1.0
24
+ - 1.0
25
+ type: NormalizeImage
26
+ - type: Permute
27
+ label_list:
28
+ - paragraph_title
29
+ - image
30
+ - text
31
+ - number
32
+ - abstract
33
+ - content
34
+ - figure_title
35
+ - formula
36
+ - table
37
+ - reference
38
+ - doc_title
39
+ - footnote
40
+ - header
41
+ - algorithm
42
+ - footer
43
+ - seal
44
+ - chart
45
+ - formula_number
46
+ - aside_text
47
+ - reference_content
48
+ Hpi:
49
+ backend_configs:
50
+ paddle_infer:
51
+ trt_dynamic_shapes: &id001
52
+ im_shape:
53
+ - - 1
54
+ - 2
55
+ - - 1
56
+ - 2
57
+ - - 8
58
+ - 2
59
+ image:
60
+ - - 1
61
+ - 3
62
+ - 800
63
+ - 800
64
+ - - 1
65
+ - 3
66
+ - 800
67
+ - 800
68
+ - - 8
69
+ - 3
70
+ - 800
71
+ - 800
72
+ scale_factor:
73
+ - - 1
74
+ - 2
75
+ - - 1
76
+ - 2
77
+ - - 8
78
+ - 2
79
+ trt_dynamic_shape_input_data:
80
+ im_shape:
81
+ - - 800
82
+ - 800
83
+ - - 800
84
+ - 800
85
+ - - 800
86
+ - 800
87
+ - 800
88
+ - 800
89
+ - 800
90
+ - 800
91
+ - 800
92
+ - 800
93
+ - 800
94
+ - 800
95
+ - 800
96
+ - 800
97
+ - 800
98
+ - 800
99
+ - 800
100
+ - 800
101
+ scale_factor:
102
+ - - 2
103
+ - 2
104
+ - - 1
105
+ - 1
106
+ - - 0.67
107
+ - 0.67
108
+ - 0.67
109
+ - 0.67
110
+ - 0.67
111
+ - 0.67
112
+ - 0.67
113
+ - 0.67
114
+ - 0.67
115
+ - 0.67
116
+ - 0.67
117
+ - 0.67
118
+ - 0.67
119
+ - 0.67
120
+ - 0.67
121
+ - 0.67
122
+ tensorrt:
123
+ dynamic_shapes: *id001
Structure/doclayout_yolo_docstructbench_imgsz1280_2501.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b152460888dc30be6db7f5dfab28bde3dcc999e5202f46187a764a1699c80be
3
+ size 39772550
Structure/layout_zh.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5acc32e5087ebb2601cf1221c7bdba960c086e1e4b009b15ce8b21c8e935fe3
3
+ size 40654210