asahi417 commited on
Commit
aedcade
·
1 Parent(s): 9cbafcc

model update

Browse files
config.json ADDED
@@ -0,0 +1,550 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "tner_output/multiconer/multi/xlm_roberta_large/model_tpkhxk/epoch_10",
3
+ "adapters": {
4
+ "adapters": {},
5
+ "config_map": {},
6
+ "fusion_config_map": {},
7
+ "fusions": {}
8
+ },
9
+ "architectures": [
10
+ "XLMRobertaForTokenClassification"
11
+ ],
12
+ "attention_probs_dropout_prob": 0.1,
13
+ "bos_token_id": 0,
14
+ "classifier_dropout": null,
15
+ "crf_state_dict": {
16
+ "_constraint_mask": [
17
+ [
18
+ 1.0,
19
+ 1.0,
20
+ 0.0,
21
+ 1.0,
22
+ 0.0,
23
+ 1.0,
24
+ 0.0,
25
+ 1.0,
26
+ 1.0,
27
+ 0.0,
28
+ 1.0,
29
+ 0.0,
30
+ 0.0,
31
+ 0.0,
32
+ 1.0
33
+ ],
34
+ [
35
+ 1.0,
36
+ 1.0,
37
+ 1.0,
38
+ 1.0,
39
+ 0.0,
40
+ 1.0,
41
+ 0.0,
42
+ 1.0,
43
+ 1.0,
44
+ 0.0,
45
+ 1.0,
46
+ 0.0,
47
+ 0.0,
48
+ 0.0,
49
+ 1.0
50
+ ],
51
+ [
52
+ 1.0,
53
+ 1.0,
54
+ 1.0,
55
+ 1.0,
56
+ 0.0,
57
+ 1.0,
58
+ 0.0,
59
+ 1.0,
60
+ 1.0,
61
+ 0.0,
62
+ 1.0,
63
+ 0.0,
64
+ 0.0,
65
+ 0.0,
66
+ 1.0
67
+ ],
68
+ [
69
+ 1.0,
70
+ 1.0,
71
+ 0.0,
72
+ 1.0,
73
+ 1.0,
74
+ 1.0,
75
+ 0.0,
76
+ 1.0,
77
+ 1.0,
78
+ 0.0,
79
+ 1.0,
80
+ 0.0,
81
+ 0.0,
82
+ 0.0,
83
+ 1.0
84
+ ],
85
+ [
86
+ 1.0,
87
+ 1.0,
88
+ 0.0,
89
+ 1.0,
90
+ 1.0,
91
+ 1.0,
92
+ 0.0,
93
+ 1.0,
94
+ 1.0,
95
+ 0.0,
96
+ 1.0,
97
+ 0.0,
98
+ 0.0,
99
+ 0.0,
100
+ 1.0
101
+ ],
102
+ [
103
+ 1.0,
104
+ 1.0,
105
+ 0.0,
106
+ 1.0,
107
+ 0.0,
108
+ 1.0,
109
+ 1.0,
110
+ 1.0,
111
+ 1.0,
112
+ 0.0,
113
+ 1.0,
114
+ 0.0,
115
+ 0.0,
116
+ 0.0,
117
+ 1.0
118
+ ],
119
+ [
120
+ 1.0,
121
+ 1.0,
122
+ 0.0,
123
+ 1.0,
124
+ 0.0,
125
+ 1.0,
126
+ 1.0,
127
+ 1.0,
128
+ 1.0,
129
+ 0.0,
130
+ 1.0,
131
+ 0.0,
132
+ 0.0,
133
+ 0.0,
134
+ 1.0
135
+ ],
136
+ [
137
+ 1.0,
138
+ 1.0,
139
+ 0.0,
140
+ 1.0,
141
+ 0.0,
142
+ 1.0,
143
+ 0.0,
144
+ 1.0,
145
+ 1.0,
146
+ 0.0,
147
+ 1.0,
148
+ 1.0,
149
+ 0.0,
150
+ 0.0,
151
+ 1.0
152
+ ],
153
+ [
154
+ 1.0,
155
+ 1.0,
156
+ 0.0,
157
+ 1.0,
158
+ 0.0,
159
+ 1.0,
160
+ 0.0,
161
+ 1.0,
162
+ 1.0,
163
+ 1.0,
164
+ 1.0,
165
+ 0.0,
166
+ 0.0,
167
+ 0.0,
168
+ 1.0
169
+ ],
170
+ [
171
+ 1.0,
172
+ 1.0,
173
+ 0.0,
174
+ 1.0,
175
+ 0.0,
176
+ 1.0,
177
+ 0.0,
178
+ 1.0,
179
+ 1.0,
180
+ 1.0,
181
+ 1.0,
182
+ 0.0,
183
+ 0.0,
184
+ 0.0,
185
+ 1.0
186
+ ],
187
+ [
188
+ 1.0,
189
+ 1.0,
190
+ 0.0,
191
+ 1.0,
192
+ 0.0,
193
+ 1.0,
194
+ 0.0,
195
+ 1.0,
196
+ 1.0,
197
+ 0.0,
198
+ 1.0,
199
+ 0.0,
200
+ 1.0,
201
+ 0.0,
202
+ 1.0
203
+ ],
204
+ [
205
+ 1.0,
206
+ 1.0,
207
+ 0.0,
208
+ 1.0,
209
+ 0.0,
210
+ 1.0,
211
+ 0.0,
212
+ 1.0,
213
+ 1.0,
214
+ 0.0,
215
+ 1.0,
216
+ 1.0,
217
+ 0.0,
218
+ 0.0,
219
+ 1.0
220
+ ],
221
+ [
222
+ 1.0,
223
+ 1.0,
224
+ 0.0,
225
+ 1.0,
226
+ 0.0,
227
+ 1.0,
228
+ 0.0,
229
+ 1.0,
230
+ 1.0,
231
+ 0.0,
232
+ 1.0,
233
+ 0.0,
234
+ 1.0,
235
+ 0.0,
236
+ 1.0
237
+ ],
238
+ [
239
+ 1.0,
240
+ 1.0,
241
+ 0.0,
242
+ 1.0,
243
+ 0.0,
244
+ 1.0,
245
+ 0.0,
246
+ 1.0,
247
+ 1.0,
248
+ 0.0,
249
+ 1.0,
250
+ 0.0,
251
+ 0.0,
252
+ 0.0,
253
+ 0.0
254
+ ],
255
+ [
256
+ 0.0,
257
+ 0.0,
258
+ 0.0,
259
+ 0.0,
260
+ 0.0,
261
+ 0.0,
262
+ 0.0,
263
+ 0.0,
264
+ 0.0,
265
+ 0.0,
266
+ 0.0,
267
+ 0.0,
268
+ 0.0,
269
+ 0.0,
270
+ 0.0
271
+ ]
272
+ ],
273
+ "end_transitions": [
274
+ -1.0464372634887695,
275
+ 1.1780683994293213,
276
+ -0.018892239779233932,
277
+ 1.130867838859558,
278
+ -0.8810096979141235,
279
+ 0.3867743909358978,
280
+ -0.2203485667705536,
281
+ 1.6087849140167236,
282
+ 0.630409300327301,
283
+ 0.38923677802085876,
284
+ 0.7220190763473511,
285
+ -0.3640587031841278,
286
+ 0.917939305305481
287
+ ],
288
+ "start_transitions": [
289
+ 1.2037677764892578,
290
+ -0.5295335650444031,
291
+ 1.743898868560791,
292
+ 0.07551981508731842,
293
+ 0.7501091361045837,
294
+ -0.03813719376921654,
295
+ 0.7444982528686523,
296
+ -0.3991338014602661,
297
+ 0.06231020390987396,
298
+ -0.5034055709838867,
299
+ -0.5263239741325378,
300
+ -1.381019949913025,
301
+ -0.28310921788215637
302
+ ],
303
+ "transitions": [
304
+ [
305
+ 0.09528060257434845,
306
+ -0.37277305126190186,
307
+ -0.25866812467575073,
308
+ 0.24822545051574707,
309
+ -0.24260400235652924,
310
+ -0.141363725066185,
311
+ -0.2408754527568817,
312
+ 0.5676054358482361,
313
+ 0.12529665231704712,
314
+ 0.04831164330244064,
315
+ -0.5429413318634033,
316
+ -0.2707440257072449,
317
+ -0.46986308693885803
318
+ ],
319
+ [
320
+ -0.11546444892883301,
321
+ 0.15093672275543213,
322
+ 0.10780151188373566,
323
+ -0.5896725654602051,
324
+ -0.6444293260574341,
325
+ 0.15699714422225952,
326
+ -0.05136333778500557,
327
+ 0.2992676794528961,
328
+ 0.4846979081630707,
329
+ 0.009639686904847622,
330
+ 0.41389113664627075,
331
+ 0.09152292460203171,
332
+ 0.7565149068832397
333
+ ],
334
+ [
335
+ -0.6054845452308655,
336
+ 0.5026692152023315,
337
+ 0.08101499825716019,
338
+ -0.09596530348062515,
339
+ -0.22281970083713531,
340
+ 0.3282443881034851,
341
+ -0.1347975730895996,
342
+ 0.0668039470911026,
343
+ -0.09752220660448074,
344
+ -0.07354750484228134,
345
+ 0.04112263023853302,
346
+ -0.38143306970596313,
347
+ 0.1599552035331726
348
+ ],
349
+ [
350
+ -0.19424262642860413,
351
+ -0.40223875641822815,
352
+ -0.11459892243146896,
353
+ 0.02577541582286358,
354
+ 0.3284653425216675,
355
+ -0.007798475679010153,
356
+ -0.12146443873643875,
357
+ -0.09523932635784149,
358
+ 0.007428421173244715,
359
+ 0.14787055552005768,
360
+ -0.1024009957909584,
361
+ 0.00570573378354311,
362
+ -0.15338386595249176
363
+ ],
364
+ [
365
+ 0.12268232554197311,
366
+ -0.3437243103981018,
367
+ 0.20003697276115417,
368
+ 0.4476252794265747,
369
+ 0.06421514600515366,
370
+ -0.592639148235321,
371
+ 0.3522963523864746,
372
+ 0.027350544929504395,
373
+ -0.10489096492528915,
374
+ 0.08435048162937164,
375
+ 0.07331147789955139,
376
+ -0.08593357354402542,
377
+ 0.03731408342719078
378
+ ],
379
+ [
380
+ 0.00524252001196146,
381
+ 0.03470692038536072,
382
+ -0.28712740540504456,
383
+ -0.1656406819820404,
384
+ -0.38783204555511475,
385
+ -0.06701929122209549,
386
+ 0.36923933029174805,
387
+ -0.2814996540546417,
388
+ -0.06497164070606232,
389
+ 0.5502147674560547,
390
+ 0.23615996539592743,
391
+ -0.038583990186452866,
392
+ 0.46467122435569763
393
+ ],
394
+ [
395
+ 0.008220738731324673,
396
+ 0.0994560569524765,
397
+ -0.22741952538490295,
398
+ 0.2880842387676239,
399
+ 0.14621806144714355,
400
+ 0.5893053412437439,
401
+ 0.0887163057923317,
402
+ 0.07120154798030853,
403
+ 0.3175446391105652,
404
+ 0.07388962805271149,
405
+ -0.3354664444923401,
406
+ -0.36641693115234375,
407
+ -0.06141874939203262
408
+ ],
409
+ [
410
+ -0.23988734185695648,
411
+ -0.42511773109436035,
412
+ -0.10005861520767212,
413
+ 0.1286901831626892,
414
+ 0.39106130599975586,
415
+ -0.2108324021100998,
416
+ -0.06749864667654037,
417
+ 0.18266062438488007,
418
+ -0.24625848233699799,
419
+ 0.23925840854644775,
420
+ 0.32043731212615967,
421
+ -0.2222135365009308,
422
+ -0.46956998109817505
423
+ ],
424
+ [
425
+ 0.32192981243133545,
426
+ -0.091096892952919,
427
+ -0.042639218270778656,
428
+ 0.28457701206207275,
429
+ -0.4472516179084778,
430
+ -0.3262014091014862,
431
+ -0.13654400408267975,
432
+ -0.14421388506889343,
433
+ 0.013985628262162209,
434
+ 0.08464387059211731,
435
+ 0.21588052809238434,
436
+ -0.0952681303024292,
437
+ -0.3312382400035858
438
+ ],
439
+ [
440
+ 0.057129133492708206,
441
+ 0.23515474796295166,
442
+ 0.24695280194282532,
443
+ -0.39212411642074585,
444
+ -0.298391729593277,
445
+ -0.24707596004009247,
446
+ 0.19766144454479218,
447
+ 0.30783167481422424,
448
+ 0.61564040184021,
449
+ 0.38697120547294617,
450
+ -0.24159395694732666,
451
+ -0.15458329021930695,
452
+ 0.20215679705142975
453
+ ],
454
+ [
455
+ -0.1168244481086731,
456
+ 0.1461421251296997,
457
+ 0.2691043019294739,
458
+ -0.5000207424163818,
459
+ 0.04115551337599754,
460
+ -0.16865290701389313,
461
+ 0.025549300014972687,
462
+ 0.009987183846533298,
463
+ 0.08454478532075882,
464
+ 0.22551782429218292,
465
+ -0.22243990004062653,
466
+ -0.5178296566009521,
467
+ -0.17889747023582458
468
+ ],
469
+ [
470
+ -0.0819055363535881,
471
+ 0.28591519594192505,
472
+ -0.3139297664165497,
473
+ -0.24641193449497223,
474
+ -0.19268277287483215,
475
+ 0.09901709854602814,
476
+ 0.2405221164226532,
477
+ -0.06007663905620575,
478
+ -0.036537591367959976,
479
+ -0.2676612138748169,
480
+ -0.11241313815116882,
481
+ 0.24845600128173828,
482
+ -0.0591643862426281
483
+ ],
484
+ [
485
+ 0.24317403137683868,
486
+ 0.526745080947876,
487
+ 0.4228650629520416,
488
+ 0.11840634793043137,
489
+ -0.23573878407478333,
490
+ 0.163038432598114,
491
+ 0.10626911371946335,
492
+ 0.14786396920681,
493
+ 0.598574697971344,
494
+ 0.019770774990320206,
495
+ 0.24404427409172058,
496
+ -0.28538644313812256,
497
+ -0.12149752676486969
498
+ ]
499
+ ]
500
+ },
501
+ "eos_token_id": 2,
502
+ "hidden_act": "gelu",
503
+ "hidden_dropout_prob": 0.1,
504
+ "hidden_size": 1024,
505
+ "id2label": {
506
+ "0": "O",
507
+ "1": "B-person",
508
+ "2": "I-person",
509
+ "3": "B-group",
510
+ "4": "I-group",
511
+ "5": "B-work of art",
512
+ "6": "I-work of art",
513
+ "7": "B-location",
514
+ "8": "B-corporation",
515
+ "9": "I-corporation",
516
+ "10": "B-product",
517
+ "11": "I-location",
518
+ "12": "I-product"
519
+ },
520
+ "initializer_range": 0.02,
521
+ "intermediate_size": 4096,
522
+ "label2id": {
523
+ "B-corporation": 8,
524
+ "B-group": 3,
525
+ "B-location": 7,
526
+ "B-person": 1,
527
+ "B-product": 10,
528
+ "B-work of art": 5,
529
+ "I-corporation": 9,
530
+ "I-group": 4,
531
+ "I-location": 11,
532
+ "I-person": 2,
533
+ "I-product": 12,
534
+ "I-work of art": 6,
535
+ "O": 0
536
+ },
537
+ "layer_norm_eps": 1e-05,
538
+ "max_position_embeddings": 514,
539
+ "model_type": "xlm-roberta",
540
+ "num_attention_heads": 16,
541
+ "num_hidden_layers": 24,
542
+ "output_past": true,
543
+ "pad_token_id": 1,
544
+ "position_embedding_type": "absolute",
545
+ "torch_dtype": "float32",
546
+ "transformers_version": "4.11.3",
547
+ "type_vocab_size": 1,
548
+ "use_cache": true,
549
+ "vocab_size": 250002
550
+ }
eval/metric.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"valid": {"micro/f1": 0.8693724015046524, "micro/recall": 0.8773275793175098, "micro/precision": 0.8615601946319259, "macro/f1": 0.8649791381276803, "macro/recall": 0.872779649803229, "macro/precision": 0.857449153082705, "per_entity_metric": {"corporation": {"precision": 0.8776102088167054, "recall": 0.8705408515535098, "f1": 0.8740612362796073}, "group": {"precision": 0.8564621798689697, "recall": 0.8778998778998779, "f1": 0.8670485378353935}, "location": {"precision": 0.8926421404682274, "recall": 0.910300136425648, "f1": 0.9013846673421142}, "person": {"precision": 0.9132825719120136, "recall": 0.921861656703672, "f1": 0.91755206119847}, "product": {"precision": 0.7850899742930592, "recall": 0.8262987012987013, "f1": 0.805167413656736}, "work of art": {"precision": 0.8196078431372549, "recall": 0.8297766749379653, "f1": 0.8246609124537608}}}}
eval/my_submission.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a12b7ea0bbd623f272d6f39c60489a5993ab97b6bcc1ce8e4c9f06660557a6ba
3
+ size 1499931
eval/prediction.valid.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fe31ef4aa3ab97dbc4a043276baf99078ca4f426b11cefae1450d34656e5846
3
+ size 2235587185
sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
3
+ size 5069051
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false}}
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "tner_output/multiconer/multi/xlm_roberta_large/model_tpkhxk/epoch_10", "tokenizer_class": "XLMRobertaTokenizer"}