{ "_name_or_path": "roberta-large-ttc", "adapters": { "adapters": {}, "config_map": {}, "fusion_config_map": {}, "fusions": {} }, "architectures": [ "RobertaForTokenClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "classifier_dropout": null, "crf_state_dict": { "_constraint_mask": [ [ 1.0, 1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0 ], [ 1.0, 1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0 ], [ 1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0 ], [ 1.0, 1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0 ], [ 1.0, 1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0 ], [ 1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0 ], [ 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0 ], [ 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0 ], [ 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0 ] ], "end_transitions": [ 0.8981873393058777, -1.6446324586868286, 0.15959225594997406, -1.065422773361206, 0.08138537406921387, 0.7564309239387512, 0.1831926703453064 ], "start_transitions": [ -1.4693340063095093, -0.9572128653526306, 0.21432557702064514, 0.6372387409210205, 0.9176182746887207, 0.3472508192062378, 0.4026823341846466 ], "transitions": [ [ -0.2170916646718979, -0.3218635320663452, 0.18208211660385132, -0.12818297743797302, 0.34018412232398987, -0.48195144534111023, -0.04889260604977608 ], [ 0.7136948108673096, -0.2505597174167633, -0.38005825877189636, -0.40438419580459595, 0.41300636529922485, -1.0484856367111206, 0.42295384407043457 ], [ 0.25146323442459106, 0.1469486951828003, -0.8152630925178528, -0.38353827595710754, 0.04554833471775055, -0.15171411633491516, -0.4340072572231293 ], [ 0.6300148963928223, -0.15564066171646118, -0.01947428472340107, -0.06949518620967865, -0.6198987364768982, -0.059607941657304764, 0.1473998874425888 ], [ 0.39721569418907166, 0.3447458744049072, -0.2558501064777374, 0.19512270390987396, 0.1556796282529831, -0.32401445508003235, -0.2883574068546295 ], [ 0.36640703678131104, 0.24679042398929596, 0.08182600140571594, -0.23832643032073975, -0.03733787685632706, -0.12589092552661896, -0.1488221138715744 ], [ 0.21845334768295288, -0.14681760966777802, 0.059233345091342926, -0.251701682806015, 0.04643300548195839, -0.06884881108999252, 0.1407143473625183 ] ] }, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 1024, "id2label": { "0": "B-LOC", "1": "B-ORG", "2": "B-PER", "3": "I-LOC", "4": "I-ORG", "5": "I-PER", "6": "O" }, "initializer_range": 0.02, "intermediate_size": 4096, "label2id": { "B-LOC": 0, "B-ORG": 1, "B-PER": 2, "I-LOC": 3, "I-ORG": 4, "I-PER": 5, "O": 6 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 16, "num_hidden_layers": 24, "pad_token_id": 1, "position_embedding_type": "absolute", "torch_dtype": "float32", "transformers_version": "4.11.3", "type_vocab_size": 1, "use_cache": true, "vocab_size": 50265 }