{ "_name_or_path": "dicta-il/dictabert", "architectures": [ "BertForJointParsing" ], "attention_probs_dropout_prob": 0.1, "classifier_dropout": null, "do_lex": true, "do_morph": true, "do_ner": true, "do_prefix": true, "do_syntax": true, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "B-ANG", "1": "B-DUC", "2": "B-EVE", "3": "B-FAC", "4": "B-GPE", "5": "B-LOC", "6": "B-ORG", "7": "B-PER", "8": "B-WOA", "9": "B-INFORMAL", "10": "B-MISC", "11": "B-TIMEX", "12": "B-TTL", "13": "I-DUC", "14": "I-EVE", "15": "I-FAC", "16": "I-GPE", "17": "I-LOC", "18": "I-ORG", "19": "I-PER", "20": "I-WOA", "21": "I-ANG", "22": "I-INFORMAL", "23": "I-MISC", "24": "I-TIMEX", "25": "I-TTL", "26": "O" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "B-ANG": 0, "B-DUC": 1, "B-EVE": 2, "B-FAC": 3, "B-GPE": 4, "B-INFORMAL": 9, "B-LOC": 5, "B-MISC": 10, "B-ORG": 6, "B-PER": 7, "B-TIMEX": 11, "B-TTL": 12, "B-WOA": 8, "I-ANG": 21, "I-DUC": 13, "I-EVE": 14, "I-FAC": 15, "I-GPE": 16, "I-INFORMAL": 22, "I-LOC": 17, "I-MISC": 23, "I-ORG": 18, "I-PER": 19, "I-TIMEX": 24, "I-TTL": 25, "I-WOA": 20, "O": 26 }, "layer_norm_eps": 1e-12, "max_position_embeddings": 512, "model_type": "bert", "newmodern": true, "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 0, "position_embedding_type": "absolute", "syntax_head_size": 128, "torch_dtype": "float32", "transformers_version": "4.36.2", "type_vocab_size": 2, "use_cache": true, "vocab_size": 128000 }