{ "_name_or_path": "hfunakura/bert-base-semtagger-gold", "architectures": [ "BertForTokenClassification" ], "attention_probs_dropout_prob": 0.1, "classifier_dropout": null, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "@@UNK@@", "1": "PRO", "2": "CTC", "3": "INT", "4": "EMP", "5": "DEC", "6": "ITJ", "7": "GRE", "8": "NEC", "9": "PFT", "10": "IMP", "11": "HAP", "12": "ROL", "13": "MOY", "14": "PRG", "15": "HAS", "16": "CLO", "17": "MOR", "18": "DEF", "19": "BUT", "20": "YOC", "21": "PRI", "22": "EQU", "23": "SUB", "24": "APX", "25": "REL", "26": "XCL", "27": "CON", "28": "GPO", "29": "QUE", "30": "DIS", "31": "IST", "32": "COL", "33": "SCO", "34": "GRP", "35": "EXS", "36": "FUT", "37": "ENS", "38": "QUC", "39": "DOM", "40": "SST", "41": "NIL", "42": "COO", "43": "QUV", "44": "PST", "45": "UNK", "46": "EXT", "47": "NTH", "48": "LIT", "49": "ORG", "50": "EXG", "51": "REF", "52": "DOW", "53": "TOP", "54": "EPS", "55": "DXT", "56": "AND", "57": "UOM", "58": "ALT", "59": "POS", "60": "PRX", "61": "GEO", "62": "BOT", "63": "DEG", "64": "ART", "65": "PER", "66": "GPE", "67": "EFS", "68": "DST", "69": "LES", "70": "ORD", "71": "NOT", "72": "NOW" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "@@UNK@@": 0, "ALT": 58, "AND": 56, "APX": 24, "ART": 64, "BOT": 62, "BUT": 19, "CLO": 16, "COL": 32, "CON": 27, "COO": 42, "CTC": 2, "DEC": 5, "DEF": 18, "DEG": 63, "DIS": 30, "DOM": 39, "DOW": 52, "DST": 68, "DXT": 55, "EFS": 67, "EMP": 4, "ENS": 37, "EPS": 54, "EQU": 22, "EXG": 50, "EXS": 35, "EXT": 46, "FUT": 36, "GEO": 61, "GPE": 66, "GPO": 28, "GRE": 7, "GRP": 34, "HAP": 11, "HAS": 15, "IMP": 10, "INT": 3, "IST": 31, "ITJ": 6, "LES": 69, "LIT": 48, "MOR": 17, "MOY": 13, "NEC": 8, "NIL": 41, "NOT": 71, "NOW": 72, "NTH": 47, "ORD": 70, "ORG": 49, "PER": 65, "PFT": 9, "POS": 59, "PRG": 14, "PRI": 21, "PRO": 1, "PRX": 60, "PST": 44, "QUC": 38, "QUE": 29, "QUV": 43, "REF": 51, "REL": 25, "ROL": 12, "SCO": 33, "SST": 40, "SUB": 23, "TOP": 53, "UNK": 45, "UOM": 57, "XCL": 26, "YOC": 20 }, "layer_norm_eps": 1e-12, "max_position_embeddings": 512, "model_type": "bert", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 0, "position_embedding_type": "absolute", "torch_dtype": "float32", "transformers_version": "4.30.0", "type_vocab_size": 2, "use_cache": true, "vocab_size": 30522 }