ArgumentMining-EN-ARI-AIF-ALBERT / tokenizer_config.json
yevhenkost's picture
add tokenizer
fcc4b45
raw
history blame
502 Bytes
{"do_lower_case": false, "remove_space": true, "keep_accents": false, "bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "<unk>", "sep_token": "[SEP]", "pad_token": "<pad>", "cls_token": "[CLS]", "mask_token": {"content": "[MASK]", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false, "__type": "AddedToken"}, "sp_model_kwargs": {}, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "final-albert-ari-outputs/", "tokenizer_class": "AlbertTokenizer"}