stefan-it's picture
tokenizer: add config (no accent stripping) and vocab
6cd52d3
raw
history blame contribute delete
82 Bytes
{"do_lower_case": true, "max_len": 512, "init_inputs": [], "strip_accents":false}