Upload tokenizer
Browse files- tokenizer.json +1 -12
tokenizer.json
CHANGED
@@ -13,18 +13,7 @@
|
|
13 |
"special": true
|
14 |
}
|
15 |
],
|
16 |
-
"normalizer":
|
17 |
-
"type": "Sequence",
|
18 |
-
"normalizers": [
|
19 |
-
{
|
20 |
-
"type": "Replace",
|
21 |
-
"pattern": {
|
22 |
-
"String": "\n"
|
23 |
-
},
|
24 |
-
"content": "<|endoftext|>"
|
25 |
-
}
|
26 |
-
]
|
27 |
-
},
|
28 |
"pre_tokenizer": {
|
29 |
"type": "Split",
|
30 |
"pattern": {
|
|
|
13 |
"special": true
|
14 |
}
|
15 |
],
|
16 |
+
"normalizer": null,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
17 |
"pre_tokenizer": {
|
18 |
"type": "Split",
|
19 |
"pattern": {
|