admko commited on
Commit
146a20c
1 Parent(s): aa5481d

End of training

Browse files
README.md ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: distilbert-base-multilingual-cased
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - precision
8
+ - recall
9
+ - f1
10
+ - accuracy
11
+ model-index:
12
+ - name: sembr2023-distilbert-base-multilingual-cased
13
+ results: []
14
+ ---
15
+
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+ # sembr2023-distilbert-base-multilingual-cased
20
+
21
+ This model is a fine-tuned version of [distilbert-base-multilingual-cased](https://huggingface.co/distilbert-base-multilingual-cased) on an unknown dataset.
22
+ It achieves the following results on the evaluation set:
23
+ - Loss: 0.2282
24
+ - Precision: 0.7986
25
+ - Recall: 0.8244
26
+ - F1: 0.8113
27
+ - Iou: 0.6825
28
+ - Accuracy: 0.9666
29
+ - Balanced Accuracy: 0.9023
30
+ - Overall Accuracy: 0.9521
31
+
32
+ ## Model description
33
+
34
+ More information needed
35
+
36
+ ## Intended uses & limitations
37
+
38
+ More information needed
39
+
40
+ ## Training and evaluation data
41
+
42
+ More information needed
43
+
44
+ ## Training procedure
45
+
46
+ ### Training hyperparameters
47
+
48
+ The following hyperparameters were used during training:
49
+ - learning_rate: 0.0001
50
+ - train_batch_size: 64
51
+ - eval_batch_size: 128
52
+ - seed: 42
53
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
54
+ - lr_scheduler_type: cosine
55
+ - training_steps: 1000
56
+
57
+ ### Training results
58
+
59
+ | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Iou | Accuracy | Balanced Accuracy | Overall Accuracy |
60
+ |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:------:|:--------:|:-----------------:|:----------------:|
61
+ | 0.4021 | 0.06 | 10 | 0.3642 | 0 | 0.0 | 0.0 | 0.0 | 0.9130 | 0.5 | 0.9130 |
62
+ | 0.2804 | 0.12 | 20 | 0.2550 | 0.7462 | 0.5589 | 0.6391 | 0.4696 | 0.9451 | 0.7704 | 0.9331 |
63
+ | 0.2173 | 0.18 | 30 | 0.2038 | 0.9157 | 0.4420 | 0.5962 | 0.4247 | 0.9479 | 0.7191 | 0.9406 |
64
+ | 0.1831 | 0.24 | 40 | 0.1824 | 0.8317 | 0.7186 | 0.7710 | 0.6274 | 0.9629 | 0.8524 | 0.9455 |
65
+ | 0.1816 | 0.29 | 50 | 0.1829 | 0.7540 | 0.7857 | 0.7695 | 0.6254 | 0.9591 | 0.8807 | 0.9405 |
66
+ | 0.1461 | 0.35 | 60 | 0.1619 | 0.8520 | 0.7210 | 0.7810 | 0.6407 | 0.9648 | 0.8545 | 0.9498 |
67
+ | 0.1424 | 0.41 | 70 | 0.1568 | 0.8091 | 0.7717 | 0.7900 | 0.6528 | 0.9643 | 0.8772 | 0.9483 |
68
+ | 0.11 | 0.47 | 80 | 0.1550 | 0.8189 | 0.7760 | 0.7969 | 0.6624 | 0.9656 | 0.8798 | 0.9504 |
69
+ | 0.1373 | 0.53 | 90 | 0.1584 | 0.8248 | 0.7640 | 0.7933 | 0.6573 | 0.9654 | 0.8743 | 0.9484 |
70
+ | 0.1202 | 0.59 | 100 | 0.1612 | 0.8292 | 0.7608 | 0.7935 | 0.6577 | 0.9656 | 0.8729 | 0.9496 |
71
+ | 0.0966 | 0.65 | 110 | 0.1390 | 0.8575 | 0.7597 | 0.8057 | 0.6746 | 0.9681 | 0.8738 | 0.9551 |
72
+ | 0.0832 | 0.71 | 120 | 0.1569 | 0.8292 | 0.7858 | 0.8069 | 0.6763 | 0.9673 | 0.8852 | 0.9508 |
73
+ | 0.0914 | 0.76 | 130 | 0.1399 | 0.8295 | 0.8009 | 0.8150 | 0.6877 | 0.9684 | 0.8926 | 0.9529 |
74
+ | 0.0793 | 0.82 | 140 | 0.1456 | 0.8161 | 0.7914 | 0.8035 | 0.6716 | 0.9663 | 0.8872 | 0.9512 |
75
+ | 0.0903 | 0.88 | 150 | 0.1461 | 0.8218 | 0.7882 | 0.8047 | 0.6732 | 0.9667 | 0.8860 | 0.9523 |
76
+ | 0.0976 | 0.94 | 160 | 0.1380 | 0.7999 | 0.8304 | 0.8148 | 0.6875 | 0.9672 | 0.9053 | 0.9533 |
77
+ | 0.0776 | 1.0 | 170 | 0.1633 | 0.8053 | 0.8144 | 0.8099 | 0.6805 | 0.9667 | 0.8978 | 0.9492 |
78
+ | 0.0722 | 1.06 | 180 | 0.1581 | 0.8569 | 0.7739 | 0.8133 | 0.6853 | 0.9691 | 0.8808 | 0.9528 |
79
+ | 0.0787 | 1.12 | 190 | 0.1563 | 0.7780 | 0.8350 | 0.8055 | 0.6743 | 0.9649 | 0.9062 | 0.9495 |
80
+ | 0.0642 | 1.18 | 200 | 0.1483 | 0.8007 | 0.8115 | 0.8060 | 0.6751 | 0.9660 | 0.8961 | 0.9527 |
81
+ | 0.0582 | 1.24 | 210 | 0.1559 | 0.8244 | 0.8067 | 0.8154 | 0.6884 | 0.9682 | 0.8952 | 0.9534 |
82
+ | 0.0617 | 1.29 | 220 | 0.1359 | 0.8158 | 0.8115 | 0.8136 | 0.6858 | 0.9677 | 0.8970 | 0.9560 |
83
+ | 0.0508 | 1.35 | 230 | 0.1638 | 0.8372 | 0.7872 | 0.8114 | 0.6827 | 0.9682 | 0.8863 | 0.9548 |
84
+ | 0.049 | 1.41 | 240 | 0.1699 | 0.8585 | 0.7432 | 0.7967 | 0.6620 | 0.9670 | 0.8657 | 0.9523 |
85
+ | 0.0604 | 1.47 | 250 | 0.1452 | 0.7935 | 0.8374 | 0.8148 | 0.6875 | 0.9669 | 0.9083 | 0.9523 |
86
+ | 0.0512 | 1.53 | 260 | 0.1551 | 0.8353 | 0.7820 | 0.8078 | 0.6776 | 0.9676 | 0.8837 | 0.9542 |
87
+ | 0.0496 | 1.59 | 270 | 0.1686 | 0.7757 | 0.8490 | 0.8107 | 0.6816 | 0.9655 | 0.9128 | 0.9496 |
88
+ | 0.0486 | 1.65 | 280 | 0.1474 | 0.8464 | 0.7834 | 0.8137 | 0.6859 | 0.9688 | 0.8849 | 0.9557 |
89
+ | 0.0529 | 1.71 | 290 | 0.1470 | 0.8114 | 0.8136 | 0.8125 | 0.6842 | 0.9673 | 0.8978 | 0.9535 |
90
+ | 0.0585 | 1.76 | 300 | 0.1625 | 0.8193 | 0.8057 | 0.8124 | 0.6841 | 0.9676 | 0.8944 | 0.9527 |
91
+ | 0.0373 | 1.82 | 310 | 0.1414 | 0.8038 | 0.8240 | 0.8138 | 0.6860 | 0.9672 | 0.9024 | 0.9555 |
92
+ | 0.0474 | 1.88 | 320 | 0.1600 | 0.8117 | 0.8073 | 0.8095 | 0.6800 | 0.9670 | 0.8947 | 0.9541 |
93
+ | 0.0475 | 1.94 | 330 | 0.1825 | 0.7709 | 0.8418 | 0.8048 | 0.6733 | 0.9645 | 0.9090 | 0.9498 |
94
+ | 0.0596 | 2.0 | 340 | 0.1688 | 0.8185 | 0.8078 | 0.8131 | 0.6851 | 0.9677 | 0.8954 | 0.9533 |
95
+ | 0.0391 | 2.06 | 350 | 0.1775 | 0.8006 | 0.8142 | 0.8073 | 0.6769 | 0.9662 | 0.8974 | 0.9514 |
96
+ | 0.0285 | 2.12 | 360 | 0.1653 | 0.8047 | 0.8165 | 0.8105 | 0.6814 | 0.9668 | 0.8988 | 0.9536 |
97
+ | 0.0288 | 2.18 | 370 | 0.1855 | 0.7697 | 0.8376 | 0.8022 | 0.6698 | 0.9641 | 0.9069 | 0.9509 |
98
+ | 0.0244 | 2.24 | 380 | 0.1813 | 0.8053 | 0.8171 | 0.8112 | 0.6823 | 0.9669 | 0.8991 | 0.9530 |
99
+ | 0.0324 | 2.29 | 390 | 0.1663 | 0.8029 | 0.8252 | 0.8139 | 0.6862 | 0.9672 | 0.9030 | 0.9538 |
100
+ | 0.0306 | 2.35 | 400 | 0.1692 | 0.7949 | 0.8291 | 0.8117 | 0.6830 | 0.9665 | 0.9044 | 0.9529 |
101
+ | 0.0279 | 2.41 | 410 | 0.1812 | 0.8171 | 0.8056 | 0.8113 | 0.6825 | 0.9674 | 0.8942 | 0.9526 |
102
+ | 0.0287 | 2.47 | 420 | 0.1768 | 0.8196 | 0.8106 | 0.8151 | 0.6879 | 0.9680 | 0.8968 | 0.9540 |
103
+ | 0.0287 | 2.53 | 430 | 0.1849 | 0.7927 | 0.8271 | 0.8095 | 0.6800 | 0.9662 | 0.9032 | 0.9532 |
104
+ | 0.0328 | 2.59 | 440 | 0.1765 | 0.8031 | 0.8197 | 0.8113 | 0.6825 | 0.9668 | 0.9003 | 0.9536 |
105
+ | 0.0226 | 2.65 | 450 | 0.1928 | 0.7879 | 0.8403 | 0.8133 | 0.6853 | 0.9664 | 0.9094 | 0.9515 |
106
+ | 0.0308 | 2.71 | 460 | 0.1905 | 0.7858 | 0.8354 | 0.8098 | 0.6804 | 0.9659 | 0.9069 | 0.9515 |
107
+ | 0.0277 | 2.76 | 470 | 0.1890 | 0.8083 | 0.8168 | 0.8125 | 0.6842 | 0.9672 | 0.8992 | 0.9537 |
108
+ | 0.0229 | 2.82 | 480 | 0.1802 | 0.8125 | 0.8177 | 0.8151 | 0.6879 | 0.9677 | 0.8999 | 0.9540 |
109
+ | 0.0218 | 2.88 | 490 | 0.1895 | 0.7805 | 0.8370 | 0.8078 | 0.6775 | 0.9654 | 0.9073 | 0.9506 |
110
+ | 0.0283 | 2.94 | 500 | 0.1926 | 0.7854 | 0.8337 | 0.8088 | 0.6790 | 0.9657 | 0.9060 | 0.9516 |
111
+ | 0.0231 | 3.0 | 510 | 0.2023 | 0.7766 | 0.8432 | 0.8085 | 0.6786 | 0.9653 | 0.9101 | 0.9509 |
112
+ | 0.0172 | 3.06 | 520 | 0.2051 | 0.7811 | 0.8305 | 0.8051 | 0.6737 | 0.9650 | 0.9042 | 0.9509 |
113
+ | 0.0206 | 3.12 | 530 | 0.1918 | 0.7894 | 0.8339 | 0.8110 | 0.6822 | 0.9662 | 0.9064 | 0.9523 |
114
+ | 0.0243 | 3.18 | 540 | 0.1982 | 0.7992 | 0.8198 | 0.8094 | 0.6798 | 0.9664 | 0.9001 | 0.9526 |
115
+ | 0.0193 | 3.24 | 550 | 0.2036 | 0.8024 | 0.8165 | 0.8094 | 0.6798 | 0.9666 | 0.8987 | 0.9517 |
116
+ | 0.0212 | 3.29 | 560 | 0.1967 | 0.8093 | 0.8175 | 0.8134 | 0.6854 | 0.9674 | 0.8996 | 0.9535 |
117
+ | 0.0188 | 3.35 | 570 | 0.1944 | 0.8056 | 0.8188 | 0.8122 | 0.6837 | 0.9671 | 0.9000 | 0.9527 |
118
+ | 0.0176 | 3.41 | 580 | 0.1975 | 0.8014 | 0.8239 | 0.8125 | 0.6842 | 0.9669 | 0.9022 | 0.9528 |
119
+ | 0.0197 | 3.47 | 590 | 0.2118 | 0.8058 | 0.8186 | 0.8122 | 0.6837 | 0.9671 | 0.8999 | 0.9529 |
120
+ | 0.0142 | 3.53 | 600 | 0.2000 | 0.8107 | 0.8187 | 0.8147 | 0.6873 | 0.9676 | 0.9003 | 0.9540 |
121
+ | 0.0145 | 3.59 | 610 | 0.2095 | 0.7950 | 0.8278 | 0.8111 | 0.6822 | 0.9665 | 0.9037 | 0.9522 |
122
+ | 0.023 | 3.65 | 620 | 0.2107 | 0.7881 | 0.8268 | 0.8070 | 0.6764 | 0.9656 | 0.9028 | 0.9511 |
123
+ | 0.0156 | 3.71 | 630 | 0.2191 | 0.7814 | 0.8366 | 0.8081 | 0.6780 | 0.9654 | 0.9072 | 0.9507 |
124
+ | 0.0175 | 3.76 | 640 | 0.2090 | 0.8081 | 0.8202 | 0.8141 | 0.6865 | 0.9674 | 0.9008 | 0.9528 |
125
+ | 0.02 | 3.82 | 650 | 0.2160 | 0.8068 | 0.8224 | 0.8145 | 0.6871 | 0.9674 | 0.9018 | 0.9524 |
126
+ | 0.0148 | 3.88 | 660 | 0.2049 | 0.7972 | 0.8306 | 0.8136 | 0.6857 | 0.9669 | 0.9052 | 0.9526 |
127
+ | 0.0184 | 3.94 | 670 | 0.2122 | 0.7908 | 0.8342 | 0.8119 | 0.6834 | 0.9664 | 0.9066 | 0.9513 |
128
+ | 0.0178 | 4.0 | 680 | 0.2090 | 0.7925 | 0.8271 | 0.8094 | 0.6799 | 0.9661 | 0.9032 | 0.9519 |
129
+ | 0.0159 | 4.06 | 690 | 0.2137 | 0.7963 | 0.8293 | 0.8125 | 0.6842 | 0.9667 | 0.9045 | 0.9528 |
130
+ | 0.0172 | 4.12 | 700 | 0.2150 | 0.7917 | 0.8317 | 0.8112 | 0.6824 | 0.9663 | 0.9054 | 0.9520 |
131
+ | 0.0163 | 4.18 | 710 | 0.2174 | 0.8010 | 0.8250 | 0.8128 | 0.6846 | 0.9670 | 0.9027 | 0.9524 |
132
+ | 0.0117 | 4.24 | 720 | 0.2175 | 0.8036 | 0.8229 | 0.8131 | 0.6851 | 0.9671 | 0.9019 | 0.9530 |
133
+ | 0.0149 | 4.29 | 730 | 0.2225 | 0.7990 | 0.8242 | 0.8114 | 0.6826 | 0.9667 | 0.9022 | 0.9515 |
134
+ | 0.014 | 4.35 | 740 | 0.2157 | 0.7916 | 0.8273 | 0.8090 | 0.6793 | 0.9660 | 0.9033 | 0.9520 |
135
+ | 0.0139 | 4.41 | 750 | 0.2217 | 0.7952 | 0.8231 | 0.8089 | 0.6792 | 0.9662 | 0.9015 | 0.9512 |
136
+ | 0.0143 | 4.47 | 760 | 0.2201 | 0.7914 | 0.8317 | 0.8111 | 0.6822 | 0.9663 | 0.9054 | 0.9518 |
137
+ | 0.0147 | 4.53 | 770 | 0.2218 | 0.7945 | 0.8264 | 0.8101 | 0.6809 | 0.9663 | 0.9030 | 0.9517 |
138
+ | 0.0124 | 4.59 | 780 | 0.2300 | 0.7915 | 0.8337 | 0.8120 | 0.6835 | 0.9664 | 0.9064 | 0.9515 |
139
+ | 0.0157 | 4.65 | 790 | 0.2244 | 0.8038 | 0.8191 | 0.8114 | 0.6826 | 0.9669 | 0.9000 | 0.9520 |
140
+ | 0.0126 | 4.71 | 800 | 0.2324 | 0.7983 | 0.8256 | 0.8117 | 0.6831 | 0.9667 | 0.9029 | 0.9514 |
141
+ | 0.0137 | 4.76 | 810 | 0.2263 | 0.8000 | 0.8226 | 0.8112 | 0.6823 | 0.9667 | 0.9015 | 0.9520 |
142
+ | 0.0133 | 4.82 | 820 | 0.2295 | 0.7932 | 0.8299 | 0.8111 | 0.6823 | 0.9664 | 0.9046 | 0.9512 |
143
+ | 0.0155 | 4.88 | 830 | 0.2266 | 0.8019 | 0.8226 | 0.8121 | 0.6837 | 0.9669 | 0.9016 | 0.9520 |
144
+ | 0.0135 | 4.94 | 840 | 0.2286 | 0.7932 | 0.8318 | 0.8121 | 0.6836 | 0.9665 | 0.9056 | 0.9518 |
145
+ | 0.0073 | 5.0 | 850 | 0.2283 | 0.8019 | 0.8235 | 0.8125 | 0.6843 | 0.9670 | 0.9021 | 0.9520 |
146
+ | 0.0147 | 5.06 | 860 | 0.2292 | 0.7965 | 0.8290 | 0.8124 | 0.6841 | 0.9667 | 0.9044 | 0.9518 |
147
+ | 0.0139 | 5.12 | 870 | 0.2293 | 0.8013 | 0.8251 | 0.8130 | 0.6850 | 0.9670 | 0.9028 | 0.9520 |
148
+ | 0.0162 | 5.18 | 880 | 0.2277 | 0.7977 | 0.8261 | 0.8116 | 0.6830 | 0.9667 | 0.9031 | 0.9518 |
149
+ | 0.0111 | 5.24 | 890 | 0.2285 | 0.7941 | 0.8288 | 0.8111 | 0.6822 | 0.9664 | 0.9042 | 0.9517 |
150
+ | 0.0128 | 5.29 | 900 | 0.2273 | 0.7987 | 0.8240 | 0.8112 | 0.6823 | 0.9666 | 0.9021 | 0.9521 |
151
+ | 0.0129 | 5.35 | 910 | 0.2282 | 0.7989 | 0.8244 | 0.8114 | 0.6827 | 0.9667 | 0.9023 | 0.9520 |
152
+ | 0.0115 | 5.41 | 920 | 0.2285 | 0.7991 | 0.8246 | 0.8117 | 0.6830 | 0.9667 | 0.9024 | 0.9520 |
153
+ | 0.011 | 5.47 | 930 | 0.2283 | 0.7999 | 0.8242 | 0.8119 | 0.6833 | 0.9668 | 0.9023 | 0.9522 |
154
+ | 0.013 | 5.53 | 940 | 0.2281 | 0.7995 | 0.8239 | 0.8115 | 0.6828 | 0.9667 | 0.9021 | 0.9522 |
155
+ | 0.0132 | 5.59 | 950 | 0.2281 | 0.7993 | 0.8239 | 0.8114 | 0.6826 | 0.9667 | 0.9021 | 0.9522 |
156
+ | 0.0104 | 5.65 | 960 | 0.2281 | 0.7992 | 0.8244 | 0.8116 | 0.6829 | 0.9667 | 0.9023 | 0.9522 |
157
+ | 0.0143 | 5.71 | 970 | 0.2279 | 0.7988 | 0.8245 | 0.8114 | 0.6827 | 0.9667 | 0.9024 | 0.9521 |
158
+ | 0.0133 | 5.76 | 980 | 0.2280 | 0.7988 | 0.8241 | 0.8113 | 0.6825 | 0.9667 | 0.9022 | 0.9521 |
159
+ | 0.0079 | 5.82 | 990 | 0.2282 | 0.7986 | 0.8244 | 0.8113 | 0.6825 | 0.9666 | 0.9023 | 0.9521 |
160
+ | 0.0108 | 5.88 | 1000 | 0.2282 | 0.7986 | 0.8244 | 0.8113 | 0.6825 | 0.9666 | 0.9023 | 0.9521 |
161
+
162
+
163
+ ### Framework versions
164
+
165
+ - Transformers 4.34.1
166
+ - Pytorch 2.0.1
167
+ - Datasets 2.14.6
168
+ - Tokenizers 0.14.1
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "[newline]": 119548,
3
+ "[percent]": 119547
4
+ }
config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-multilingual-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForTokenClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "off",
13
+ "1": "space-0",
14
+ "2": "space-1",
15
+ "3": "space-2",
16
+ "4": "space-3",
17
+ "5": "nospace-0",
18
+ "6": "nospace-1",
19
+ "7": "nospace-2",
20
+ "8": "nospace-3"
21
+ },
22
+ "initializer_range": 0.02,
23
+ "label2id": {
24
+ "nospace-0": 5,
25
+ "nospace-1": 6,
26
+ "nospace-2": 7,
27
+ "nospace-3": 8,
28
+ "off": 0,
29
+ "space-0": 1,
30
+ "space-1": 2,
31
+ "space-2": 3,
32
+ "space-3": 4
33
+ },
34
+ "max_indent": 3,
35
+ "max_position_embeddings": 512,
36
+ "model_type": "distilbert",
37
+ "n_heads": 12,
38
+ "n_layers": 6,
39
+ "output_past": true,
40
+ "pad_token_id": 0,
41
+ "qa_dropout": 0.1,
42
+ "seq_classif_dropout": 0.2,
43
+ "sinusoidal_pos_embds": false,
44
+ "tie_weights_": true,
45
+ "torch_dtype": "float32",
46
+ "transformers_version": "4.34.1",
47
+ "vocab_size": 119549
48
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1a2572ec6403ebf80c583f159b5a3c895b1a6d17cb75fba69c5234746962759
3
+ size 539004965
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "119547": {
44
+ "content": "[percent]",
45
+ "lstrip": false,
46
+ "normalized": true,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": false
50
+ },
51
+ "119548": {
52
+ "content": "[newline]",
53
+ "lstrip": false,
54
+ "normalized": true,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": false
58
+ }
59
+ },
60
+ "clean_up_tokenization_spaces": true,
61
+ "cls_token": "[CLS]",
62
+ "do_lower_case": false,
63
+ "mask_token": "[MASK]",
64
+ "model_max_length": 512,
65
+ "pad_token": "[PAD]",
66
+ "sep_token": "[SEP]",
67
+ "strip_accents": null,
68
+ "tokenize_chinese_chars": true,
69
+ "tokenizer_class": "DistilBertTokenizer",
70
+ "unk_token": "[UNK]"
71
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48b8bb1f1df30f10b77fd707ae10f8f3989df3b876002e0eb799597e17378015
3
+ size 4219
vocab.txt ADDED
The diff for this file is too large to render. See raw diff