ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k6_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8110
  • Qwk: 0.6535
  • Mse: 0.8110
  • Rmse: 0.9006

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0625 2 5.3844 -0.0007 5.3844 2.3204
No log 0.125 4 3.9217 0.0383 3.9217 1.9803
No log 0.1875 6 2.2289 0.1065 2.2289 1.4929
No log 0.25 8 2.1066 -0.0205 2.1066 1.4514
No log 0.3125 10 1.8636 0.0313 1.8636 1.3651
No log 0.375 12 1.1524 0.2115 1.1524 1.0735
No log 0.4375 14 1.1698 0.3324 1.1698 1.0816
No log 0.5 16 1.1904 0.3240 1.1904 1.0910
No log 0.5625 18 1.0889 0.2332 1.0889 1.0435
No log 0.625 20 1.1004 0.2115 1.1004 1.0490
No log 0.6875 22 1.0934 0.2080 1.0934 1.0457
No log 0.75 24 1.0741 0.2896 1.0741 1.0364
No log 0.8125 26 1.0624 0.3101 1.0624 1.0307
No log 0.875 28 1.0725 0.3972 1.0725 1.0356
No log 0.9375 30 0.9757 0.3822 0.9757 0.9878
No log 1.0 32 0.9084 0.3667 0.9084 0.9531
No log 1.0625 34 0.9896 0.2322 0.9896 0.9948
No log 1.125 36 1.4602 0.0166 1.4602 1.2084
No log 1.1875 38 1.4314 0.0558 1.4314 1.1964
No log 1.25 40 0.9263 0.4055 0.9263 0.9624
No log 1.3125 42 1.1152 0.3228 1.1152 1.0560
No log 1.375 44 1.7099 0.2089 1.7099 1.3076
No log 1.4375 46 1.8020 0.2023 1.8020 1.3424
No log 1.5 48 1.5302 0.2239 1.5302 1.2370
No log 1.5625 50 1.0929 0.3228 1.0929 1.0454
No log 1.625 52 0.8997 0.4645 0.8997 0.9485
No log 1.6875 54 0.8755 0.5566 0.8755 0.9357
No log 1.75 56 0.8120 0.5575 0.8120 0.9011
No log 1.8125 58 0.7467 0.6284 0.7467 0.8641
No log 1.875 60 0.6988 0.6308 0.6988 0.8360
No log 1.9375 62 0.7103 0.6353 0.7103 0.8428
No log 2.0 64 0.9418 0.5729 0.9418 0.9705
No log 2.0625 66 0.9890 0.5886 0.9890 0.9945
No log 2.125 68 0.9588 0.6107 0.9588 0.9792
No log 2.1875 70 0.9071 0.6108 0.9071 0.9524
No log 2.25 72 0.6961 0.6760 0.6961 0.8343
No log 2.3125 74 0.6922 0.6987 0.6922 0.8320
No log 2.375 76 0.7839 0.6683 0.7839 0.8854
No log 2.4375 78 0.7893 0.6798 0.7893 0.8884
No log 2.5 80 0.7390 0.6886 0.7390 0.8596
No log 2.5625 82 1.0370 0.5685 1.0370 1.0183
No log 2.625 84 1.0171 0.5841 1.0171 1.0085
No log 2.6875 86 0.7566 0.6809 0.7566 0.8698
No log 2.75 88 0.7145 0.6416 0.7145 0.8453
No log 2.8125 90 0.8564 0.6110 0.8564 0.9254
No log 2.875 92 0.7612 0.5952 0.7612 0.8725
No log 2.9375 94 0.6394 0.6232 0.6394 0.7996
No log 3.0 96 0.6195 0.6703 0.6195 0.7871
No log 3.0625 98 0.6545 0.6540 0.6545 0.8090
No log 3.125 100 0.6717 0.6509 0.6717 0.8195
No log 3.1875 102 0.6339 0.6869 0.6339 0.7962
No log 3.25 104 0.7026 0.6763 0.7026 0.8382
No log 3.3125 106 0.8779 0.6265 0.8779 0.9369
No log 3.375 108 1.0884 0.5542 1.0884 1.0433
No log 3.4375 110 0.8730 0.6206 0.8730 0.9344
No log 3.5 112 0.6497 0.6768 0.6497 0.8060
No log 3.5625 114 0.6891 0.6416 0.6891 0.8301
No log 3.625 116 0.6981 0.6192 0.6981 0.8355
No log 3.6875 118 0.6717 0.6408 0.6717 0.8196
No log 3.75 120 0.6997 0.5691 0.6997 0.8365
No log 3.8125 122 0.7485 0.5088 0.7485 0.8652
No log 3.875 124 0.8046 0.5167 0.8046 0.8970
No log 3.9375 126 0.9736 0.5326 0.9736 0.9867
No log 4.0 128 1.2351 0.2990 1.2351 1.1114
No log 4.0625 130 1.2003 0.3859 1.2003 1.0956
No log 4.125 132 0.9471 0.5191 0.9471 0.9732
No log 4.1875 134 0.7354 0.5683 0.7354 0.8575
No log 4.25 136 0.7089 0.5859 0.7089 0.8420
No log 4.3125 138 0.7920 0.6153 0.7920 0.8899
No log 4.375 140 0.7509 0.6647 0.7509 0.8665
No log 4.4375 142 0.6597 0.7010 0.6597 0.8122
No log 4.5 144 0.6526 0.7254 0.6526 0.8078
No log 4.5625 146 0.6192 0.6999 0.6192 0.7869
No log 4.625 148 0.6232 0.7042 0.6232 0.7894
No log 4.6875 150 0.5952 0.6957 0.5952 0.7715
No log 4.75 152 0.6080 0.6788 0.6080 0.7797
No log 4.8125 154 0.6463 0.6732 0.6463 0.8039
No log 4.875 156 0.6296 0.6839 0.6296 0.7935
No log 4.9375 158 0.6296 0.6788 0.6296 0.7935
No log 5.0 160 0.6714 0.6795 0.6714 0.8194
No log 5.0625 162 0.6559 0.6970 0.6559 0.8099
No log 5.125 164 0.6875 0.7013 0.6875 0.8292
No log 5.1875 166 0.7153 0.7014 0.7153 0.8457
No log 5.25 168 0.7516 0.6880 0.7516 0.8669
No log 5.3125 170 0.7774 0.6657 0.7774 0.8817
No log 5.375 172 0.7181 0.6803 0.7181 0.8474
No log 5.4375 174 0.7758 0.6843 0.7758 0.8808
No log 5.5 176 0.8746 0.6593 0.8746 0.9352
No log 5.5625 178 0.8438 0.6571 0.8438 0.9186
No log 5.625 180 0.7118 0.6545 0.7118 0.8437
No log 5.6875 182 0.6350 0.7218 0.6350 0.7969
No log 5.75 184 0.6249 0.7177 0.6249 0.7905
No log 5.8125 186 0.6252 0.7000 0.6252 0.7907
No log 5.875 188 0.6185 0.7045 0.6185 0.7865
No log 5.9375 190 0.6091 0.7130 0.6091 0.7804
No log 6.0 192 0.6412 0.6820 0.6412 0.8008
No log 6.0625 194 0.6115 0.6884 0.6115 0.7820
No log 6.125 196 0.6404 0.6925 0.6404 0.8002
No log 6.1875 198 0.7120 0.6787 0.7120 0.8438
No log 6.25 200 0.7661 0.6476 0.7661 0.8753
No log 6.3125 202 0.6717 0.6580 0.6717 0.8196
No log 6.375 204 0.6693 0.6427 0.6693 0.8181
No log 6.4375 206 0.6671 0.6133 0.6671 0.8168
No log 6.5 208 0.6789 0.6601 0.6789 0.8239
No log 6.5625 210 0.8382 0.6223 0.8382 0.9155
No log 6.625 212 0.8678 0.5910 0.8678 0.9316
No log 6.6875 214 0.7113 0.6520 0.7113 0.8434
No log 6.75 216 0.6530 0.6054 0.6530 0.8081
No log 6.8125 218 0.7539 0.6308 0.7539 0.8683
No log 6.875 220 0.7216 0.6364 0.7216 0.8495
No log 6.9375 222 0.6175 0.6509 0.6175 0.7858
No log 7.0 224 0.6433 0.7223 0.6433 0.8020
No log 7.0625 226 0.6750 0.7020 0.6750 0.8216
No log 7.125 228 0.6318 0.7355 0.6318 0.7948
No log 7.1875 230 0.6341 0.7073 0.6341 0.7963
No log 7.25 232 0.6423 0.7148 0.6423 0.8014
No log 7.3125 234 0.6967 0.7068 0.6967 0.8347
No log 7.375 236 0.8553 0.6026 0.8553 0.9248
No log 7.4375 238 0.9150 0.5679 0.9150 0.9566
No log 7.5 240 0.8781 0.5899 0.8781 0.9371
No log 7.5625 242 0.7495 0.6703 0.7495 0.8658
No log 7.625 244 0.7188 0.7135 0.7188 0.8478
No log 7.6875 246 0.7854 0.6492 0.7854 0.8862
No log 7.75 248 0.8179 0.6638 0.8179 0.9044
No log 7.8125 250 0.7609 0.7141 0.7609 0.8723
No log 7.875 252 0.6988 0.6827 0.6988 0.8359
No log 7.9375 254 0.7043 0.6832 0.7043 0.8392
No log 8.0 256 0.7420 0.6794 0.7420 0.8614
No log 8.0625 258 0.8430 0.6135 0.8430 0.9181
No log 8.125 260 0.8420 0.6068 0.8420 0.9176
No log 8.1875 262 0.7439 0.6601 0.7439 0.8625
No log 8.25 264 0.6886 0.6764 0.6886 0.8298
No log 8.3125 266 0.6815 0.6661 0.6815 0.8255
No log 8.375 268 0.7278 0.6853 0.7278 0.8531
No log 8.4375 270 0.7744 0.7096 0.7744 0.8800
No log 8.5 272 0.7232 0.7063 0.7232 0.8504
No log 8.5625 274 0.6607 0.7074 0.6607 0.8128
No log 8.625 276 0.6593 0.7052 0.6593 0.8120
No log 8.6875 278 0.7128 0.6971 0.7128 0.8443
No log 8.75 280 0.7888 0.6676 0.7888 0.8882
No log 8.8125 282 0.7429 0.6874 0.7429 0.8619
No log 8.875 284 0.6792 0.6974 0.6792 0.8241
No log 8.9375 286 0.6438 0.6911 0.6438 0.8024
No log 9.0 288 0.6642 0.6657 0.6642 0.8150
No log 9.0625 290 0.6400 0.6737 0.6400 0.8000
No log 9.125 292 0.6825 0.7263 0.6825 0.8262
No log 9.1875 294 0.8287 0.6543 0.8287 0.9103
No log 9.25 296 0.8409 0.6695 0.8409 0.9170
No log 9.3125 298 0.7399 0.6704 0.7399 0.8602
No log 9.375 300 0.6872 0.6943 0.6872 0.8290
No log 9.4375 302 0.6793 0.6957 0.6793 0.8242
No log 9.5 304 0.7228 0.6402 0.7228 0.8502
No log 9.5625 306 0.8447 0.6373 0.8447 0.9191
No log 9.625 308 0.9367 0.6205 0.9367 0.9678
No log 9.6875 310 0.8635 0.6818 0.8635 0.9293
No log 9.75 312 0.7064 0.7583 0.7064 0.8405
No log 9.8125 314 0.6485 0.7525 0.6485 0.8053
No log 9.875 316 0.6345 0.7259 0.6345 0.7965
No log 9.9375 318 0.6421 0.6909 0.6421 0.8013
No log 10.0 320 0.6574 0.6616 0.6574 0.8108
No log 10.0625 322 0.6562 0.6782 0.6562 0.8100
No log 10.125 324 0.6637 0.6895 0.6637 0.8147
No log 10.1875 326 0.6679 0.7193 0.6679 0.8172
No log 10.25 328 0.6747 0.7283 0.6747 0.8214
No log 10.3125 330 0.6934 0.7058 0.6934 0.8327
No log 10.375 332 0.7240 0.6269 0.7240 0.8509
No log 10.4375 334 0.7126 0.6560 0.7126 0.8442
No log 10.5 336 0.6738 0.6660 0.6738 0.8209
No log 10.5625 338 0.7271 0.7002 0.7271 0.8527
No log 10.625 340 0.7607 0.6846 0.7607 0.8722
No log 10.6875 342 0.7544 0.7098 0.7544 0.8685
No log 10.75 344 0.6993 0.7294 0.6993 0.8362
No log 10.8125 346 0.6716 0.6598 0.6716 0.8195
No log 10.875 348 0.7102 0.6558 0.7102 0.8427
No log 10.9375 350 0.6697 0.6372 0.6697 0.8183
No log 11.0 352 0.6541 0.7405 0.6541 0.8087
No log 11.0625 354 0.7343 0.7021 0.7343 0.8569
No log 11.125 356 0.8312 0.6002 0.8312 0.9117
No log 11.1875 358 0.7945 0.5937 0.7945 0.8913
No log 11.25 360 0.7262 0.6433 0.7262 0.8522
No log 11.3125 362 0.6582 0.6641 0.6582 0.8113
No log 11.375 364 0.6394 0.6913 0.6394 0.7997
No log 11.4375 366 0.6334 0.6928 0.6334 0.7959
No log 11.5 368 0.6394 0.7164 0.6394 0.7996
No log 11.5625 370 0.6694 0.7203 0.6694 0.8182
No log 11.625 372 0.7632 0.7117 0.7632 0.8736
No log 11.6875 374 0.8389 0.6615 0.8389 0.9159
No log 11.75 376 0.7600 0.6906 0.7600 0.8718
No log 11.8125 378 0.7130 0.7100 0.7130 0.8444
No log 11.875 380 0.7139 0.6855 0.7139 0.8449
No log 11.9375 382 0.7449 0.6578 0.7449 0.8631
No log 12.0 384 0.8071 0.6357 0.8071 0.8984
No log 12.0625 386 0.7932 0.6306 0.7932 0.8906
No log 12.125 388 0.7347 0.6561 0.7347 0.8571
No log 12.1875 390 0.6973 0.7055 0.6973 0.8351
No log 12.25 392 0.6726 0.7208 0.6726 0.8201
No log 12.3125 394 0.6521 0.7464 0.6521 0.8075
No log 12.375 396 0.6488 0.7404 0.6488 0.8055
No log 12.4375 398 0.6461 0.7260 0.6461 0.8038
No log 12.5 400 0.6468 0.7130 0.6468 0.8042
No log 12.5625 402 0.6482 0.6935 0.6482 0.8051
No log 12.625 404 0.6875 0.6948 0.6875 0.8291
No log 12.6875 406 0.7055 0.6797 0.7055 0.8400
No log 12.75 408 0.6770 0.6929 0.6770 0.8228
No log 12.8125 410 0.6535 0.6831 0.6535 0.8084
No log 12.875 412 0.6252 0.7227 0.6252 0.7907
No log 12.9375 414 0.6153 0.7259 0.6153 0.7844
No log 13.0 416 0.6131 0.7365 0.6131 0.7830
No log 13.0625 418 0.6228 0.7486 0.6228 0.7892
No log 13.125 420 0.6291 0.7350 0.6291 0.7932
No log 13.1875 422 0.6268 0.7153 0.6268 0.7917
No log 13.25 424 0.6377 0.7029 0.6377 0.7986
No log 13.3125 426 0.6675 0.6828 0.6675 0.8170
No log 13.375 428 0.6747 0.6885 0.6747 0.8214
No log 13.4375 430 0.6577 0.7088 0.6577 0.8110
No log 13.5 432 0.6290 0.6927 0.6290 0.7931
No log 13.5625 434 0.6390 0.6876 0.6390 0.7994
No log 13.625 436 0.6368 0.7093 0.6368 0.7980
No log 13.6875 438 0.6678 0.7213 0.6678 0.8172
No log 13.75 440 0.7893 0.6610 0.7893 0.8884
No log 13.8125 442 0.8389 0.6402 0.8389 0.9159
No log 13.875 444 0.7687 0.6386 0.7687 0.8767
No log 13.9375 446 0.6742 0.6821 0.6742 0.8211
No log 14.0 448 0.6398 0.7233 0.6398 0.7999
No log 14.0625 450 0.6234 0.7147 0.6234 0.7895
No log 14.125 452 0.6156 0.7102 0.6156 0.7846
No log 14.1875 454 0.6140 0.7109 0.6140 0.7836
No log 14.25 456 0.6441 0.7176 0.6441 0.8025
No log 14.3125 458 0.6635 0.6797 0.6635 0.8145
No log 14.375 460 0.6320 0.7127 0.6320 0.7950
No log 14.4375 462 0.6009 0.7317 0.6009 0.7752
No log 14.5 464 0.6122 0.6833 0.6122 0.7824
No log 14.5625 466 0.6341 0.6676 0.6341 0.7963
No log 14.625 468 0.6163 0.6862 0.6163 0.7850
No log 14.6875 470 0.6030 0.7235 0.6030 0.7766
No log 14.75 472 0.6066 0.7324 0.6066 0.7788
No log 14.8125 474 0.6154 0.7511 0.6154 0.7844
No log 14.875 476 0.6427 0.7270 0.6427 0.8017
No log 14.9375 478 0.6505 0.7552 0.6505 0.8065
No log 15.0 480 0.6495 0.7396 0.6495 0.8059
No log 15.0625 482 0.6281 0.7252 0.6281 0.7926
No log 15.125 484 0.6292 0.7140 0.6292 0.7932
No log 15.1875 486 0.6403 0.7221 0.6403 0.8002
No log 15.25 488 0.6990 0.7157 0.6990 0.8361
No log 15.3125 490 0.7847 0.6695 0.7847 0.8858
No log 15.375 492 0.7947 0.6489 0.7947 0.8915
No log 15.4375 494 0.7333 0.7294 0.7333 0.8564
No log 15.5 496 0.7046 0.7238 0.7046 0.8394
No log 15.5625 498 0.6843 0.7092 0.6843 0.8272
0.4225 15.625 500 0.6964 0.6947 0.6964 0.8345
0.4225 15.6875 502 0.7060 0.6952 0.7060 0.8402
0.4225 15.75 504 0.7509 0.6859 0.7509 0.8665
0.4225 15.8125 506 0.8372 0.6343 0.8372 0.9150
0.4225 15.875 508 0.8242 0.6443 0.8242 0.9079
0.4225 15.9375 510 0.8484 0.6433 0.8484 0.9211
0.4225 16.0 512 0.8110 0.6535 0.8110 0.9006

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k6_task1_organization

Finetuned
(4204)
this model