ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k3_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9478
  • Qwk: 0.5380
  • Mse: 0.9478
  • Rmse: 0.9735

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1333 2 5.1014 -0.0137 5.1014 2.2586
No log 0.2667 4 3.5702 0.0723 3.5702 1.8895
No log 0.4 6 2.0572 0.0957 2.0572 1.4343
No log 0.5333 8 1.8291 0.0248 1.8291 1.3525
No log 0.6667 10 1.3573 0.1208 1.3573 1.1650
No log 0.8 12 1.1986 0.2232 1.1986 1.0948
No log 0.9333 14 1.1793 0.1771 1.1793 1.0860
No log 1.0667 16 1.1839 0.0918 1.1839 1.0881
No log 1.2 18 1.1708 0.0967 1.1708 1.0820
No log 1.3333 20 1.1497 0.1584 1.1497 1.0723
No log 1.4667 22 1.1107 0.2587 1.1107 1.0539
No log 1.6 24 1.0294 0.3188 1.0294 1.0146
No log 1.7333 26 0.9963 0.2802 0.9963 0.9982
No log 1.8667 28 1.2567 0.2757 1.2567 1.1210
No log 2.0 30 1.1841 0.3143 1.1841 1.0882
No log 2.1333 32 0.9474 0.3166 0.9474 0.9734
No log 2.2667 34 0.9094 0.4223 0.9094 0.9536
No log 2.4 36 0.9260 0.4167 0.9260 0.9623
No log 2.5333 38 0.8822 0.4539 0.8822 0.9392
No log 2.6667 40 0.8368 0.4914 0.8368 0.9148
No log 2.8 42 0.8631 0.4427 0.8631 0.9290
No log 2.9333 44 0.9219 0.4409 0.9219 0.9601
No log 3.0667 46 0.8554 0.4369 0.8554 0.9249
No log 3.2 48 0.8410 0.5207 0.8410 0.9171
No log 3.3333 50 0.8403 0.4709 0.8403 0.9167
No log 3.4667 52 0.7445 0.5313 0.7445 0.8629
No log 3.6 54 0.7240 0.5935 0.7240 0.8509
No log 3.7333 56 0.8421 0.5654 0.8421 0.9177
No log 3.8667 58 0.9731 0.4889 0.9731 0.9865
No log 4.0 60 0.9843 0.4698 0.9843 0.9921
No log 4.1333 62 1.0109 0.5102 1.0109 1.0054
No log 4.2667 64 1.0027 0.5744 1.0027 1.0013
No log 4.4 66 0.7808 0.6795 0.7808 0.8836
No log 4.5333 68 0.6736 0.7128 0.6736 0.8208
No log 4.6667 70 0.7573 0.6730 0.7573 0.8702
No log 4.8 72 0.7737 0.6655 0.7737 0.8796
No log 4.9333 74 0.6173 0.7349 0.6173 0.7857
No log 5.0667 76 0.7886 0.6778 0.7886 0.8880
No log 5.2 78 0.8951 0.6128 0.8951 0.9461
No log 5.3333 80 0.8614 0.5940 0.8614 0.9281
No log 5.4667 82 0.7638 0.6018 0.7638 0.8739
No log 5.6 84 0.7731 0.6372 0.7731 0.8793
No log 5.7333 86 1.0199 0.5831 1.0199 1.0099
No log 5.8667 88 1.0002 0.5625 1.0002 1.0001
No log 6.0 90 0.7479 0.6644 0.7479 0.8648
No log 6.1333 92 0.8345 0.5345 0.8345 0.9135
No log 6.2667 94 0.7712 0.5443 0.7712 0.8782
No log 6.4 96 0.6644 0.6408 0.6644 0.8151
No log 6.5333 98 0.7768 0.6727 0.7768 0.8814
No log 6.6667 100 0.9828 0.5770 0.9828 0.9914
No log 6.8 102 1.0346 0.5732 1.0346 1.0171
No log 6.9333 104 1.1637 0.5872 1.1637 1.0787
No log 7.0667 106 1.0844 0.5918 1.0844 1.0414
No log 7.2 108 0.9006 0.6315 0.9006 0.9490
No log 7.3333 110 0.7327 0.6617 0.7327 0.8560
No log 7.4667 112 0.7304 0.6686 0.7304 0.8546
No log 7.6 114 0.8822 0.6532 0.8822 0.9393
No log 7.7333 116 0.9152 0.6523 0.9152 0.9567
No log 7.8667 118 0.8226 0.6395 0.8226 0.9070
No log 8.0 120 0.8147 0.6212 0.8147 0.9026
No log 8.1333 122 0.7126 0.6487 0.7126 0.8441
No log 8.2667 124 0.6444 0.6873 0.6444 0.8027
No log 8.4 126 0.6550 0.6602 0.6550 0.8093
No log 8.5333 128 0.6432 0.7129 0.6432 0.8020
No log 8.6667 130 0.7659 0.6336 0.7659 0.8752
No log 8.8 132 1.1970 0.5487 1.1970 1.0941
No log 8.9333 134 1.3963 0.4597 1.3963 1.1817
No log 9.0667 136 1.2894 0.4941 1.2894 1.1355
No log 9.2 138 0.9859 0.5601 0.9859 0.9929
No log 9.3333 140 0.7392 0.7012 0.7392 0.8598
No log 9.4667 142 0.6778 0.6875 0.6778 0.8233
No log 9.6 144 0.6967 0.6998 0.6967 0.8347
No log 9.7333 146 0.7626 0.6534 0.7626 0.8733
No log 9.8667 148 0.9039 0.6039 0.9039 0.9507
No log 10.0 150 0.9780 0.5771 0.9780 0.9889
No log 10.1333 152 0.9273 0.5679 0.9273 0.9630
No log 10.2667 154 0.8045 0.5688 0.8045 0.8970
No log 10.4 156 0.7713 0.6012 0.7713 0.8782
No log 10.5333 158 0.7664 0.6135 0.7664 0.8754
No log 10.6667 160 0.7937 0.6814 0.7937 0.8909
No log 10.8 162 0.8413 0.6619 0.8413 0.9172
No log 10.9333 164 0.7903 0.6657 0.7903 0.8890
No log 11.0667 166 0.7414 0.6458 0.7414 0.8611
No log 11.2 168 0.6768 0.7058 0.6768 0.8227
No log 11.3333 170 0.6628 0.6745 0.6628 0.8141
No log 11.4667 172 0.7042 0.6831 0.7042 0.8392
No log 11.6 174 0.8381 0.5851 0.8381 0.9155
No log 11.7333 176 0.9962 0.5189 0.9962 0.9981
No log 11.8667 178 0.9663 0.5149 0.9663 0.9830
No log 12.0 180 0.8926 0.5292 0.8926 0.9448
No log 12.1333 182 0.8258 0.5622 0.8258 0.9088
No log 12.2667 184 0.7873 0.5996 0.7873 0.8873
No log 12.4 186 0.7588 0.6279 0.7588 0.8711
No log 12.5333 188 0.8636 0.6303 0.8636 0.9293
No log 12.6667 190 1.0423 0.5494 1.0423 1.0209
No log 12.8 192 1.1374 0.5394 1.1374 1.0665
No log 12.9333 194 0.9805 0.5580 0.9805 0.9902
No log 13.0667 196 0.7638 0.6786 0.7638 0.8740
No log 13.2 198 0.6801 0.6902 0.6801 0.8247
No log 13.3333 200 0.7181 0.6814 0.7181 0.8474
No log 13.4667 202 0.9209 0.6037 0.9209 0.9596
No log 13.6 204 1.0366 0.5866 1.0366 1.0181
No log 13.7333 206 0.8782 0.6409 0.8782 0.9371
No log 13.8667 208 0.7125 0.6623 0.7125 0.8441
No log 14.0 210 0.7198 0.6623 0.7198 0.8484
No log 14.1333 212 0.8062 0.6473 0.8062 0.8979
No log 14.2667 214 0.9816 0.6046 0.9816 0.9908
No log 14.4 216 1.0717 0.6041 1.0717 1.0352
No log 14.5333 218 1.1566 0.5905 1.1566 1.0755
No log 14.6667 220 1.0433 0.6041 1.0433 1.0214
No log 14.8 222 0.7804 0.6451 0.7804 0.8834
No log 14.9333 224 0.7023 0.6621 0.7023 0.8380
No log 15.0667 226 0.7212 0.6664 0.7212 0.8492
No log 15.2 228 0.8576 0.5890 0.8576 0.9260
No log 15.3333 230 1.0608 0.5991 1.0608 1.0300
No log 15.4667 232 1.1233 0.5558 1.1233 1.0599
No log 15.6 234 0.9683 0.5968 0.9683 0.9840
No log 15.7333 236 0.8918 0.6011 0.8918 0.9444
No log 15.8667 238 0.8034 0.6194 0.8034 0.8963
No log 16.0 240 0.8689 0.5968 0.8689 0.9321
No log 16.1333 242 0.9349 0.5743 0.9349 0.9669
No log 16.2667 244 0.9210 0.5601 0.9210 0.9597
No log 16.4 246 0.8756 0.5809 0.8756 0.9358
No log 16.5333 248 0.8627 0.5875 0.8627 0.9288
No log 16.6667 250 0.8918 0.5782 0.8918 0.9443
No log 16.8 252 1.0114 0.5543 1.0114 1.0057
No log 16.9333 254 1.1582 0.5547 1.1582 1.0762
No log 17.0667 256 1.0835 0.5666 1.0835 1.0409
No log 17.2 258 0.9699 0.5944 0.9699 0.9848
No log 17.3333 260 0.9218 0.6018 0.9218 0.9601
No log 17.4667 262 0.8693 0.6301 0.8693 0.9324
No log 17.6 264 0.7771 0.6436 0.7771 0.8816
No log 17.7333 266 0.8108 0.6318 0.8108 0.9004
No log 17.8667 268 0.9391 0.5829 0.9391 0.9691
No log 18.0 270 0.9957 0.5574 0.9957 0.9978
No log 18.1333 272 0.8805 0.6015 0.8805 0.9384
No log 18.2667 274 0.7243 0.6511 0.7243 0.8511
No log 18.4 276 0.6881 0.6573 0.6881 0.8295
No log 18.5333 278 0.7431 0.6389 0.7431 0.8621
No log 18.6667 280 0.9332 0.5945 0.9332 0.9660
No log 18.8 282 1.1373 0.4983 1.1373 1.0664
No log 18.9333 284 1.1789 0.4603 1.1789 1.0858
No log 19.0667 286 1.1012 0.4973 1.1012 1.0494
No log 19.2 288 1.0141 0.5506 1.0141 1.0070
No log 19.3333 290 0.9121 0.5756 0.9121 0.9550
No log 19.4667 292 0.8448 0.6032 0.8448 0.9191
No log 19.6 294 0.8703 0.5947 0.8703 0.9329
No log 19.7333 296 1.0412 0.5430 1.0412 1.0204
No log 19.8667 298 1.1441 0.5301 1.1441 1.0696
No log 20.0 300 1.0337 0.5669 1.0337 1.0167
No log 20.1333 302 0.8553 0.5557 0.8553 0.9248
No log 20.2667 304 0.7892 0.6262 0.7892 0.8884
No log 20.4 306 0.8321 0.6014 0.8321 0.9122
No log 20.5333 308 0.9341 0.5452 0.9341 0.9665
No log 20.6667 310 0.9477 0.5481 0.9477 0.9735
No log 20.8 312 0.9159 0.5884 0.9159 0.9570
No log 20.9333 314 0.8631 0.6018 0.8631 0.9290
No log 21.0667 316 0.7864 0.6234 0.7864 0.8868
No log 21.2 318 0.8403 0.6140 0.8403 0.9167
No log 21.3333 320 0.9765 0.5463 0.9765 0.9882
No log 21.4667 322 1.1500 0.5528 1.1500 1.0724
No log 21.6 324 1.1956 0.5256 1.1956 1.0934
No log 21.7333 326 1.0958 0.5271 1.0958 1.0468
No log 21.8667 328 0.9253 0.5514 0.9253 0.9620
No log 22.0 330 0.8549 0.6309 0.8549 0.9246
No log 22.1333 332 0.9136 0.6120 0.9136 0.9558
No log 22.2667 334 1.0979 0.5427 1.0979 1.0478
No log 22.4 336 1.2203 0.5236 1.2203 1.1047
No log 22.5333 338 1.1317 0.5497 1.1317 1.0638
No log 22.6667 340 1.0750 0.5319 1.0750 1.0368
No log 22.8 342 1.0288 0.5542 1.0288 1.0143
No log 22.9333 344 0.8819 0.5953 0.8819 0.9391
No log 23.0667 346 0.7201 0.6553 0.7201 0.8486
No log 23.2 348 0.6753 0.6726 0.6753 0.8218
No log 23.3333 350 0.7261 0.6540 0.7261 0.8521
No log 23.4667 352 0.9067 0.6120 0.9067 0.9522
No log 23.6 354 1.1019 0.5722 1.1019 1.0497
No log 23.7333 356 1.2135 0.5352 1.2135 1.1016
No log 23.8667 358 1.1886 0.4879 1.1886 1.0902
No log 24.0 360 1.0463 0.4993 1.0463 1.0229
No log 24.1333 362 0.9107 0.5610 0.9107 0.9543
No log 24.2667 364 0.8479 0.6004 0.8479 0.9208
No log 24.4 366 0.9120 0.5692 0.9120 0.9550
No log 24.5333 368 1.0035 0.5119 1.0035 1.0017
No log 24.6667 370 0.9427 0.5631 0.9427 0.9709
No log 24.8 372 0.8954 0.6009 0.8954 0.9463
No log 24.9333 374 0.8419 0.6192 0.8419 0.9176
No log 25.0667 376 0.8517 0.6094 0.8517 0.9229
No log 25.2 378 0.9031 0.6087 0.9031 0.9503
No log 25.3333 380 0.8997 0.6087 0.8997 0.9485
No log 25.4667 382 0.8170 0.6132 0.8170 0.9039
No log 25.6 384 0.7277 0.6440 0.7277 0.8531
No log 25.7333 386 0.7084 0.6261 0.7084 0.8417
No log 25.8667 388 0.7199 0.6279 0.7199 0.8485
No log 26.0 390 0.7802 0.6036 0.7802 0.8833
No log 26.1333 392 0.8784 0.5978 0.8784 0.9372
No log 26.2667 394 0.9394 0.5981 0.9394 0.9692
No log 26.4 396 0.9289 0.5769 0.9289 0.9638
No log 26.5333 398 0.8831 0.6039 0.8831 0.9397
No log 26.6667 400 0.9006 0.5730 0.9006 0.9490
No log 26.8 402 1.0012 0.5260 1.0012 1.0006
No log 26.9333 404 1.1272 0.4907 1.1272 1.0617
No log 27.0667 406 1.2982 0.4899 1.2982 1.1394
No log 27.2 408 1.4419 0.4755 1.4419 1.2008
No log 27.3333 410 1.4425 0.4597 1.4425 1.2011
No log 27.4667 412 1.2631 0.5070 1.2631 1.1239
No log 27.6 414 0.9748 0.5744 0.9748 0.9873
No log 27.7333 416 0.7481 0.6707 0.7481 0.8649
No log 27.8667 418 0.6948 0.6933 0.6948 0.8336
No log 28.0 420 0.7266 0.6985 0.7266 0.8524
No log 28.1333 422 0.8370 0.6475 0.8370 0.9149
No log 28.2667 424 1.0672 0.5895 1.0672 1.0330
No log 28.4 426 1.1562 0.5397 1.1562 1.0753
No log 28.5333 428 1.0600 0.5655 1.0600 1.0296
No log 28.6667 430 0.8828 0.5821 0.8828 0.9396
No log 28.8 432 0.7547 0.6214 0.7547 0.8688
No log 28.9333 434 0.6918 0.6576 0.6918 0.8317
No log 29.0667 436 0.6931 0.6486 0.6931 0.8325
No log 29.2 438 0.7465 0.6156 0.7465 0.8640
No log 29.3333 440 0.8818 0.5745 0.8818 0.9390
No log 29.4667 442 1.0698 0.4678 1.0698 1.0343
No log 29.6 444 1.1079 0.4809 1.1079 1.0526
No log 29.7333 446 1.0420 0.4987 1.0420 1.0208
No log 29.8667 448 0.9064 0.5957 0.9064 0.9521
No log 30.0 450 0.7407 0.6389 0.7407 0.8606
No log 30.1333 452 0.6719 0.6868 0.6719 0.8197
No log 30.2667 454 0.6622 0.6959 0.6622 0.8138
No log 30.4 456 0.7032 0.6958 0.7032 0.8386
No log 30.5333 458 0.8392 0.6128 0.8392 0.9161
No log 30.6667 460 0.9981 0.5563 0.9981 0.9990
No log 30.8 462 1.0127 0.5393 1.0127 1.0063
No log 30.9333 464 0.9240 0.5745 0.9240 0.9613
No log 31.0667 466 0.7881 0.6349 0.7881 0.8877
No log 31.2 468 0.7057 0.6698 0.7057 0.8401
No log 31.3333 470 0.7089 0.6611 0.7089 0.8419
No log 31.4667 472 0.7563 0.6176 0.7563 0.8697
No log 31.6 474 0.8547 0.5954 0.8547 0.9245
No log 31.7333 476 0.8899 0.5731 0.8899 0.9434
No log 31.8667 478 0.9023 0.5421 0.9023 0.9499
No log 32.0 480 0.8353 0.6005 0.8353 0.9139
No log 32.1333 482 0.7776 0.6281 0.7776 0.8818
No log 32.2667 484 0.7890 0.6202 0.7890 0.8882
No log 32.4 486 0.8900 0.5771 0.8900 0.9434
No log 32.5333 488 1.0895 0.5110 1.0895 1.0438
No log 32.6667 490 1.1851 0.5105 1.1851 1.0886
No log 32.8 492 1.1151 0.5036 1.1151 1.0560
No log 32.9333 494 0.9996 0.5440 0.9996 0.9998
No log 33.0667 496 0.9125 0.5745 0.9125 0.9552
No log 33.2 498 0.8651 0.6446 0.8651 0.9301
0.3675 33.3333 500 0.9014 0.6024 0.9014 0.9494
0.3675 33.4667 502 1.0042 0.5306 1.0042 1.0021
0.3675 33.6 504 1.0358 0.5485 1.0358 1.0178
0.3675 33.7333 506 0.9684 0.5668 0.9684 0.9841
0.3675 33.8667 508 0.8883 0.5969 0.8883 0.9425
0.3675 34.0 510 0.8778 0.6114 0.8778 0.9369
0.3675 34.1333 512 0.8545 0.6114 0.8545 0.9244
0.3675 34.2667 514 0.8288 0.6139 0.8288 0.9104
0.3675 34.4 516 0.8503 0.6241 0.8503 0.9221
0.3675 34.5333 518 0.8560 0.6114 0.8560 0.9252
0.3675 34.6667 520 0.8866 0.5977 0.8866 0.9416
0.3675 34.8 522 0.9155 0.5531 0.9155 0.9568
0.3675 34.9333 524 0.8503 0.6247 0.8503 0.9221
0.3675 35.0667 526 0.8026 0.6517 0.8026 0.8959
0.3675 35.2 528 0.7951 0.6588 0.7951 0.8917
0.3675 35.3333 530 0.8057 0.6474 0.8057 0.8976
0.3675 35.4667 532 0.8140 0.6039 0.8140 0.9022
0.3675 35.6 534 0.7897 0.6176 0.7897 0.8886
0.3675 35.7333 536 0.8166 0.5902 0.8166 0.9036
0.3675 35.8667 538 0.8815 0.5773 0.8815 0.9389
0.3675 36.0 540 0.8949 0.5854 0.8949 0.9460
0.3675 36.1333 542 0.8262 0.6037 0.8262 0.9089
0.3675 36.2667 544 0.7551 0.6194 0.7551 0.8690
0.3675 36.4 546 0.7761 0.6176 0.7761 0.8810
0.3675 36.5333 548 0.8670 0.5806 0.8670 0.9311
0.3675 36.6667 550 0.9897 0.5385 0.9897 0.9948
0.3675 36.8 552 1.0611 0.5529 1.0611 1.0301
0.3675 36.9333 554 1.0555 0.5554 1.0555 1.0274
0.3675 37.0667 556 0.9478 0.5380 0.9478 0.9735

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
6
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k3_task1_organization

Finetuned
(4222)
this model