ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k2_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1908
  • Qwk: 0.4951
  • Mse: 1.1908
  • Rmse: 1.0912

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1667 2 5.3284 0.0021 5.3284 2.3083
No log 0.3333 4 3.6943 0.0513 3.6943 1.9221
No log 0.5 6 1.9996 0.0983 1.9996 1.4141
No log 0.6667 8 1.4828 0.0896 1.4828 1.2177
No log 0.8333 10 1.2332 0.2944 1.2332 1.1105
No log 1.0 12 1.1344 0.2438 1.1344 1.0651
No log 1.1667 14 1.2409 0.2333 1.2409 1.1139
No log 1.3333 16 1.3370 0.1778 1.3370 1.1563
No log 1.5 18 1.0870 0.3501 1.0870 1.0426
No log 1.6667 20 1.1082 0.2115 1.1082 1.0527
No log 1.8333 22 1.0991 0.2166 1.0991 1.0484
No log 2.0 24 1.0328 0.2900 1.0328 1.0163
No log 2.1667 26 1.0598 0.3813 1.0598 1.0295
No log 2.3333 28 1.4317 0.2389 1.4317 1.1965
No log 2.5 30 1.5612 0.2002 1.5612 1.2495
No log 2.6667 32 1.2388 0.2660 1.2388 1.1130
No log 2.8333 34 0.9678 0.3861 0.9678 0.9838
No log 3.0 36 1.1487 0.3607 1.1487 1.0718
No log 3.1667 38 1.2653 0.4233 1.2653 1.1248
No log 3.3333 40 1.0990 0.4582 1.0990 1.0483
No log 3.5 42 0.9949 0.5102 0.9949 0.9974
No log 3.6667 44 1.2129 0.4193 1.2129 1.1013
No log 3.8333 46 1.1282 0.4617 1.1282 1.0622
No log 4.0 48 0.9541 0.5095 0.9541 0.9768
No log 4.1667 50 1.1361 0.4121 1.1361 1.0659
No log 4.3333 52 1.2856 0.3452 1.2856 1.1338
No log 4.5 54 1.1697 0.4269 1.1697 1.0815
No log 4.6667 56 0.9671 0.4530 0.9671 0.9834
No log 4.8333 58 0.9964 0.4358 0.9964 0.9982
No log 5.0 60 0.9765 0.4854 0.9765 0.9882
No log 5.1667 62 0.9567 0.4657 0.9567 0.9781
No log 5.3333 64 1.2777 0.4586 1.2777 1.1303
No log 5.5 66 1.4408 0.3753 1.4408 1.2004
No log 5.6667 68 1.3942 0.4764 1.3942 1.1807
No log 5.8333 70 1.1558 0.5203 1.1558 1.0751
No log 6.0 72 1.0295 0.5658 1.0295 1.0146
No log 6.1667 74 1.0331 0.5958 1.0331 1.0164
No log 6.3333 76 1.0887 0.5752 1.0887 1.0434
No log 6.5 78 0.9745 0.5802 0.9745 0.9872
No log 6.6667 80 0.9564 0.5387 0.9564 0.9780
No log 6.8333 82 1.0058 0.4884 1.0058 1.0029
No log 7.0 84 0.9467 0.5221 0.9467 0.9730
No log 7.1667 86 0.9508 0.5185 0.9508 0.9751
No log 7.3333 88 0.9234 0.5401 0.9234 0.9609
No log 7.5 90 0.9633 0.4998 0.9633 0.9815
No log 7.6667 92 1.0346 0.5039 1.0346 1.0171
No log 7.8333 94 0.9789 0.5642 0.9789 0.9894
No log 8.0 96 0.8815 0.5531 0.8815 0.9389
No log 8.1667 98 0.8766 0.5531 0.8766 0.9363
No log 8.3333 100 0.8060 0.5887 0.8060 0.8978
No log 8.5 102 0.8340 0.5778 0.8340 0.9133
No log 8.6667 104 0.9978 0.5813 0.9978 0.9989
No log 8.8333 106 0.9602 0.6128 0.9602 0.9799
No log 9.0 108 1.0380 0.5717 1.0380 1.0188
No log 9.1667 110 1.0433 0.5810 1.0433 1.0214
No log 9.3333 112 1.0957 0.5489 1.0957 1.0468
No log 9.5 114 1.1710 0.4954 1.1710 1.0821
No log 9.6667 116 1.3349 0.4608 1.3349 1.1554
No log 9.8333 118 1.3581 0.4587 1.3581 1.1654
No log 10.0 120 1.0969 0.5209 1.0969 1.0474
No log 10.1667 122 0.9703 0.5433 0.9703 0.9851
No log 10.3333 124 0.9623 0.5489 0.9623 0.9810
No log 10.5 126 1.0944 0.5646 1.0944 1.0461
No log 10.6667 128 1.0680 0.5645 1.0680 1.0334
No log 10.8333 130 0.9567 0.5900 0.9567 0.9781
No log 11.0 132 0.8453 0.5775 0.8453 0.9194
No log 11.1667 134 0.8790 0.6096 0.8790 0.9375
No log 11.3333 136 1.1018 0.5762 1.1018 1.0496
No log 11.5 138 1.1498 0.5566 1.1498 1.0723
No log 11.6667 140 0.9877 0.5774 0.9877 0.9938
No log 11.8333 142 0.8255 0.6451 0.8255 0.9086
No log 12.0 144 0.7383 0.6226 0.7383 0.8592
No log 12.1667 146 0.7408 0.6000 0.7408 0.8607
No log 12.3333 148 0.8513 0.6098 0.8513 0.9227
No log 12.5 150 0.9897 0.5489 0.9897 0.9948
No log 12.6667 152 1.0994 0.5155 1.0994 1.0485
No log 12.8333 154 0.9720 0.5655 0.9720 0.9859
No log 13.0 156 0.8350 0.6508 0.8350 0.9138
No log 13.1667 158 0.8573 0.6377 0.8573 0.9259
No log 13.3333 160 0.8109 0.6469 0.8109 0.9005
No log 13.5 162 0.7856 0.6467 0.7856 0.8864
No log 13.6667 164 0.9105 0.6261 0.9105 0.9542
No log 13.8333 166 1.0291 0.5741 1.0291 1.0144
No log 14.0 168 1.1094 0.5358 1.1094 1.0533
No log 14.1667 170 1.1152 0.5400 1.1152 1.0560
No log 14.3333 172 0.9484 0.6050 0.9484 0.9739
No log 14.5 174 0.7578 0.6386 0.7578 0.8705
No log 14.6667 176 0.7451 0.6467 0.7451 0.8632
No log 14.8333 178 0.7653 0.6231 0.7653 0.8748
No log 15.0 180 0.7703 0.6463 0.7703 0.8777
No log 15.1667 182 0.8314 0.6033 0.8314 0.9118
No log 15.3333 184 0.9628 0.6016 0.9628 0.9812
No log 15.5 186 1.2032 0.5243 1.2032 1.0969
No log 15.6667 188 1.4047 0.4422 1.4047 1.1852
No log 15.8333 190 1.2780 0.4831 1.2780 1.1305
No log 16.0 192 1.1750 0.5157 1.1750 1.0840
No log 16.1667 194 1.0362 0.5312 1.0362 1.0180
No log 16.3333 196 0.9845 0.5634 0.9845 0.9922
No log 16.5 198 1.1342 0.5469 1.1342 1.0650
No log 16.6667 200 1.4355 0.4853 1.4355 1.1981
No log 16.8333 202 1.6859 0.4073 1.6859 1.2984
No log 17.0 204 1.5542 0.4286 1.5542 1.2467
No log 17.1667 206 1.2900 0.4743 1.2900 1.1358
No log 17.3333 208 0.9588 0.5315 0.9588 0.9792
No log 17.5 210 0.8032 0.6110 0.8032 0.8962
No log 17.6667 212 0.7897 0.6138 0.7897 0.8887
No log 17.8333 214 0.8578 0.5858 0.8578 0.9262
No log 18.0 216 0.9631 0.5597 0.9631 0.9814
No log 18.1667 218 0.9426 0.5567 0.9426 0.9709
No log 18.3333 220 0.8108 0.5851 0.8108 0.9005
No log 18.5 222 0.7972 0.5651 0.7972 0.8928
No log 18.6667 224 0.8812 0.5686 0.8812 0.9387
No log 18.8333 226 1.0589 0.5495 1.0589 1.0290
No log 19.0 228 1.2877 0.4971 1.2877 1.1348
No log 19.1667 230 1.2074 0.4896 1.2074 1.0988
No log 19.3333 232 1.0061 0.5566 1.0061 1.0030
No log 19.5 234 0.8604 0.5448 0.8604 0.9276
No log 19.6667 236 0.8068 0.5477 0.8068 0.8982
No log 19.8333 238 0.8362 0.5878 0.8362 0.9145
No log 20.0 240 0.9071 0.6507 0.9071 0.9524
No log 20.1667 242 1.0063 0.5456 1.0063 1.0032
No log 20.3333 244 1.1496 0.5174 1.1496 1.0722
No log 20.5 246 1.2034 0.5002 1.2034 1.0970
No log 20.6667 248 1.1027 0.5041 1.1027 1.0501
No log 20.8333 250 0.8446 0.6175 0.8446 0.9190
No log 21.0 252 0.7721 0.5722 0.7721 0.8787
No log 21.1667 254 0.8099 0.5792 0.8099 0.9000
No log 21.3333 256 1.0018 0.5792 1.0018 1.0009
No log 21.5 258 1.2179 0.4683 1.2179 1.1036
No log 21.6667 260 1.2729 0.4868 1.2729 1.1282
No log 21.8333 262 1.1465 0.5157 1.1465 1.0708
No log 22.0 264 1.0551 0.5441 1.0551 1.0272
No log 22.1667 266 0.9533 0.6065 0.9533 0.9763
No log 22.3333 268 0.8957 0.6293 0.8957 0.9464
No log 22.5 270 0.9165 0.6155 0.9165 0.9573
No log 22.6667 272 0.9886 0.5965 0.9886 0.9943
No log 22.8333 274 0.9735 0.5992 0.9735 0.9866
No log 23.0 276 0.8723 0.6025 0.8723 0.9340
No log 23.1667 278 0.7905 0.6004 0.7905 0.8891
No log 23.3333 280 0.7188 0.6165 0.7188 0.8478
No log 23.5 282 0.7158 0.6011 0.7158 0.8460
No log 23.6667 284 0.7597 0.6330 0.7597 0.8716
No log 23.8333 286 0.8969 0.6142 0.8969 0.9471
No log 24.0 288 1.0341 0.5944 1.0341 1.0169
No log 24.1667 290 0.9823 0.5855 0.9823 0.9911
No log 24.3333 292 0.8807 0.5956 0.8807 0.9384
No log 24.5 294 0.8703 0.5553 0.8703 0.9329
No log 24.6667 296 0.9197 0.5743 0.9197 0.9590
No log 24.8333 298 0.9465 0.5940 0.9465 0.9729
No log 25.0 300 0.9288 0.6026 0.9288 0.9637
No log 25.1667 302 0.9780 0.6009 0.9780 0.9890
No log 25.3333 304 0.9036 0.6156 0.9036 0.9506
No log 25.5 306 0.7796 0.6731 0.7796 0.8830
No log 25.6667 308 0.7544 0.6721 0.7544 0.8686
No log 25.8333 310 0.8109 0.6495 0.8109 0.9005
No log 26.0 312 0.8323 0.6477 0.8323 0.9123
No log 26.1667 314 0.9667 0.5762 0.9667 0.9832
No log 26.3333 316 1.0880 0.5227 1.0880 1.0431
No log 26.5 318 1.1064 0.5036 1.1064 1.0518
No log 26.6667 320 1.0028 0.5570 1.0028 1.0014
No log 26.8333 322 1.0375 0.5364 1.0375 1.0186
No log 27.0 324 1.1381 0.4868 1.1381 1.0668
No log 27.1667 326 1.1166 0.5043 1.1166 1.0567
No log 27.3333 328 1.0151 0.5705 1.0151 1.0075
No log 27.5 330 0.9857 0.5718 0.9857 0.9928
No log 27.6667 332 1.1457 0.4943 1.1457 1.0704
No log 27.8333 334 1.2990 0.5024 1.2990 1.1397
No log 28.0 336 1.3859 0.4556 1.3859 1.1773
No log 28.1667 338 1.3107 0.4645 1.3107 1.1449
No log 28.3333 340 1.0769 0.5479 1.0769 1.0377
No log 28.5 342 0.8578 0.5673 0.8578 0.9262
No log 28.6667 344 0.7843 0.6064 0.7843 0.8856
No log 28.8333 346 0.7807 0.6064 0.7807 0.8836
No log 29.0 348 0.8626 0.6261 0.8626 0.9288
No log 29.1667 350 1.0270 0.5572 1.0270 1.0134
No log 29.3333 352 1.1850 0.5016 1.1850 1.0886
No log 29.5 354 1.4000 0.4500 1.4000 1.1832
No log 29.6667 356 1.3459 0.4736 1.3459 1.1601
No log 29.8333 358 1.1131 0.5114 1.1131 1.0550
No log 30.0 360 0.8908 0.6173 0.8908 0.9438
No log 30.1667 362 0.8267 0.6271 0.8267 0.9092
No log 30.3333 364 0.8628 0.6245 0.8628 0.9289
No log 30.5 366 0.9623 0.6100 0.9623 0.9810
No log 30.6667 368 0.9424 0.6100 0.9424 0.9708
No log 30.8333 370 0.8517 0.6154 0.8517 0.9229
No log 31.0 372 0.8255 0.6321 0.8255 0.9086
No log 31.1667 374 0.8710 0.5928 0.8710 0.9333
No log 31.3333 376 0.9316 0.5938 0.9316 0.9652
No log 31.5 378 0.9227 0.5878 0.9227 0.9606
No log 31.6667 380 0.9145 0.5919 0.9145 0.9563
No log 31.8333 382 0.8406 0.5878 0.8406 0.9168
No log 32.0 384 0.8386 0.6256 0.8386 0.9158
No log 32.1667 386 0.8850 0.5952 0.8850 0.9407
No log 32.3333 388 0.8950 0.6045 0.8950 0.9460
No log 32.5 390 0.9765 0.5872 0.9765 0.9882
No log 32.6667 392 0.9752 0.5860 0.9752 0.9875
No log 32.8333 394 0.8886 0.5939 0.8886 0.9427
No log 33.0 396 0.7714 0.5971 0.7714 0.8783
No log 33.1667 398 0.7442 0.6037 0.7442 0.8627
No log 33.3333 400 0.7773 0.6009 0.7773 0.8816
No log 33.5 402 0.8777 0.5890 0.8777 0.9369
No log 33.6667 404 1.0397 0.5686 1.0397 1.0197
No log 33.8333 406 1.1583 0.5361 1.1583 1.0763
No log 34.0 408 1.1712 0.5276 1.1712 1.0822
No log 34.1667 410 1.0899 0.5378 1.0899 1.0440
No log 34.3333 412 0.9092 0.6180 0.9092 0.9535
No log 34.5 414 0.8205 0.6048 0.8205 0.9058
No log 34.6667 416 0.8106 0.6176 0.8106 0.9003
No log 34.8333 418 0.8708 0.5769 0.8708 0.9331
No log 35.0 420 0.9949 0.5829 0.9949 0.9975
No log 35.1667 422 1.0961 0.5309 1.0961 1.0469
No log 35.3333 424 1.1708 0.5163 1.1708 1.0820
No log 35.5 426 1.1060 0.5267 1.1060 1.0517
No log 35.6667 428 1.0217 0.5752 1.0217 1.0108
No log 35.8333 430 0.9874 0.5892 0.9874 0.9937
No log 36.0 432 0.9540 0.6086 0.9540 0.9767
No log 36.1667 434 0.8689 0.6119 0.8689 0.9321
No log 36.3333 436 0.8541 0.6211 0.8541 0.9242
No log 36.5 438 0.9280 0.6274 0.9280 0.9633
No log 36.6667 440 1.0559 0.5802 1.0559 1.0276
No log 36.8333 442 1.1794 0.5010 1.1794 1.0860
No log 37.0 444 1.2043 0.5100 1.2043 1.0974
No log 37.1667 446 1.0534 0.5754 1.0534 1.0263
No log 37.3333 448 0.9352 0.6293 0.9352 0.9670
No log 37.5 450 0.8308 0.6406 0.8308 0.9115
No log 37.6667 452 0.7867 0.6326 0.7867 0.8870
No log 37.8333 454 0.8170 0.6148 0.8170 0.9039
No log 38.0 456 0.9314 0.6278 0.9314 0.9651
No log 38.1667 458 1.0612 0.5495 1.0612 1.0301
No log 38.3333 460 1.0770 0.5663 1.0770 1.0378
No log 38.5 462 1.0041 0.5708 1.0041 1.0021
No log 38.6667 464 0.9335 0.6246 0.9335 0.9662
No log 38.8333 466 0.8254 0.6441 0.8254 0.9085
No log 39.0 468 0.7692 0.5849 0.7692 0.8771
No log 39.1667 470 0.7596 0.5911 0.7596 0.8716
No log 39.3333 472 0.7908 0.5967 0.7908 0.8893
No log 39.5 474 0.8695 0.6204 0.8695 0.9325
No log 39.6667 476 1.0479 0.5707 1.0479 1.0236
No log 39.8333 478 1.2878 0.4732 1.2878 1.1348
No log 40.0 480 1.4204 0.4415 1.4204 1.1918
No log 40.1667 482 1.4722 0.4293 1.4722 1.2133
No log 40.3333 484 1.3330 0.4818 1.3330 1.1545
No log 40.5 486 1.1272 0.5495 1.1272 1.0617
No log 40.6667 488 0.9395 0.6347 0.9395 0.9693
No log 40.8333 490 0.8147 0.6271 0.8147 0.9026
No log 41.0 492 0.7834 0.6405 0.7834 0.8851
No log 41.1667 494 0.8070 0.6164 0.8070 0.8983
No log 41.3333 496 0.8648 0.6256 0.8648 0.9300
No log 41.5 498 0.9086 0.5968 0.9086 0.9532
0.3298 41.6667 500 0.9752 0.5783 0.9752 0.9875
0.3298 41.8333 502 1.0090 0.5805 1.0090 1.0045
0.3298 42.0 504 1.0389 0.5916 1.0389 1.0193
0.3298 42.1667 506 1.0323 0.5573 1.0323 1.0160
0.3298 42.3333 508 0.9446 0.6051 0.9446 0.9719
0.3298 42.5 510 0.8819 0.6281 0.8819 0.9391
0.3298 42.6667 512 0.8680 0.6397 0.8680 0.9317
0.3298 42.8333 514 0.8912 0.6139 0.8912 0.9440
0.3298 43.0 516 0.9985 0.5553 0.9985 0.9992
0.3298 43.1667 518 1.0470 0.5573 1.0470 1.0232
0.3298 43.3333 520 1.1167 0.5329 1.1167 1.0567
0.3298 43.5 522 1.2175 0.4952 1.2175 1.1034
0.3298 43.6667 524 1.1908 0.4951 1.1908 1.0912

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
6
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k2_task1_organization

Finetuned
(4222)
this model