ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k2_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8113
  • Qwk: 0.5917
  • Mse: 0.8113
  • Rmse: 0.9007

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1667 2 5.4358 -0.0297 5.4358 2.3315
No log 0.3333 4 3.5345 0.0746 3.5345 1.8800
No log 0.5 6 2.1615 0.0475 2.1615 1.4702
No log 0.6667 8 2.6902 -0.1733 2.6902 1.6402
No log 0.8333 10 2.4000 -0.1173 2.4000 1.5492
No log 1.0 12 1.4761 0.0593 1.4761 1.2150
No log 1.1667 14 1.2640 0.2135 1.2640 1.1243
No log 1.3333 16 1.2667 0.1532 1.2667 1.1255
No log 1.5 18 1.3127 0.1374 1.3127 1.1457
No log 1.6667 20 1.3180 0.1944 1.3180 1.1480
No log 1.8333 22 1.2969 0.1908 1.2969 1.1388
No log 2.0 24 1.2834 0.0697 1.2834 1.1329
No log 2.1667 26 1.2662 0.0465 1.2662 1.1253
No log 2.3333 28 1.2301 0.0465 1.2301 1.1091
No log 2.5 30 1.2121 0.0465 1.2121 1.1010
No log 2.6667 32 1.1312 0.1510 1.1312 1.0636
No log 2.8333 34 1.1129 0.2149 1.1129 1.0549
No log 3.0 36 1.1962 0.2534 1.1962 1.0937
No log 3.1667 38 1.2411 0.2574 1.2411 1.1140
No log 3.3333 40 1.0605 0.3625 1.0605 1.0298
No log 3.5 42 0.9339 0.3152 0.9339 0.9664
No log 3.6667 44 1.0453 0.3908 1.0453 1.0224
No log 3.8333 46 0.9523 0.4582 0.9523 0.9759
No log 4.0 48 0.8697 0.4953 0.8697 0.9326
No log 4.1667 50 0.9218 0.4893 0.9218 0.9601
No log 4.3333 52 0.9982 0.4199 0.9982 0.9991
No log 4.5 54 0.7975 0.5952 0.7975 0.8930
No log 4.6667 56 0.9277 0.4860 0.9277 0.9632
No log 4.8333 58 1.1444 0.3983 1.1444 1.0698
No log 5.0 60 0.9519 0.4318 0.9519 0.9756
No log 5.1667 62 0.7670 0.6397 0.7670 0.8758
No log 5.3333 64 0.7390 0.6216 0.7390 0.8596
No log 5.5 66 0.8863 0.4820 0.8863 0.9414
No log 5.6667 68 0.7626 0.5668 0.7626 0.8733
No log 5.8333 70 0.6962 0.5909 0.6962 0.8344
No log 6.0 72 0.7098 0.6265 0.7098 0.8425
No log 6.1667 74 0.7143 0.6480 0.7143 0.8451
No log 6.3333 76 0.7739 0.6659 0.7739 0.8797
No log 6.5 78 0.7998 0.6589 0.7998 0.8943
No log 6.6667 80 0.7811 0.6567 0.7811 0.8838
No log 6.8333 82 0.8077 0.6232 0.8077 0.8987
No log 7.0 84 0.7737 0.6350 0.7737 0.8796
No log 7.1667 86 0.7401 0.6390 0.7401 0.8603
No log 7.3333 88 0.9133 0.6049 0.9133 0.9557
No log 7.5 90 1.2633 0.4641 1.2633 1.1240
No log 7.6667 92 1.1792 0.4868 1.1792 1.0859
No log 7.8333 94 0.8222 0.6085 0.8222 0.9067
No log 8.0 96 0.6225 0.6417 0.6225 0.7890
No log 8.1667 98 0.6168 0.6913 0.6168 0.7854
No log 8.3333 100 0.6399 0.6643 0.6399 0.8000
No log 8.5 102 0.9269 0.6197 0.9269 0.9628
No log 8.6667 104 1.2865 0.5281 1.2865 1.1342
No log 8.8333 106 1.0829 0.5766 1.0829 1.0406
No log 9.0 108 0.7073 0.6242 0.7073 0.8410
No log 9.1667 110 0.6491 0.6714 0.6491 0.8056
No log 9.3333 112 0.6406 0.6672 0.6406 0.8004
No log 9.5 114 0.7337 0.6166 0.7337 0.8565
No log 9.6667 116 1.0648 0.5550 1.0648 1.0319
No log 9.8333 118 1.0653 0.5348 1.0653 1.0321
No log 10.0 120 0.8412 0.6026 0.8412 0.9172
No log 10.1667 122 0.6323 0.7018 0.6323 0.7951
No log 10.3333 124 0.6464 0.6898 0.6464 0.8040
No log 10.5 126 0.6445 0.6848 0.6445 0.8028
No log 10.6667 128 0.7044 0.6529 0.7044 0.8393
No log 10.8333 130 0.8204 0.6027 0.8204 0.9057
No log 11.0 132 0.8548 0.5865 0.8548 0.9246
No log 11.1667 134 0.7625 0.6070 0.7625 0.8732
No log 11.3333 136 0.6841 0.6330 0.6841 0.8271
No log 11.5 138 0.7023 0.6347 0.7023 0.8381
No log 11.6667 140 0.8313 0.5822 0.8313 0.9118
No log 11.8333 142 1.0535 0.5212 1.0535 1.0264
No log 12.0 144 1.3821 0.4625 1.3821 1.1756
No log 12.1667 146 1.4354 0.4101 1.4354 1.1981
No log 12.3333 148 1.1759 0.4490 1.1759 1.0844
No log 12.5 150 0.8859 0.5341 0.8859 0.9412
No log 12.6667 152 0.7790 0.5626 0.7790 0.8826
No log 12.8333 154 0.7612 0.5973 0.7612 0.8725
No log 13.0 156 0.8651 0.5785 0.8651 0.9301
No log 13.1667 158 0.8617 0.5664 0.8617 0.9283
No log 13.3333 160 0.8320 0.5606 0.8320 0.9122
No log 13.5 162 0.7525 0.5816 0.7525 0.8674
No log 13.6667 164 0.6900 0.6226 0.6900 0.8307
No log 13.8333 166 0.6990 0.5711 0.6990 0.8361
No log 14.0 168 0.6764 0.6290 0.6764 0.8224
No log 14.1667 170 0.6704 0.6611 0.6704 0.8188
No log 14.3333 172 0.6667 0.6997 0.6667 0.8165
No log 14.5 174 0.7612 0.6648 0.7612 0.8725
No log 14.6667 176 0.7549 0.6752 0.7549 0.8688
No log 14.8333 178 0.7025 0.7210 0.7025 0.8381
No log 15.0 180 0.6301 0.7367 0.6301 0.7938
No log 15.1667 182 0.6391 0.6900 0.6391 0.7994
No log 15.3333 184 0.6610 0.6452 0.6610 0.8130
No log 15.5 186 0.7492 0.5639 0.7492 0.8655
No log 15.6667 188 0.8604 0.5427 0.8604 0.9276
No log 15.8333 190 0.8448 0.5442 0.8448 0.9191
No log 16.0 192 0.7778 0.5438 0.7778 0.8819
No log 16.1667 194 0.8290 0.5518 0.8290 0.9105
No log 16.3333 196 0.7873 0.5955 0.7873 0.8873
No log 16.5 198 0.7723 0.5971 0.7723 0.8788
No log 16.6667 200 0.8684 0.5622 0.8684 0.9319
No log 16.8333 202 1.0396 0.5520 1.0396 1.0196
No log 17.0 204 0.9377 0.5426 0.9377 0.9683
No log 17.1667 206 0.8025 0.5645 0.8025 0.8958
No log 17.3333 208 0.6542 0.6628 0.6542 0.8088
No log 17.5 210 0.6493 0.6381 0.6493 0.8058
No log 17.6667 212 0.6324 0.6963 0.6324 0.7952
No log 17.8333 214 0.6225 0.6768 0.6225 0.7890
No log 18.0 216 0.8171 0.6533 0.8171 0.9039
No log 18.1667 218 1.0291 0.5457 1.0291 1.0145
No log 18.3333 220 1.0137 0.5568 1.0137 1.0068
No log 18.5 222 0.8438 0.6062 0.8438 0.9186
No log 18.6667 224 0.6663 0.6846 0.6663 0.8163
No log 18.8333 226 0.6319 0.6682 0.6319 0.7949
No log 19.0 228 0.6348 0.6573 0.6348 0.7967
No log 19.1667 230 0.7295 0.6445 0.7295 0.8541
No log 19.3333 232 0.8660 0.5548 0.8660 0.9306
No log 19.5 234 0.9169 0.5435 0.9169 0.9575
No log 19.6667 236 0.9751 0.5435 0.9751 0.9875
No log 19.8333 238 0.8674 0.5435 0.8674 0.9313
No log 20.0 240 0.7471 0.5802 0.7471 0.8644
No log 20.1667 242 0.7188 0.6167 0.7188 0.8478
No log 20.3333 244 0.7928 0.5814 0.7928 0.8904
No log 20.5 246 0.9584 0.5470 0.9584 0.9790
No log 20.6667 248 0.9150 0.5601 0.9150 0.9565
No log 20.8333 250 0.7226 0.6376 0.7226 0.8501
No log 21.0 252 0.6289 0.6683 0.6289 0.7930
No log 21.1667 254 0.6473 0.6932 0.6473 0.8046
No log 21.3333 256 0.7521 0.6604 0.7521 0.8673
No log 21.5 258 0.8457 0.5879 0.8457 0.9196
No log 21.6667 260 0.8053 0.5962 0.8053 0.8974
No log 21.8333 262 0.7148 0.6352 0.7148 0.8455
No log 22.0 264 0.6928 0.6813 0.6928 0.8323
No log 22.1667 266 0.7757 0.6279 0.7757 0.8807
No log 22.3333 268 0.8422 0.5910 0.8422 0.9177
No log 22.5 270 0.8650 0.5821 0.8650 0.9300
No log 22.6667 272 0.8244 0.5853 0.8244 0.9080
No log 22.8333 274 0.8274 0.5758 0.8274 0.9096
No log 23.0 276 0.8393 0.5575 0.8393 0.9161
No log 23.1667 278 0.8158 0.5632 0.8158 0.9032
No log 23.3333 280 0.8229 0.5518 0.8229 0.9071
No log 23.5 282 0.8591 0.5605 0.8591 0.9269
No log 23.6667 284 0.8451 0.5661 0.8451 0.9193
No log 23.8333 286 0.7564 0.6083 0.7564 0.8697
No log 24.0 288 0.7104 0.6489 0.7104 0.8429
No log 24.1667 290 0.7228 0.6042 0.7228 0.8501
No log 24.3333 292 0.7775 0.5702 0.7775 0.8817
No log 24.5 294 0.8217 0.5856 0.8217 0.9065
No log 24.6667 296 0.7911 0.6118 0.7911 0.8895
No log 24.8333 298 0.7607 0.6373 0.7607 0.8722
No log 25.0 300 0.6736 0.6859 0.6736 0.8207
No log 25.1667 302 0.6354 0.7067 0.6354 0.7971
No log 25.3333 304 0.6595 0.6480 0.6595 0.8121
No log 25.5 306 0.6590 0.6691 0.6590 0.8118
No log 25.6667 308 0.6960 0.6448 0.6960 0.8343
No log 25.8333 310 0.7901 0.6346 0.7901 0.8889
No log 26.0 312 0.9454 0.5258 0.9454 0.9723
No log 26.1667 314 1.0207 0.4823 1.0207 1.0103
No log 26.3333 316 1.0253 0.4877 1.0253 1.0126
No log 26.5 318 0.9545 0.5357 0.9545 0.9770
No log 26.6667 320 0.8690 0.5840 0.8690 0.9322
No log 26.8333 322 0.8077 0.6062 0.8077 0.8987
No log 27.0 324 0.7957 0.6329 0.7957 0.8920
No log 27.1667 326 0.8694 0.5826 0.8694 0.9324
No log 27.3333 328 0.9884 0.5829 0.9884 0.9942
No log 27.5 330 1.0618 0.5482 1.0618 1.0305
No log 27.6667 332 1.0405 0.5089 1.0405 1.0200
No log 27.8333 334 0.9306 0.5706 0.9306 0.9647
No log 28.0 336 0.7739 0.6088 0.7739 0.8797
No log 28.1667 338 0.7049 0.6707 0.7049 0.8396
No log 28.3333 340 0.7045 0.6707 0.7045 0.8393
No log 28.5 342 0.7709 0.6176 0.7709 0.8780
No log 28.6667 344 0.8741 0.5884 0.8741 0.9349
No log 28.8333 346 0.8960 0.5851 0.8960 0.9466
No log 29.0 348 0.8455 0.6169 0.8455 0.9195
No log 29.1667 350 0.7515 0.6147 0.7515 0.8669
No log 29.3333 352 0.7071 0.6417 0.7071 0.8409
No log 29.5 354 0.7248 0.6430 0.7248 0.8513
No log 29.6667 356 0.8064 0.6124 0.8064 0.8980
No log 29.8333 358 0.9054 0.5849 0.9054 0.9515
No log 30.0 360 0.9521 0.5745 0.9521 0.9758
No log 30.1667 362 0.9109 0.6022 0.9109 0.9544
No log 30.3333 364 0.8722 0.5959 0.8722 0.9339
No log 30.5 366 0.7801 0.6263 0.7801 0.8832
No log 30.6667 368 0.6999 0.6633 0.6999 0.8366
No log 30.8333 370 0.6939 0.6789 0.6939 0.8330
No log 31.0 372 0.6673 0.6871 0.6673 0.8169
No log 31.1667 374 0.6907 0.6789 0.6907 0.8311
No log 31.3333 376 0.7701 0.6457 0.7701 0.8776
No log 31.5 378 0.8110 0.6144 0.8110 0.9006
No log 31.6667 380 0.8706 0.5941 0.8706 0.9331
No log 31.8333 382 0.9169 0.5451 0.9169 0.9575
No log 32.0 384 0.9688 0.5383 0.9688 0.9843
No log 32.1667 386 1.0293 0.5614 1.0293 1.0146
No log 32.3333 388 1.0389 0.5564 1.0389 1.0193
No log 32.5 390 1.0227 0.5388 1.0227 1.0113
No log 32.6667 392 0.9477 0.6056 0.9477 0.9735
No log 32.8333 394 0.8902 0.6264 0.8902 0.9435
No log 33.0 396 0.8964 0.6058 0.8964 0.9468
No log 33.1667 398 0.9414 0.5584 0.9414 0.9702
No log 33.3333 400 0.9490 0.5439 0.9490 0.9742
No log 33.5 402 0.9850 0.5173 0.9850 0.9924
No log 33.6667 404 1.0209 0.5032 1.0209 1.0104
No log 33.8333 406 0.9755 0.5159 0.9755 0.9877
No log 34.0 408 0.9287 0.5239 0.9287 0.9637
No log 34.1667 410 0.7877 0.6361 0.7877 0.8875
No log 34.3333 412 0.7149 0.6579 0.7149 0.8455
No log 34.5 414 0.7035 0.6633 0.7035 0.8388
No log 34.6667 416 0.7609 0.6446 0.7609 0.8723
No log 34.8333 418 0.7794 0.6758 0.7794 0.8828
No log 35.0 420 0.8135 0.6518 0.8135 0.9020
No log 35.1667 422 0.7712 0.6306 0.7712 0.8782
No log 35.3333 424 0.7007 0.6543 0.7007 0.8371
No log 35.5 426 0.7052 0.6390 0.7052 0.8398
No log 35.6667 428 0.7541 0.5933 0.7541 0.8684
No log 35.8333 430 0.8246 0.5789 0.8246 0.9081
No log 36.0 432 0.8216 0.5746 0.8216 0.9064
No log 36.1667 434 0.7991 0.5802 0.7991 0.8939
No log 36.3333 436 0.7379 0.6392 0.7379 0.8590
No log 36.5 438 0.7124 0.6709 0.7124 0.8440
No log 36.6667 440 0.7115 0.6590 0.7115 0.8435
No log 36.8333 442 0.7318 0.6655 0.7318 0.8555
No log 37.0 444 0.8158 0.6423 0.8158 0.9032
No log 37.1667 446 0.8767 0.6009 0.8767 0.9363
No log 37.3333 448 0.9009 0.5651 0.9009 0.9492
No log 37.5 450 0.8529 0.5635 0.8529 0.9235
No log 37.6667 452 0.8567 0.5494 0.8567 0.9256
No log 37.8333 454 0.8453 0.5483 0.8453 0.9194
No log 38.0 456 0.8539 0.5392 0.8539 0.9241
No log 38.1667 458 0.8584 0.5392 0.8584 0.9265
No log 38.3333 460 0.9139 0.5522 0.9139 0.9560
No log 38.5 462 0.9381 0.5852 0.9381 0.9685
No log 38.6667 464 0.9010 0.5932 0.9010 0.9492
No log 38.8333 466 0.8134 0.6298 0.8134 0.9019
No log 39.0 468 0.8115 0.6465 0.8115 0.9008
No log 39.1667 470 0.8612 0.6043 0.8612 0.9280
No log 39.3333 472 0.9315 0.5651 0.9315 0.9651
No log 39.5 474 0.9647 0.5155 0.9647 0.9822
No log 39.6667 476 0.8907 0.5717 0.8907 0.9438
No log 39.8333 478 0.8141 0.6147 0.8141 0.9023
No log 40.0 480 0.7793 0.6329 0.7793 0.8828
No log 40.1667 482 0.8073 0.6184 0.8073 0.8985
No log 40.3333 484 0.8508 0.6002 0.8508 0.9224
No log 40.5 486 0.8487 0.5979 0.8487 0.9213
No log 40.6667 488 0.8427 0.5786 0.8427 0.9180
No log 40.8333 490 0.9014 0.5336 0.9014 0.9494
No log 41.0 492 0.8850 0.5149 0.8850 0.9407
No log 41.1667 494 0.8151 0.5348 0.8151 0.9028
No log 41.3333 496 0.7535 0.6042 0.7535 0.8680
No log 41.5 498 0.7285 0.6211 0.7285 0.8535
0.3659 41.6667 500 0.7605 0.6070 0.7605 0.8720
0.3659 41.8333 502 0.8407 0.6009 0.8407 0.9169
0.3659 42.0 504 0.8931 0.6248 0.8931 0.9450
0.3659 42.1667 506 0.8874 0.6195 0.8874 0.9420
0.3659 42.3333 508 0.8406 0.6127 0.8406 0.9169
0.3659 42.5 510 0.8118 0.5843 0.8118 0.9010
0.3659 42.6667 512 0.8256 0.5716 0.8256 0.9086
0.3659 42.8333 514 0.8136 0.5947 0.8136 0.9020
0.3659 43.0 516 0.7964 0.6139 0.7964 0.8924
0.3659 43.1667 518 0.7576 0.6095 0.7576 0.8704
0.3659 43.3333 520 0.7656 0.6268 0.7656 0.8750
0.3659 43.5 522 0.8197 0.6303 0.8197 0.9054
0.3659 43.6667 524 0.8702 0.6218 0.8702 0.9328
0.3659 43.8333 526 0.8796 0.6108 0.8796 0.9379
0.3659 44.0 528 0.8497 0.6108 0.8497 0.9218
0.3659 44.1667 530 0.7770 0.6321 0.7770 0.8815
0.3659 44.3333 532 0.7298 0.6497 0.7298 0.8543
0.3659 44.5 534 0.7334 0.6357 0.7334 0.8564
0.3659 44.6667 536 0.7655 0.6238 0.7655 0.8749
0.3659 44.8333 538 0.8307 0.5995 0.8307 0.9114
0.3659 45.0 540 0.8848 0.5760 0.8848 0.9406
0.3659 45.1667 542 0.9111 0.5608 0.9111 0.9545
0.3659 45.3333 544 0.8834 0.5567 0.8834 0.9399
0.3659 45.5 546 0.8113 0.5917 0.8113 0.9007

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
7
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k2_task1_organization

Finetuned
(4222)
this model