ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k4_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8913
  • Qwk: 0.6715
  • Mse: 0.8913
  • Rmse: 0.9441

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1053 2 6.8696 0.0061 6.8696 2.6210
No log 0.2105 4 4.6169 0.0591 4.6169 2.1487
No log 0.3158 6 3.5045 0.0109 3.5045 1.8720
No log 0.4211 8 2.5481 0.0392 2.5481 1.5963
No log 0.5263 10 2.3105 0.2267 2.3105 1.5200
No log 0.6316 12 2.1995 0.2958 2.1995 1.4831
No log 0.7368 14 1.6595 0.2712 1.6595 1.2882
No log 0.8421 16 1.6182 0.2037 1.6182 1.2721
No log 0.9474 18 1.6610 0.2364 1.6610 1.2888
No log 1.0526 20 1.5654 0.3025 1.5654 1.2512
No log 1.1579 22 1.7794 0.3008 1.7794 1.3340
No log 1.2632 24 1.7775 0.3008 1.7775 1.3332
No log 1.3684 26 1.4108 0.4 1.4108 1.1878
No log 1.4737 28 1.3059 0.3103 1.3059 1.1428
No log 1.5789 30 1.2291 0.4098 1.2291 1.1086
No log 1.6842 32 1.3498 0.4776 1.3498 1.1618
No log 1.7895 34 1.8655 0.3576 1.8655 1.3658
No log 1.8947 36 2.7085 0.2816 2.7085 1.6458
No log 2.0 38 2.5008 0.3317 2.5008 1.5814
No log 2.1053 40 1.6455 0.475 1.6455 1.2828
No log 2.2105 42 1.2633 0.4733 1.2633 1.1240
No log 2.3158 44 1.1954 0.4923 1.1954 1.0933
No log 2.4211 46 1.2253 0.5255 1.2253 1.1069
No log 2.5263 48 1.5997 0.4744 1.5997 1.2648
No log 2.6316 50 2.0884 0.3353 2.0884 1.4451
No log 2.7368 52 2.0324 0.3436 2.0324 1.4256
No log 2.8421 54 1.2982 0.5379 1.2982 1.1394
No log 2.9474 56 0.7814 0.6715 0.7814 0.8840
No log 3.0526 58 0.7442 0.6957 0.7442 0.8627
No log 3.1579 60 0.9107 0.6569 0.9107 0.9543
No log 3.2632 62 1.0364 0.6377 1.0364 1.0181
No log 3.3684 64 1.1188 0.6383 1.1188 1.0577
No log 3.4737 66 0.9442 0.6569 0.9442 0.9717
No log 3.5789 68 0.8980 0.6569 0.8980 0.9476
No log 3.6842 70 0.8518 0.7034 0.8518 0.9229
No log 3.7895 72 0.7911 0.6928 0.7911 0.8894
No log 3.8947 74 0.7580 0.7361 0.7580 0.8707
No log 4.0 76 0.7512 0.7436 0.7512 0.8667
No log 4.1053 78 0.7992 0.7143 0.7992 0.8940
No log 4.2105 80 0.9068 0.6800 0.9068 0.9523
No log 4.3158 82 0.9256 0.6338 0.9256 0.9621
No log 4.4211 84 0.8496 0.7068 0.8496 0.9217
No log 4.5263 86 0.8022 0.6565 0.8022 0.8957
No log 4.6316 88 0.9640 0.5882 0.9640 0.9818
No log 4.7368 90 0.9217 0.5926 0.9217 0.9601
No log 4.8421 92 0.7407 0.6866 0.7407 0.8607
No log 4.9474 94 0.6941 0.7143 0.6941 0.8331
No log 5.0526 96 0.6450 0.7042 0.6450 0.8031
No log 5.1579 98 0.6129 0.7361 0.6129 0.7829
No log 5.2632 100 0.6093 0.7413 0.6093 0.7806
No log 5.3684 102 0.7118 0.7607 0.7118 0.8437
No log 5.4737 104 0.7808 0.7356 0.7808 0.8836
No log 5.5789 106 0.6837 0.7532 0.6837 0.8269
No log 5.6842 108 0.6284 0.7361 0.6284 0.7927
No log 5.7895 110 0.6613 0.7183 0.6613 0.8132
No log 5.8947 112 0.6306 0.7465 0.6306 0.7941
No log 6.0 114 0.8659 0.7771 0.8659 0.9306
No log 6.1053 116 1.1189 0.6631 1.1189 1.0578
No log 6.2105 118 0.8722 0.7640 0.8722 0.9339
No log 6.3158 120 0.6307 0.7826 0.6307 0.7942
No log 6.4211 122 0.7350 0.7260 0.7350 0.8573
No log 6.5263 124 0.7206 0.7172 0.7206 0.8489
No log 6.6316 126 0.7214 0.7550 0.7214 0.8494
No log 6.7368 128 0.9886 0.7152 0.9886 0.9943
No log 6.8421 130 1.0730 0.6272 1.0730 1.0358
No log 6.9474 132 0.8266 0.7317 0.8266 0.9092
No log 7.0526 134 0.6995 0.7260 0.6995 0.8364
No log 7.1579 136 0.7380 0.7172 0.7380 0.8591
No log 7.2632 138 0.6891 0.7361 0.6891 0.8301
No log 7.3684 140 0.9548 0.6788 0.9548 0.9771
No log 7.4737 142 1.3232 0.6036 1.3232 1.1503
No log 7.5789 144 1.1591 0.6182 1.1591 1.0766
No log 7.6842 146 0.8727 0.6710 0.8727 0.9342
No log 7.7895 148 0.7659 0.7034 0.7659 0.8751
No log 7.8947 150 0.7691 0.7133 0.7691 0.8770
No log 8.0 152 0.8098 0.7044 0.8098 0.8999
No log 8.1053 154 0.8690 0.6835 0.8690 0.9322
No log 8.2105 156 0.7654 0.7013 0.7654 0.8749
No log 8.3158 158 0.7615 0.7013 0.7615 0.8726
No log 8.4211 160 0.8082 0.6928 0.8082 0.8990
No log 8.5263 162 0.7780 0.6980 0.7780 0.8821
No log 8.6316 164 0.8332 0.7179 0.8332 0.9128
No log 8.7368 166 0.8274 0.7179 0.8274 0.9096
No log 8.8421 168 0.7857 0.6968 0.7857 0.8864
No log 8.9474 170 0.7784 0.6887 0.7784 0.8823
No log 9.0526 172 0.8070 0.7179 0.8070 0.8984
No log 9.1579 174 0.7264 0.7342 0.7264 0.8523
No log 9.2632 176 0.7129 0.7692 0.7129 0.8443
No log 9.3684 178 0.6077 0.7976 0.6077 0.7796
No log 9.4737 180 0.5954 0.8095 0.5954 0.7716
No log 9.5789 182 0.6395 0.8047 0.6395 0.7997
No log 9.6842 184 0.6766 0.8023 0.6766 0.8225
No log 9.7895 186 0.8062 0.7345 0.8062 0.8979
No log 9.8947 188 0.9078 0.6941 0.9078 0.9528
No log 10.0 190 0.8041 0.7059 0.8041 0.8967
No log 10.1053 192 0.7284 0.7194 0.7284 0.8534
No log 10.2105 194 0.7333 0.6763 0.7333 0.8563
No log 10.3158 196 0.7919 0.6711 0.7919 0.8899
No log 10.4211 198 0.8148 0.6928 0.8148 0.9027
No log 10.5263 200 0.7534 0.6809 0.7534 0.8680
No log 10.6316 202 0.7209 0.6714 0.7209 0.8491
No log 10.7368 204 0.7976 0.6761 0.7976 0.8931
No log 10.8421 206 0.9284 0.6800 0.9284 0.9635
No log 10.9474 208 1.1067 0.6 1.1067 1.0520
No log 11.0526 210 1.1071 0.6 1.1071 1.0522
No log 11.1579 212 0.9248 0.6928 0.9248 0.9617
No log 11.2632 214 0.8350 0.7067 0.8350 0.9138
No log 11.3684 216 0.8205 0.7190 0.8205 0.9058
No log 11.4737 218 0.8292 0.6928 0.8292 0.9106
No log 11.5789 220 0.8629 0.7044 0.8629 0.9290
No log 11.6842 222 0.8947 0.7160 0.8947 0.9459
No log 11.7895 224 0.8980 0.7152 0.8980 0.9476
No log 11.8947 226 1.0454 0.6267 1.0454 1.0224
No log 12.0 228 1.0161 0.625 1.0161 1.0080
No log 12.1053 230 0.7577 0.7285 0.7577 0.8705
No log 12.2105 232 0.6805 0.7451 0.6805 0.8249
No log 12.3158 234 0.6709 0.7692 0.6709 0.8191
No log 12.4211 236 0.6840 0.7792 0.6840 0.8270
No log 12.5263 238 0.6921 0.7550 0.6921 0.8319
No log 12.6316 240 0.7327 0.7310 0.7327 0.8560
No log 12.7368 242 0.7992 0.7234 0.7992 0.8940
No log 12.8421 244 0.8859 0.6571 0.8859 0.9412
No log 12.9474 246 1.0556 0.5481 1.0556 1.0274
No log 13.0526 248 1.0729 0.5481 1.0729 1.0358
No log 13.1579 250 0.9671 0.6043 0.9671 0.9834
No log 13.2632 252 0.7938 0.6906 0.7938 0.8910
No log 13.3684 254 0.7104 0.7050 0.7104 0.8428
No log 13.4737 256 0.6819 0.7183 0.6819 0.8257
No log 13.5789 258 0.7395 0.7143 0.7395 0.8599
No log 13.6842 260 0.7634 0.6795 0.7634 0.8738
No log 13.7895 262 0.6800 0.7237 0.6800 0.8246
No log 13.8947 264 0.6716 0.7517 0.6716 0.8195
No log 14.0 266 0.7051 0.7310 0.7051 0.8397
No log 14.1053 268 0.7487 0.7133 0.7487 0.8653
No log 14.2105 270 0.7954 0.6901 0.7954 0.8919
No log 14.3158 272 0.8442 0.6571 0.8442 0.9188
No log 14.4211 274 0.8603 0.6260 0.8603 0.9275
No log 14.5263 276 0.8342 0.6667 0.8342 0.9134
No log 14.6316 278 0.7813 0.6901 0.7813 0.8839
No log 14.7368 280 0.7453 0.7051 0.7453 0.8633
No log 14.8421 282 0.6901 0.7051 0.6901 0.8307
No log 14.9474 284 0.6502 0.7722 0.6502 0.8063
No log 15.0526 286 0.6641 0.7722 0.6641 0.8149
No log 15.1579 288 0.6402 0.7722 0.6402 0.8001
No log 15.2632 290 0.6462 0.7683 0.6462 0.8039
No log 15.3684 292 0.6858 0.7683 0.6858 0.8281
No log 15.4737 294 0.7060 0.7607 0.7060 0.8403
No log 15.5789 296 0.7208 0.7722 0.7208 0.8490
No log 15.6842 298 0.7769 0.7059 0.7769 0.8814
No log 15.7895 300 0.8508 0.6887 0.8508 0.9224
No log 15.8947 302 0.9488 0.6623 0.9488 0.9741
No log 16.0 304 0.9805 0.6626 0.9805 0.9902
No log 16.1053 306 0.8563 0.7081 0.8563 0.9254
No log 16.2105 308 0.7279 0.7355 0.7279 0.8532
No log 16.3158 310 0.7397 0.7179 0.7397 0.8601
No log 16.4211 312 0.7407 0.7059 0.7407 0.8606
No log 16.5263 314 0.6937 0.7673 0.6937 0.8329
No log 16.6316 316 0.6865 0.7799 0.6865 0.8286
No log 16.7368 318 0.7355 0.7407 0.7355 0.8576
No log 16.8421 320 0.7798 0.7186 0.7798 0.8831
No log 16.9474 322 0.7994 0.7195 0.7994 0.8941
No log 17.0526 324 0.7775 0.7329 0.7775 0.8818
No log 17.1579 326 0.7807 0.7375 0.7807 0.8836
No log 17.2632 328 0.7859 0.7436 0.7859 0.8865
No log 17.3684 330 0.7683 0.7152 0.7683 0.8765
No log 17.4737 332 0.7579 0.7152 0.7579 0.8706
No log 17.5789 334 0.7481 0.7162 0.7481 0.8650
No log 17.6842 336 0.7693 0.6849 0.7693 0.8771
No log 17.7895 338 0.8245 0.6759 0.8245 0.9080
No log 17.8947 340 0.8247 0.6761 0.8247 0.9081
No log 18.0 342 0.7546 0.6809 0.7546 0.8687
No log 18.1053 344 0.7441 0.6809 0.7441 0.8626
No log 18.2105 346 0.7143 0.7286 0.7143 0.8451
No log 18.3158 348 0.6997 0.7083 0.6997 0.8365
No log 18.4211 350 0.6887 0.7172 0.6887 0.8299
No log 18.5263 352 0.6856 0.7133 0.6856 0.8280
No log 18.6316 354 0.7048 0.7273 0.7048 0.8395
No log 18.7368 356 0.7097 0.7183 0.7097 0.8424
No log 18.8421 358 0.7100 0.7413 0.7100 0.8426
No log 18.9474 360 0.8071 0.6849 0.8071 0.8984
No log 19.0526 362 0.9550 0.7020 0.9550 0.9772
No log 19.1579 364 0.9477 0.6933 0.9477 0.9735
No log 19.2632 366 0.8372 0.6849 0.8372 0.9150
No log 19.3684 368 0.7612 0.7234 0.7612 0.8725
No log 19.4737 370 0.7293 0.7234 0.7293 0.8540
No log 19.5789 372 0.7150 0.7448 0.7150 0.8456
No log 19.6842 374 0.7483 0.7179 0.7483 0.8650
No log 19.7895 376 0.7749 0.7342 0.7749 0.8803
No log 19.8947 378 0.7992 0.6879 0.7992 0.8940
No log 20.0 380 0.7713 0.7320 0.7713 0.8782
No log 20.1053 382 0.7898 0.7105 0.7898 0.8887
No log 20.2105 384 0.7819 0.6980 0.7819 0.8843
No log 20.3158 386 0.7430 0.7162 0.7430 0.8620
No log 20.4211 388 0.7235 0.7413 0.7235 0.8506
No log 20.5263 390 0.7347 0.7234 0.7347 0.8572
No log 20.6316 392 0.7282 0.7234 0.7282 0.8533
No log 20.7368 394 0.6991 0.7465 0.6991 0.8361
No log 20.8421 396 0.6973 0.7417 0.6973 0.8351
No log 20.9474 398 0.6881 0.7333 0.6881 0.8295
No log 21.0526 400 0.6682 0.7417 0.6682 0.8174
No log 21.1579 402 0.6541 0.7733 0.6541 0.8087
No log 21.2632 404 0.6602 0.7733 0.6602 0.8125
No log 21.3684 406 0.6695 0.7898 0.6695 0.8182
No log 21.4737 408 0.6716 0.7925 0.6716 0.8195
No log 21.5789 410 0.6945 0.7799 0.6945 0.8334
No log 21.6842 412 0.7082 0.7561 0.7082 0.8416
No log 21.7895 414 0.7288 0.7381 0.7288 0.8537
No log 21.8947 416 0.6789 0.7654 0.6789 0.8240
No log 22.0 418 0.6346 0.7771 0.6346 0.7966
No log 22.1053 420 0.6253 0.7875 0.6253 0.7908
No log 22.2105 422 0.6702 0.7625 0.6702 0.8186
No log 22.3158 424 0.6994 0.7485 0.6994 0.8363
No log 22.4211 426 0.6625 0.7771 0.6625 0.8139
No log 22.5263 428 0.6731 0.7875 0.6731 0.8204
No log 22.6316 430 0.7078 0.7205 0.7078 0.8413
No log 22.7368 432 0.8430 0.7209 0.8430 0.9182
No log 22.8421 434 0.9298 0.6821 0.9298 0.9643
No log 22.9474 436 0.8571 0.7030 0.8571 0.9258
No log 23.0526 438 0.7670 0.7407 0.7670 0.8758
No log 23.1579 440 0.7128 0.7453 0.7128 0.8443
No log 23.2632 442 0.6736 0.7831 0.6736 0.8207
No log 23.3684 444 0.6867 0.7784 0.6867 0.8287
No log 23.4737 446 0.7617 0.7186 0.7617 0.8728
No log 23.5789 448 0.7782 0.7117 0.7782 0.8822
No log 23.6842 450 0.7214 0.7273 0.7214 0.8493
No log 23.7895 452 0.7008 0.7347 0.7008 0.8372
No log 23.8947 454 0.7136 0.7361 0.7136 0.8448
No log 24.0 456 0.7166 0.7483 0.7166 0.8465
No log 24.1053 458 0.7455 0.7532 0.7455 0.8634
No log 24.2105 460 0.8306 0.6879 0.8306 0.9114
No log 24.3158 462 0.8417 0.6879 0.8417 0.9175
No log 24.4211 464 0.8272 0.6757 0.8272 0.9095
No log 24.5263 466 0.8259 0.6763 0.8259 0.9088
No log 24.6316 468 0.8114 0.6716 0.8114 0.9008
No log 24.7368 470 0.8215 0.6617 0.8215 0.9064
No log 24.8421 472 0.8568 0.6429 0.8568 0.9256
No log 24.9474 474 0.8686 0.6757 0.8686 0.9320
No log 25.0526 476 0.8895 0.6803 0.8895 0.9431
No log 25.1579 478 0.8645 0.6980 0.8645 0.9298
No log 25.2632 480 0.8642 0.6968 0.8642 0.9296
No log 25.3684 482 0.8036 0.7389 0.8036 0.8964
No log 25.4737 484 0.7587 0.7261 0.7587 0.8710
No log 25.5789 486 0.7420 0.7296 0.7420 0.8614
No log 25.6842 488 0.7376 0.7308 0.7376 0.8588
No log 25.7895 490 0.7366 0.7237 0.7366 0.8583
No log 25.8947 492 0.7359 0.7237 0.7359 0.8578
No log 26.0 494 0.7339 0.7342 0.7339 0.8567
No log 26.1053 496 0.7417 0.7484 0.7417 0.8612
No log 26.2105 498 0.7649 0.7285 0.7649 0.8746
0.3124 26.3158 500 0.7969 0.7067 0.7969 0.8927
0.3124 26.4211 502 0.8026 0.6846 0.8026 0.8959
0.3124 26.5263 504 0.7937 0.6846 0.7937 0.8909
0.3124 26.6316 506 0.7713 0.7067 0.7713 0.8782
0.3124 26.7368 508 0.7573 0.7152 0.7573 0.8702
0.3124 26.8421 510 0.7472 0.7237 0.7472 0.8644
0.3124 26.9474 512 0.7364 0.7134 0.7364 0.8581
0.3124 27.0526 514 0.7615 0.6968 0.7615 0.8726
0.3124 27.1579 516 0.7938 0.6797 0.7938 0.8909
0.3124 27.2632 518 0.8107 0.6577 0.8107 0.9004
0.3124 27.3684 520 0.8467 0.6714 0.8467 0.9201
0.3124 27.4737 522 0.8913 0.6715 0.8913 0.9441

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
182
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k4_task1_organization

Finetuned
(4222)
this model