ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k9_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9405
  • Qwk: 0.5669
  • Mse: 0.9405
  • Rmse: 0.9698

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0476 2 6.9937 -0.0056 6.9937 2.6446
No log 0.0952 4 4.9773 0.0449 4.9773 2.2310
No log 0.1429 6 3.2285 0.0606 3.2285 1.7968
No log 0.1905 8 3.2540 -0.0113 3.2540 1.8039
No log 0.2381 10 2.3518 0.1791 2.3518 1.5336
No log 0.2857 12 1.9102 0.1724 1.9102 1.3821
No log 0.3333 14 2.0098 0.1739 2.0098 1.4177
No log 0.3810 16 1.9523 0.1455 1.9523 1.3973
No log 0.4286 18 1.9261 0.2906 1.9261 1.3878
No log 0.4762 20 2.0530 0.2923 2.0530 1.4328
No log 0.5238 22 2.1998 0.1429 2.1998 1.4832
No log 0.5714 24 2.1184 0.25 2.1184 1.4555
No log 0.6190 26 2.0691 0.2519 2.0691 1.4384
No log 0.6667 28 2.3789 0.1867 2.3789 1.5424
No log 0.7143 30 2.9783 0.1183 2.9783 1.7258
No log 0.7619 32 2.8852 0.1183 2.8852 1.6986
No log 0.8095 34 2.3513 0.2375 2.3513 1.5334
No log 0.8571 36 1.6827 0.3453 1.6827 1.2972
No log 0.9048 38 1.4126 0.384 1.4126 1.1885
No log 0.9524 40 1.3256 0.4032 1.3256 1.1513
No log 1.0 42 1.3635 0.4308 1.3635 1.1677
No log 1.0476 44 1.3611 0.4361 1.3611 1.1667
No log 1.0952 46 1.4554 0.4296 1.4554 1.2064
No log 1.1429 48 1.4497 0.4493 1.4497 1.2040
No log 1.1905 50 1.4786 0.4375 1.4786 1.2160
No log 1.2381 52 1.5278 0.4409 1.5278 1.2360
No log 1.2857 54 1.5598 0.4394 1.5598 1.2489
No log 1.3333 56 1.7453 0.4058 1.7453 1.3211
No log 1.3810 58 1.9344 0.3097 1.9344 1.3908
No log 1.4286 60 2.2661 0.3006 2.2661 1.5054
No log 1.4762 62 2.2867 0.3095 2.2867 1.5122
No log 1.5238 64 2.4212 0.3095 2.4212 1.5560
No log 1.5714 66 2.0451 0.3095 2.0451 1.4301
No log 1.6190 68 1.7252 0.3797 1.7252 1.3135
No log 1.6667 70 1.5447 0.4336 1.5447 1.2429
No log 1.7143 72 1.3248 0.4662 1.3248 1.1510
No log 1.7619 74 1.2932 0.4848 1.2932 1.1372
No log 1.8095 76 1.2699 0.4848 1.2699 1.1269
No log 1.8571 78 1.3201 0.5315 1.3201 1.1489
No log 1.9048 80 1.4469 0.5235 1.4469 1.2029
No log 1.9524 82 1.4105 0.5235 1.4105 1.1877
No log 2.0 84 1.1617 0.4427 1.1617 1.0778
No log 2.0476 86 1.0823 0.4 1.0823 1.0403
No log 2.0952 88 1.0451 0.5806 1.0451 1.0223
No log 2.1429 90 1.0040 0.5758 1.0040 1.0020
No log 2.1905 92 1.0617 0.5224 1.0617 1.0304
No log 2.2381 94 1.3775 0.4818 1.3775 1.1737
No log 2.2857 96 1.7681 0.4304 1.7681 1.3297
No log 2.3333 98 1.6312 0.4713 1.6312 1.2772
No log 2.3810 100 1.2848 0.5455 1.2848 1.1335
No log 2.4286 102 0.9554 0.5077 0.9554 0.9774
No log 2.4762 104 0.9655 0.4918 0.9655 0.9826
No log 2.5238 106 0.9790 0.608 0.9790 0.9894
No log 2.5714 108 0.8916 0.6667 0.8916 0.9442
No log 2.6190 110 0.9158 0.6383 0.9158 0.9570
No log 2.6667 112 0.9624 0.6667 0.9624 0.9810
No log 2.7143 114 0.9050 0.6438 0.9050 0.9513
No log 2.7619 116 0.9940 0.6202 0.9940 0.9970
No log 2.8095 118 1.2378 0.4444 1.2378 1.1126
No log 2.8571 120 1.2718 0.4724 1.2718 1.1277
No log 2.9048 122 1.1150 0.5606 1.1150 1.0559
No log 2.9524 124 1.0310 0.6154 1.0310 1.0154
No log 3.0 126 1.0967 0.6316 1.0967 1.0472
No log 3.0476 128 1.1075 0.6053 1.1075 1.0524
No log 3.0952 130 1.0776 0.5753 1.0776 1.0381
No log 3.1429 132 1.2206 0.5789 1.2206 1.1048
No log 3.1905 134 1.5631 0.4878 1.5631 1.2502
No log 3.2381 136 1.5977 0.4878 1.5977 1.2640
No log 3.2857 138 1.1629 0.5694 1.1629 1.0784
No log 3.3333 140 0.9882 0.5714 0.9882 0.9941
No log 3.3810 142 0.9913 0.6154 0.9913 0.9957
No log 3.4286 144 0.9761 0.6061 0.9761 0.9880
No log 3.4762 146 0.9615 0.5985 0.9615 0.9806
No log 3.5238 148 1.0058 0.5865 1.0058 1.0029
No log 3.5714 150 1.1215 0.5152 1.1215 1.0590
No log 3.6190 152 1.1637 0.5077 1.1637 1.0788
No log 3.6667 154 1.1979 0.4806 1.1979 1.0945
No log 3.7143 156 1.2464 0.5113 1.2464 1.1164
No log 3.7619 158 1.3786 0.4681 1.3786 1.1741
No log 3.8095 160 1.5380 0.4540 1.5380 1.2401
No log 3.8571 162 1.5631 0.4734 1.5631 1.2502
No log 3.9048 164 1.3009 0.5139 1.3009 1.1406
No log 3.9524 166 1.0880 0.5344 1.0880 1.0431
No log 4.0 168 1.1841 0.5042 1.1841 1.0882
No log 4.0476 170 1.2186 0.5041 1.2186 1.1039
No log 4.0952 172 1.0965 0.5645 1.0965 1.0472
No log 4.1429 174 1.0878 0.5693 1.0878 1.0430
No log 4.1905 176 1.2922 0.5395 1.2922 1.1367
No log 4.2381 178 1.3185 0.5548 1.3185 1.1483
No log 4.2857 180 1.0728 0.5571 1.0728 1.0358
No log 4.3333 182 0.9947 0.5985 0.9947 0.9974
No log 4.3810 184 0.9857 0.6412 0.9857 0.9928
No log 4.4286 186 0.9445 0.6364 0.9445 0.9719
No log 4.4762 188 0.9443 0.5846 0.9443 0.9718
No log 4.5238 190 0.9245 0.6364 0.9245 0.9615
No log 4.5714 192 0.9660 0.5909 0.9660 0.9828
No log 4.6190 194 0.9654 0.6074 0.9654 0.9825
No log 4.6667 196 0.9147 0.6434 0.9147 0.9564
No log 4.7143 198 0.9539 0.6486 0.9539 0.9767
No log 4.7619 200 0.9517 0.6974 0.9517 0.9755
No log 4.8095 202 0.9010 0.6939 0.9010 0.9492
No log 4.8571 204 0.8685 0.6939 0.8685 0.9319
No log 4.9048 206 0.8378 0.7389 0.8378 0.9153
No log 4.9524 208 0.8106 0.7285 0.8106 0.9003
No log 5.0 210 0.8547 0.6957 0.8547 0.9245
No log 5.0476 212 0.9953 0.6222 0.9953 0.9977
No log 5.0952 214 0.9890 0.6471 0.9890 0.9945
No log 5.1429 216 0.8621 0.6617 0.8621 0.9285
No log 5.1905 218 0.8004 0.7260 0.8004 0.8946
No log 5.2381 220 0.7978 0.6906 0.7978 0.8932
No log 5.2857 222 0.8506 0.6423 0.8506 0.9223
No log 5.3333 224 0.8448 0.6667 0.8448 0.9191
No log 5.3810 226 0.8033 0.6963 0.8033 0.8963
No log 5.4286 228 0.8100 0.7007 0.8100 0.9000
No log 5.4762 230 0.8448 0.6260 0.8448 0.9191
No log 5.5238 232 0.8618 0.6212 0.8618 0.9284
No log 5.5714 234 0.8699 0.6269 0.8699 0.9327
No log 5.6190 236 0.8527 0.6212 0.8527 0.9234
No log 5.6667 238 0.8964 0.6667 0.8964 0.9468
No log 5.7143 240 0.9040 0.6567 0.9040 0.9508
No log 5.7619 242 0.8907 0.6715 0.8907 0.9437
No log 5.8095 244 0.9416 0.5821 0.9416 0.9704
No log 5.8571 246 0.9955 0.5303 0.9955 0.9977
No log 5.9048 248 1.0780 0.5652 1.0780 1.0382
No log 5.9524 250 1.1215 0.5588 1.1215 1.0590
No log 6.0 252 1.1432 0.5874 1.1432 1.0692
No log 6.0476 254 1.0145 0.6338 1.0145 1.0072
No log 6.0952 256 0.9076 0.6286 0.9076 0.9527
No log 6.1429 258 0.8639 0.6619 0.8639 0.9295
No log 6.1905 260 0.8295 0.6809 0.8295 0.9108
No log 6.2381 262 0.7895 0.7083 0.7895 0.8886
No log 6.2857 264 0.8056 0.6571 0.8056 0.8976
No log 6.3333 266 0.8518 0.6573 0.8518 0.9229
No log 6.3810 268 0.8515 0.6713 0.8515 0.9227
No log 6.4286 270 0.8142 0.6331 0.8142 0.9023
No log 6.4762 272 0.8439 0.7 0.8439 0.9186
No log 6.5238 274 0.8382 0.6906 0.8382 0.9155
No log 6.5714 276 0.7527 0.7133 0.7527 0.8676
No log 6.6190 278 0.6961 0.7619 0.6961 0.8344
No log 6.6667 280 0.7278 0.7578 0.7278 0.8531
No log 6.7143 282 0.7384 0.7389 0.7384 0.8593
No log 6.7619 284 0.7293 0.7347 0.7293 0.8540
No log 6.8095 286 0.7503 0.7222 0.7503 0.8662
No log 6.8571 288 0.8441 0.6316 0.8441 0.9188
No log 6.9048 290 0.8717 0.6370 0.8717 0.9336
No log 6.9524 292 0.8575 0.6412 0.8575 0.9260
No log 7.0 294 0.8191 0.6515 0.8191 0.9051
No log 7.0476 296 0.7941 0.6866 0.7941 0.8911
No log 7.0952 298 0.7909 0.7042 0.7909 0.8893
No log 7.1429 300 0.8062 0.6761 0.8062 0.8979
No log 7.1905 302 0.8419 0.6286 0.8419 0.9176
No log 7.2381 304 0.8582 0.7160 0.8582 0.9264
No log 7.2857 306 0.8251 0.6759 0.8251 0.9083
No log 7.3333 308 0.7948 0.6853 0.7948 0.8915
No log 7.3810 310 0.7921 0.6525 0.7921 0.8900
No log 7.4286 312 0.8171 0.6479 0.8171 0.9039
No log 7.4762 314 0.8220 0.6667 0.8220 0.9067
No log 7.5238 316 0.8149 0.6522 0.8148 0.9027
No log 7.5714 318 0.8212 0.6761 0.8212 0.9062
No log 7.6190 320 0.9163 0.6438 0.9163 0.9572
No log 7.6667 322 1.0493 0.6093 1.0493 1.0244
No log 7.7143 324 0.9308 0.6395 0.9308 0.9648
No log 7.7619 326 0.8265 0.6571 0.8265 0.9091
No log 7.8095 328 0.8164 0.6853 0.8164 0.9036
No log 7.8571 330 0.7936 0.7083 0.7936 0.8908
No log 7.9048 332 0.7891 0.6853 0.7891 0.8883
No log 7.9524 334 0.7981 0.7083 0.7981 0.8934
No log 8.0 336 0.8077 0.7222 0.8077 0.8987
No log 8.0476 338 0.8133 0.6853 0.8133 0.9018
No log 8.0952 340 0.9132 0.6164 0.9132 0.9556
No log 8.1429 342 0.9788 0.6415 0.9788 0.9893
No log 8.1905 344 0.9108 0.65 0.9108 0.9544
No log 8.2381 346 0.8261 0.6533 0.8261 0.9089
No log 8.2857 348 0.7365 0.7722 0.7365 0.8582
No log 8.3333 350 0.7079 0.7682 0.7079 0.8413
No log 8.3810 352 0.6945 0.7821 0.6945 0.8334
No log 8.4286 354 0.7019 0.76 0.7019 0.8378
No log 8.4762 356 0.7366 0.7568 0.7366 0.8582
No log 8.5238 358 0.8075 0.7092 0.8075 0.8986
No log 8.5714 360 0.8390 0.7015 0.8390 0.9160
No log 8.6190 362 0.8466 0.6667 0.8466 0.9201
No log 8.6667 364 0.8161 0.6567 0.8161 0.9034
No log 8.7143 366 0.7786 0.7083 0.7786 0.8824
No log 8.7619 368 0.7347 0.75 0.7347 0.8572
No log 8.8095 370 0.6813 0.7848 0.6813 0.8254
No log 8.8571 372 0.6545 0.7898 0.6545 0.8090
No log 8.9048 374 0.6761 0.7361 0.6761 0.8223
No log 8.9524 376 0.7207 0.7194 0.7207 0.8490
No log 9.0 378 0.7354 0.7143 0.7354 0.8575
No log 9.0476 380 0.7234 0.7338 0.7234 0.8505
No log 9.0952 382 0.7292 0.7234 0.7292 0.8539
No log 9.1429 384 0.7922 0.7143 0.7922 0.8901
No log 9.1905 386 0.8679 0.6861 0.8679 0.9316
No log 9.2381 388 0.8740 0.6861 0.8740 0.9349
No log 9.2857 390 0.7704 0.7143 0.7704 0.8777
No log 9.3333 392 0.6829 0.7755 0.6829 0.8264
No log 9.3810 394 0.6725 0.8025 0.6725 0.8201
No log 9.4286 396 0.6778 0.8 0.6778 0.8233
No log 9.4762 398 0.7055 0.7273 0.7055 0.8399
No log 9.5238 400 0.7596 0.6861 0.7596 0.8715
No log 9.5714 402 0.7969 0.6861 0.7969 0.8927
No log 9.6190 404 0.8283 0.6569 0.8283 0.9101
No log 9.6667 406 0.8421 0.6269 0.8421 0.9177
No log 9.7143 408 0.8597 0.5778 0.8597 0.9272
No log 9.7619 410 0.8625 0.5926 0.8625 0.9287
No log 9.8095 412 0.8426 0.6029 0.8426 0.9179
No log 9.8571 414 0.8160 0.6087 0.8160 0.9033
No log 9.9048 416 0.8004 0.6241 0.8004 0.8947
No log 9.9524 418 0.8236 0.6483 0.8236 0.9075
No log 10.0 420 0.7485 0.7020 0.7485 0.8652
No log 10.0476 422 0.6942 0.7368 0.6942 0.8332
No log 10.0952 424 0.6915 0.7448 0.6915 0.8316
No log 10.1429 426 0.7266 0.7111 0.7266 0.8524
No log 10.1905 428 0.7238 0.7111 0.7238 0.8508
No log 10.2381 430 0.6806 0.7550 0.6806 0.8250
No log 10.2857 432 0.6983 0.7550 0.6983 0.8356
No log 10.3333 434 0.7256 0.7347 0.7256 0.8518
No log 10.3810 436 0.7825 0.6812 0.7825 0.8846
No log 10.4286 438 0.8013 0.7133 0.8013 0.8951
No log 10.4762 440 0.8069 0.7050 0.8069 0.8983
No log 10.5238 442 0.8694 0.6815 0.8694 0.9324
No log 10.5714 444 0.9044 0.6618 0.9044 0.9510
No log 10.6190 446 0.8631 0.6861 0.8631 0.9290
No log 10.6667 448 0.7440 0.7194 0.7440 0.8626
No log 10.7143 450 0.6741 0.7417 0.6741 0.8210
No log 10.7619 452 0.6736 0.7792 0.6736 0.8207
No log 10.8095 454 0.6883 0.7417 0.6883 0.8296
No log 10.8571 456 0.7501 0.7092 0.7501 0.8661
No log 10.9048 458 0.8131 0.6667 0.8131 0.9017
No log 10.9524 460 0.7942 0.7034 0.7942 0.8912
No log 11.0 462 0.7815 0.7034 0.7815 0.8840
No log 11.0476 464 0.7726 0.7285 0.7726 0.8790
No log 11.0952 466 0.7649 0.7451 0.7649 0.8746
No log 11.1429 468 0.7601 0.7183 0.7601 0.8718
No log 11.1905 470 0.7644 0.7101 0.7644 0.8743
No log 11.2381 472 0.7425 0.6901 0.7425 0.8617
No log 11.2857 474 0.7216 0.7123 0.7216 0.8495
No log 11.3333 476 0.6927 0.7355 0.6927 0.8323
No log 11.3810 478 0.7122 0.7657 0.7122 0.8439
No log 11.4286 480 0.8340 0.7356 0.8340 0.9133
No log 11.4762 482 0.8255 0.7356 0.8255 0.9085
No log 11.5238 484 0.7274 0.7547 0.7274 0.8529
No log 11.5714 486 0.6956 0.75 0.6956 0.8340
No log 11.6190 488 0.7530 0.6765 0.7530 0.8678
No log 11.6667 490 0.7981 0.6765 0.7981 0.8934
No log 11.7143 492 0.8027 0.7015 0.8027 0.8959
No log 11.7619 494 0.7954 0.7194 0.7954 0.8919
No log 11.8095 496 0.8357 0.6029 0.8357 0.9142
No log 11.8571 498 0.8497 0.6479 0.8497 0.9218
0.3968 11.9048 500 0.8060 0.6857 0.8060 0.8978
0.3968 11.9524 502 0.7609 0.7518 0.7609 0.8723
0.3968 12.0 504 0.7321 0.7534 0.7321 0.8556
0.3968 12.0476 506 0.6980 0.7671 0.6980 0.8355
0.3968 12.0952 508 0.6723 0.7183 0.6723 0.8200
0.3968 12.1429 510 0.6591 0.7234 0.6591 0.8118
0.3968 12.1905 512 0.6614 0.7234 0.6614 0.8133
0.3968 12.2381 514 0.6569 0.7234 0.6569 0.8105
0.3968 12.2857 516 0.6647 0.7101 0.6647 0.8153
0.3968 12.3333 518 0.7142 0.6963 0.7142 0.8451
0.3968 12.3810 520 0.7622 0.6866 0.7622 0.8730
0.3968 12.4286 522 0.7714 0.6866 0.7714 0.8783
0.3968 12.4762 524 0.7804 0.6970 0.7804 0.8834
0.3968 12.5238 526 0.7938 0.6716 0.7938 0.8909
0.3968 12.5714 528 0.8070 0.6617 0.8070 0.8983
0.3968 12.6190 530 0.7825 0.6716 0.7825 0.8846
0.3968 12.6667 532 0.7482 0.7092 0.7482 0.8650
0.3968 12.7143 534 0.7361 0.7448 0.7361 0.8580
0.3968 12.7619 536 0.7237 0.7361 0.7237 0.8507
0.3968 12.8095 538 0.6954 0.7619 0.6954 0.8339
0.3968 12.8571 540 0.6766 0.7564 0.6766 0.8225
0.3968 12.9048 542 0.7142 0.7654 0.7142 0.8451
0.3968 12.9524 544 0.8040 0.7262 0.8040 0.8967
0.3968 13.0 546 0.8411 0.7326 0.8411 0.9171
0.3968 13.0476 548 0.7929 0.7657 0.7929 0.8905
0.3968 13.0952 550 0.7145 0.7805 0.7145 0.8453
0.3968 13.1429 552 0.7202 0.7211 0.7202 0.8486
0.3968 13.1905 554 0.7965 0.6906 0.7965 0.8925
0.3968 13.2381 556 0.9150 0.6567 0.9150 0.9566
0.3968 13.2857 558 0.9774 0.6515 0.9774 0.9886
0.3968 13.3333 560 1.0291 0.5714 1.0291 1.0144
0.3968 13.3810 562 1.0234 0.5203 1.0234 1.0116
0.3968 13.4286 564 0.9405 0.5669 0.9405 0.9698

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
182
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k9_task1_organization

Finetuned
(4222)
this model