ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k16_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8071
  • Qwk: -0.2030
  • Mse: 0.8071
  • Rmse: 0.8984

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0465 2 3.5954 -0.0057 3.5954 1.8962
No log 0.0930 4 2.1535 -0.0299 2.1535 1.4675
No log 0.1395 6 1.3280 0.0 1.3280 1.1524
No log 0.1860 8 0.8135 -0.0371 0.8135 0.9020
No log 0.2326 10 0.9803 -0.0261 0.9803 0.9901
No log 0.2791 12 0.9586 -0.0245 0.9586 0.9791
No log 0.3256 14 0.6774 0.1444 0.6774 0.8230
No log 0.3721 16 0.8110 0.1024 0.8110 0.9005
No log 0.4186 18 1.0336 0.0119 1.0336 1.0167
No log 0.4651 20 1.0362 0.0119 1.0362 1.0179
No log 0.5116 22 1.0978 0.0083 1.0978 1.0477
No log 0.5581 24 0.8590 -0.1271 0.8590 0.9268
No log 0.6047 26 0.7933 0.0071 0.7933 0.8907
No log 0.6512 28 1.1431 0.0569 1.1431 1.0692
No log 0.6977 30 0.7683 -0.0287 0.7683 0.8765
No log 0.7442 32 0.6734 0.0 0.6734 0.8206
No log 0.7907 34 0.6676 0.0 0.6676 0.8171
No log 0.8372 36 0.7490 0.0909 0.7490 0.8655
No log 0.8837 38 1.1585 -0.0479 1.1585 1.0763
No log 0.9302 40 1.4865 0.0 1.4865 1.2192
No log 0.9767 42 1.3285 -0.0247 1.3285 1.1526
No log 1.0233 44 1.0414 -0.0728 1.0414 1.0205
No log 1.0698 46 0.8461 0.0316 0.8461 0.9198
No log 1.1163 48 0.8227 -0.0408 0.8227 0.9070
No log 1.1628 50 0.7843 -0.0240 0.7843 0.8856
No log 1.2093 52 0.7206 -0.0035 0.7206 0.8489
No log 1.2558 54 0.7151 -0.0035 0.7151 0.8456
No log 1.3023 56 0.7378 -0.0131 0.7378 0.8590
No log 1.3488 58 0.7007 -0.0101 0.7007 0.8371
No log 1.3953 60 0.6836 -0.0101 0.6836 0.8268
No log 1.4419 62 0.7547 0.0628 0.7547 0.8687
No log 1.4884 64 0.7731 0.0043 0.7731 0.8792
No log 1.5349 66 0.8702 0.1947 0.8702 0.9328
No log 1.5814 68 0.7377 0.0334 0.7377 0.8589
No log 1.6279 70 0.7297 0.0416 0.7297 0.8542
No log 1.6744 72 0.7738 -0.0287 0.7738 0.8796
No log 1.7209 74 0.7428 -0.0035 0.7428 0.8619
No log 1.7674 76 0.7648 0.0 0.7648 0.8746
No log 1.8140 78 0.7640 -0.0035 0.7640 0.8741
No log 1.8605 80 0.7700 -0.0152 0.7700 0.8775
No log 1.9070 82 0.7959 0.0680 0.7959 0.8921
No log 1.9535 84 0.7898 -0.0627 0.7898 0.8887
No log 2.0 86 0.9057 -0.1474 0.9057 0.9517
No log 2.0465 88 0.9509 -0.2455 0.9509 0.9751
No log 2.0930 90 0.8540 -0.1905 0.8540 0.9241
No log 2.1395 92 0.8133 0.0303 0.8133 0.9018
No log 2.1860 94 0.8774 -0.1703 0.8774 0.9367
No log 2.2326 96 0.8713 -0.1583 0.8713 0.9334
No log 2.2791 98 0.8329 -0.1468 0.8329 0.9126
No log 2.3256 100 0.8666 -0.0870 0.8666 0.9309
No log 2.3721 102 0.9345 -0.0970 0.9345 0.9667
No log 2.4186 104 0.8735 -0.0870 0.8735 0.9346
No log 2.4651 106 0.8613 0.0959 0.8613 0.9281
No log 2.5116 108 0.8845 -0.0896 0.8845 0.9405
No log 2.5581 110 1.0859 -0.0820 1.0859 1.0421
No log 2.6047 112 1.0144 -0.1099 1.0144 1.0072
No log 2.6512 114 0.7966 -0.1610 0.7966 0.8925
No log 2.6977 116 0.8230 0.0099 0.8230 0.9072
No log 2.7442 118 0.7904 -0.1542 0.7904 0.8891
No log 2.7907 120 0.8578 -0.1263 0.8578 0.9262
No log 2.8372 122 0.8482 -0.1470 0.8482 0.9210
No log 2.8837 124 0.8849 0.0154 0.8849 0.9407
No log 2.9302 126 0.8717 -0.0302 0.8717 0.9336
No log 2.9767 128 0.9598 0.0107 0.9598 0.9797
No log 3.0233 130 0.9730 0.1149 0.9730 0.9864
No log 3.0698 132 0.8307 0.0066 0.8307 0.9114
No log 3.1163 134 0.8292 -0.0511 0.8292 0.9106
No log 3.1628 136 0.9597 0.0404 0.9597 0.9796
No log 3.2093 138 1.0092 0.0502 1.0092 1.0046
No log 3.2558 140 0.8717 -0.0941 0.8717 0.9337
No log 3.3023 142 0.8608 -0.0541 0.8608 0.9278
No log 3.3488 144 0.8167 -0.1529 0.8167 0.9037
No log 3.3953 146 0.9073 -0.0317 0.9073 0.9525
No log 3.4419 148 0.8833 -0.1142 0.8833 0.9398
No log 3.4884 150 0.7385 -0.1137 0.7385 0.8593
No log 3.5349 152 0.7693 0.0549 0.7693 0.8771
No log 3.5814 154 0.7710 0.0549 0.7710 0.8781
No log 3.6279 156 0.7355 -0.0069 0.7355 0.8576
No log 3.6744 158 0.7940 -0.0571 0.7940 0.8911
No log 3.7209 160 0.8048 -0.2144 0.8048 0.8971
No log 3.7674 162 0.8485 0.0152 0.8485 0.9212
No log 3.8140 164 0.8738 0.0216 0.8738 0.9348
No log 3.8605 166 0.9616 0.0107 0.9616 0.9806
No log 3.9070 168 1.1397 -0.0058 1.1397 1.0676
No log 3.9535 170 1.0364 -0.0181 1.0364 1.0180
No log 4.0 172 0.8195 0.0026 0.8195 0.9053
No log 4.0465 174 0.8465 0.0113 0.8465 0.9200
No log 4.0930 176 0.7969 -0.0465 0.7969 0.8927
No log 4.1395 178 0.8580 -0.0989 0.8580 0.9263
No log 4.1860 180 0.9453 -0.0363 0.9453 0.9723
No log 4.2326 182 1.1076 0.0487 1.1076 1.0524
No log 4.2791 184 0.9976 0.0734 0.9976 0.9988
No log 4.3256 186 0.9986 0.1078 0.9986 0.9993
No log 4.3721 188 1.0132 0.0431 1.0132 1.0066
No log 4.4186 190 0.9178 -0.0543 0.9178 0.9580
No log 4.4651 192 0.9674 0.0676 0.9674 0.9836
No log 4.5116 194 0.9277 0.0081 0.9277 0.9632
No log 4.5581 196 1.0345 0.0454 1.0345 1.0171
No log 4.6047 198 1.1201 -0.0721 1.1201 1.0583
No log 4.6512 200 0.9425 -0.0492 0.9425 0.9708
No log 4.6977 202 0.8929 -0.0465 0.8929 0.9449
No log 4.7442 204 0.9339 -0.0079 0.9339 0.9664
No log 4.7907 206 0.9593 -0.1166 0.9593 0.9794
No log 4.8372 208 1.1049 -0.1278 1.1049 1.0511
No log 4.8837 210 1.2938 -0.0283 1.2938 1.1374
No log 4.9302 212 1.2094 -0.0116 1.2094 1.0997
No log 4.9767 214 1.0368 0.0239 1.0368 1.0182
No log 5.0233 216 0.9897 0.0239 0.9897 0.9949
No log 5.0698 218 1.0241 -0.0551 1.0241 1.0120
No log 5.1163 220 0.9975 -0.0583 0.9975 0.9987
No log 5.1628 222 0.9682 -0.1011 0.9682 0.9840
No log 5.2093 224 0.8370 -0.1871 0.8370 0.9149
No log 5.2558 226 0.8720 0.0205 0.8720 0.9338
No log 5.3023 228 0.8636 -0.1470 0.8636 0.9293
No log 5.3488 230 1.0441 0.0180 1.0441 1.0218
No log 5.3953 232 1.0206 -0.0563 1.0206 1.0103
No log 5.4419 234 0.9095 -0.2128 0.9095 0.9537
No log 5.4884 236 0.9573 -0.0946 0.9573 0.9784
No log 5.5349 238 0.9488 -0.1760 0.9488 0.9741
No log 5.5814 240 1.1666 0.0226 1.1666 1.0801
No log 5.6279 242 1.1333 0.0175 1.1333 1.0646
No log 5.6744 244 0.9856 -0.1471 0.9856 0.9928
No log 5.7209 246 0.9423 -0.2302 0.9423 0.9707
No log 5.7674 248 0.9535 -0.0949 0.9535 0.9764
No log 5.8140 250 1.0311 -0.0606 1.0311 1.0154
No log 5.8605 252 0.9756 -0.0473 0.9756 0.9877
No log 5.9070 254 0.9434 -0.0904 0.9434 0.9713
No log 5.9535 256 0.9000 -0.1956 0.9000 0.9487
No log 6.0 258 0.9053 -0.1183 0.9053 0.9515
No log 6.0465 260 0.9103 -0.0117 0.9103 0.9541
No log 6.0930 262 0.9285 -0.1885 0.9285 0.9636
No log 6.1395 264 1.0147 0.0753 1.0147 1.0073
No log 6.1860 266 1.0594 0.0758 1.0594 1.0293
No log 6.2326 268 0.9534 -0.1466 0.9534 0.9764
No log 6.2791 270 0.9543 -0.0582 0.9543 0.9769
No log 6.3256 272 0.8997 -0.1128 0.8997 0.9485
No log 6.3721 274 0.9094 -0.0543 0.9094 0.9536
No log 6.4186 276 0.9114 -0.0528 0.9114 0.9547
No log 6.4651 278 0.8466 -0.2123 0.8466 0.9201
No log 6.5116 280 0.8173 -0.1396 0.8173 0.9040
No log 6.5581 282 0.8275 -0.1653 0.8275 0.9097
No log 6.6047 284 0.8493 -0.1659 0.8493 0.9216
No log 6.6512 286 0.8457 -0.0907 0.8457 0.9196
No log 6.6977 288 0.8721 -0.1214 0.8721 0.9338
No log 6.7442 290 0.9426 -0.1637 0.9426 0.9709
No log 6.7907 292 0.9070 -0.0620 0.9070 0.9524
No log 6.8372 294 0.9455 0.0084 0.9455 0.9724
No log 6.8837 296 0.9707 0.0084 0.9707 0.9853
No log 6.9302 298 0.9265 -0.1116 0.9265 0.9626
No log 6.9767 300 0.9053 -0.0426 0.9053 0.9515
No log 7.0233 302 0.9877 0.0986 0.9877 0.9938
No log 7.0698 304 0.9833 0.0587 0.9833 0.9916
No log 7.1163 306 0.8189 0.0257 0.8189 0.9050
No log 7.1628 308 0.7793 -0.1332 0.7793 0.8828
No log 7.2093 310 0.7662 -0.0870 0.7662 0.8753
No log 7.2558 312 0.7970 -0.1266 0.7970 0.8927
No log 7.3023 314 0.8459 -0.0941 0.8459 0.9198
No log 7.3488 316 0.9187 0.1116 0.9187 0.9585
No log 7.3953 318 0.8377 -0.0389 0.8377 0.9153
No log 7.4419 320 0.8704 -0.0049 0.8704 0.9330
No log 7.4884 322 0.9381 -0.0159 0.9381 0.9686
No log 7.5349 324 0.9646 -0.0128 0.9646 0.9821
No log 7.5814 326 0.9576 -0.0082 0.9576 0.9786
No log 7.6279 328 0.8843 0.0347 0.8843 0.9404
No log 7.6744 330 0.8580 0.0361 0.8580 0.9263
No log 7.7209 332 0.8487 0.0474 0.8487 0.9212
No log 7.7674 334 0.8616 -0.1472 0.8616 0.9282
No log 7.8140 336 0.8825 -0.1412 0.8825 0.9394
No log 7.8605 338 0.9501 0.0129 0.9501 0.9748
No log 7.9070 340 0.8692 -0.1472 0.8692 0.9323
No log 7.9535 342 0.8588 -0.1470 0.8588 0.9267
No log 8.0 344 0.9360 -0.0595 0.9360 0.9675
No log 8.0465 346 0.9167 -0.0015 0.9167 0.9575
No log 8.0930 348 0.8419 -0.0851 0.8419 0.9175
No log 8.1395 350 0.8330 -0.0054 0.8330 0.9127
No log 8.1860 352 0.8218 -0.0082 0.8218 0.9065
No log 8.2326 354 0.8484 -0.1592 0.8484 0.9211
No log 8.2791 356 0.8823 -0.1045 0.8823 0.9393
No log 8.3256 358 1.0452 -0.0253 1.0452 1.0224
No log 8.3721 360 0.9484 -0.0015 0.9484 0.9739
No log 8.4186 362 0.9043 -0.0711 0.9043 0.9509
No log 8.4651 364 0.9556 0.0922 0.9556 0.9776
No log 8.5116 366 1.1528 0.0175 1.1528 1.0737
No log 8.5581 368 1.0316 -0.0253 1.0316 1.0157
No log 8.6047 370 0.8448 -0.0860 0.8448 0.9191
No log 8.6512 372 0.8249 0.0549 0.8249 0.9083
No log 8.6977 374 0.7747 0.0776 0.7747 0.8802
No log 8.7442 376 0.7954 -0.1753 0.7954 0.8918
No log 8.7907 378 0.8594 -0.0408 0.8594 0.9270
No log 8.8372 380 0.8243 -0.1722 0.8243 0.9079
No log 8.8837 382 0.8642 -0.1643 0.8642 0.9296
No log 8.9302 384 0.9132 -0.1302 0.9132 0.9556
No log 8.9767 386 1.0015 0.0142 1.0015 1.0007
No log 9.0233 388 0.9231 -0.0829 0.9231 0.9608
No log 9.0698 390 0.8754 -0.1222 0.8754 0.9356
No log 9.1163 392 0.9169 -0.0373 0.9169 0.9575
No log 9.1628 394 0.8580 0.0574 0.8580 0.9263
No log 9.2093 396 0.8534 -0.1979 0.8534 0.9238
No log 9.2558 398 0.8550 -0.1979 0.8550 0.9247
No log 9.3023 400 0.8054 -0.0889 0.8054 0.8974
No log 9.3488 402 0.8417 0.0909 0.8417 0.9174
No log 9.3953 404 0.8462 0.0909 0.8462 0.9199
No log 9.4419 406 0.8028 -0.1795 0.8028 0.8960
No log 9.4884 408 0.8203 -0.2108 0.8203 0.9057
No log 9.5349 410 0.8385 -0.0156 0.8385 0.9157
No log 9.5814 412 0.8514 -0.0178 0.8514 0.9227
No log 9.6279 414 0.8413 -0.1648 0.8413 0.9172
No log 9.6744 416 0.8359 -0.1592 0.8359 0.9143
No log 9.7209 418 0.8227 -0.1333 0.8227 0.9070
No log 9.7674 420 0.8082 -0.0170 0.8082 0.8990
No log 9.8140 422 0.7830 -0.1094 0.7830 0.8849
No log 9.8605 424 0.7759 -0.1010 0.7759 0.8809
No log 9.9070 426 0.7846 -0.1010 0.7846 0.8858
No log 9.9535 428 0.7975 -0.0532 0.7975 0.8930
No log 10.0 430 0.8339 0.0650 0.8339 0.9132
No log 10.0465 432 0.8837 0.0424 0.8837 0.9401
No log 10.0930 434 0.8264 0.0295 0.8264 0.9091
No log 10.1395 436 0.9324 0.0092 0.9324 0.9656
No log 10.1860 438 0.9468 0.0470 0.9468 0.9730
No log 10.2326 440 0.8144 -0.1967 0.8144 0.9025
No log 10.2791 442 0.8274 0.0438 0.8274 0.9096
No log 10.3256 444 0.8712 0.0676 0.8712 0.9334
No log 10.3721 446 0.7817 0.0289 0.7817 0.8842
No log 10.4186 448 0.7930 -0.1397 0.7930 0.8905
No log 10.4651 450 0.8451 -0.0551 0.8451 0.9193
No log 10.5116 452 0.8409 -0.0989 0.8409 0.9170
No log 10.5581 454 0.8100 -0.0138 0.8100 0.9000
No log 10.6047 456 0.9069 -0.0031 0.9069 0.9523
No log 10.6512 458 0.8850 0.0068 0.8850 0.9408
No log 10.6977 460 0.8498 -0.1956 0.8498 0.9218
No log 10.7442 462 0.8745 -0.0852 0.8745 0.9351
No log 10.7907 464 0.8074 -0.0462 0.8074 0.8985
No log 10.8372 466 0.7900 0.0670 0.7900 0.8888
No log 10.8837 468 0.7945 0.0999 0.7945 0.8914
No log 10.9302 470 0.7469 0.0628 0.7469 0.8642
No log 10.9767 472 0.7252 -0.0065 0.7252 0.8516
No log 11.0233 474 0.7810 -0.0591 0.7810 0.8838
No log 11.0698 476 0.7655 -0.2224 0.7655 0.8749
No log 11.1163 478 0.7909 -0.2056 0.7909 0.8893
No log 11.1628 480 0.7824 -0.2123 0.7824 0.8845
No log 11.2093 482 0.7945 -0.2154 0.7945 0.8914
No log 11.2558 484 0.7602 -0.0902 0.7602 0.8719
No log 11.3023 486 0.7792 -0.2154 0.7792 0.8827
No log 11.3488 488 0.7982 -0.2138 0.7982 0.8934
No log 11.3953 490 0.7783 -0.0849 0.7783 0.8822
No log 11.4419 492 0.7654 0.0357 0.7654 0.8749
No log 11.4884 494 0.7730 -0.0849 0.7730 0.8792
No log 11.5349 496 0.7748 -0.0849 0.7748 0.8802
No log 11.5814 498 0.7996 -0.1601 0.7996 0.8942
0.3255 11.6279 500 0.7814 -0.0984 0.7814 0.8840
0.3255 11.6744 502 0.7855 -0.1040 0.7855 0.8863
0.3255 11.7209 504 0.8157 -0.0947 0.8157 0.9032
0.3255 11.7674 506 0.8455 -0.0517 0.8455 0.9195
0.3255 11.8140 508 0.8660 -0.0533 0.8660 0.9306
0.3255 11.8605 510 0.8914 -0.2837 0.8914 0.9442
0.3255 11.9070 512 0.9067 -0.1697 0.9067 0.9522
0.3255 11.9535 514 0.8685 -0.1219 0.8685 0.9319
0.3255 12.0 516 0.8576 -0.0806 0.8576 0.9261
0.3255 12.0465 518 0.8524 -0.1526 0.8524 0.9233
0.3255 12.0930 520 0.8224 -0.1217 0.8224 0.9069
0.3255 12.1395 522 0.8109 -0.1653 0.8109 0.9005
0.3255 12.1860 524 0.8346 -0.1039 0.8346 0.9136
0.3255 12.2326 526 0.8071 -0.2030 0.8071 0.8984

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
9
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k16_task3_organization

Finetuned
(4222)
this model