ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k3_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9513
  • Qwk: 0.1396
  • Mse: 0.9513
  • Rmse: 0.9753

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1818 2 3.4939 -0.0047 3.4939 1.8692
No log 0.3636 4 1.8714 0.0943 1.8714 1.3680
No log 0.5455 6 2.1638 0.0229 2.1638 1.4710
No log 0.7273 8 1.5643 0.0799 1.5643 1.2507
No log 0.9091 10 1.0555 -0.2432 1.0555 1.0274
No log 1.0909 12 0.8411 -0.1233 0.8411 0.9171
No log 1.2727 14 0.7997 -0.0695 0.7997 0.8943
No log 1.4545 16 0.8166 -0.0240 0.8166 0.9036
No log 1.6364 18 0.9232 -0.0218 0.9232 0.9608
No log 1.8182 20 1.2420 0.0 1.2420 1.1145
No log 2.0 22 1.4854 0.0425 1.4854 1.2188
No log 2.1818 24 1.4237 0.0425 1.4237 1.1932
No log 2.3636 26 1.1898 0.0 1.1898 1.0908
No log 2.5455 28 0.9591 -0.0648 0.9591 0.9793
No log 2.7273 30 0.8541 0.0409 0.8541 0.9242
No log 2.9091 32 0.8223 -0.0790 0.8223 0.9068
No log 3.0909 34 0.8434 -0.1244 0.8434 0.9184
No log 3.2727 36 1.2476 0.0022 1.2476 1.1170
No log 3.4545 38 2.2842 0.0369 2.2842 1.5113
No log 3.6364 40 2.4973 -0.0168 2.4973 1.5803
No log 3.8182 42 1.5182 -0.0199 1.5182 1.2322
No log 4.0 44 0.7928 -0.1233 0.7928 0.8904
No log 4.1818 46 0.7167 -0.0069 0.7167 0.8466
No log 4.3636 48 0.7486 -0.1765 0.7486 0.8652
No log 4.5455 50 0.8937 0.0111 0.8937 0.9454
No log 4.7273 52 1.0836 0.0279 1.0836 1.0410
No log 4.9091 54 1.2995 0.0 1.2995 1.1400
No log 5.0909 56 1.1691 0.0317 1.1691 1.0813
No log 5.2727 58 0.8939 -0.0886 0.8939 0.9455
No log 5.4545 60 0.9205 -0.1270 0.9205 0.9594
No log 5.6364 62 0.9142 -0.1270 0.9142 0.9561
No log 5.8182 64 0.8945 -0.0878 0.8945 0.9458
No log 6.0 66 0.7288 0.0506 0.7288 0.8537
No log 6.1818 68 0.7340 0.0506 0.7340 0.8567
No log 6.3636 70 1.0509 -0.0122 1.0509 1.0251
No log 6.5455 72 1.8272 0.0327 1.8272 1.3517
No log 6.7273 74 1.4334 -0.0656 1.4334 1.1973
No log 6.9091 76 0.7844 0.1148 0.7844 0.8857
No log 7.0909 78 0.7334 0.0460 0.7334 0.8564
No log 7.2727 80 0.7384 0.0964 0.7384 0.8593
No log 7.4545 82 0.8568 0.0909 0.8568 0.9257
No log 7.6364 84 1.2439 -0.0334 1.2439 1.1153
No log 7.8182 86 1.0075 0.0175 1.0075 1.0037
No log 8.0 88 0.7965 0.1506 0.7965 0.8925
No log 8.1818 90 0.7465 0.1254 0.7465 0.8640
No log 8.3636 92 0.9853 0.0986 0.9853 0.9926
No log 8.5455 94 1.7284 0.0839 1.7284 1.3147
No log 8.7273 96 1.3406 0.0671 1.3406 1.1579
No log 8.9091 98 0.7797 0.0236 0.7797 0.8830
No log 9.0909 100 0.8303 0.0279 0.8303 0.9112
No log 9.2727 102 0.7967 0.0683 0.7967 0.8926
No log 9.4545 104 0.7150 0.0460 0.7150 0.8456
No log 9.6364 106 1.2258 -0.0351 1.2258 1.1071
No log 9.8182 108 1.9932 0.0694 1.9932 1.4118
No log 10.0 110 1.8935 0.0796 1.8935 1.3761
No log 10.1818 112 1.1256 0.0006 1.1256 1.0609
No log 10.3636 114 0.6903 0.1318 0.6903 0.8308
No log 10.5455 116 0.7286 0.0496 0.7286 0.8536
No log 10.7273 118 0.7175 0.0496 0.7175 0.8470
No log 10.9091 120 0.7128 0.2180 0.7128 0.8443
No log 11.0909 122 1.0053 0.0566 1.0053 1.0027
No log 11.2727 124 1.0685 0.0758 1.0685 1.0337
No log 11.4545 126 0.8311 0.1593 0.8311 0.9116
No log 11.6364 128 0.7081 0.2711 0.7081 0.8415
No log 11.8182 130 0.6944 0.1433 0.6944 0.8333
No log 12.0 132 0.6819 0.1758 0.6819 0.8258
No log 12.1818 134 0.9022 0.0676 0.9022 0.9498
No log 12.3636 136 1.3169 0.0350 1.3169 1.1476
No log 12.5455 138 1.2702 0.0152 1.2702 1.1270
No log 12.7273 140 0.8888 -0.0122 0.8888 0.9428
No log 12.9091 142 0.7417 0.1047 0.7417 0.8612
No log 13.0909 144 0.7018 0.1541 0.7018 0.8378
No log 13.2727 146 0.7837 0.1379 0.7837 0.8853
No log 13.4545 148 1.0590 0.0044 1.0590 1.0291
No log 13.6364 150 1.2980 -0.0119 1.2980 1.1393
No log 13.8182 152 1.0841 -0.0306 1.0841 1.0412
No log 14.0 154 0.7533 0.1095 0.7533 0.8679
No log 14.1818 156 0.7157 0.2239 0.7157 0.8460
No log 14.3636 158 0.7877 0.1144 0.7877 0.8875
No log 14.5455 160 0.9546 0.0293 0.9546 0.9770
No log 14.7273 162 1.0262 0.0953 1.0262 1.0130
No log 14.9091 164 0.8707 0.1964 0.8707 0.9331
No log 15.0909 166 0.8245 0.2053 0.8245 0.9080
No log 15.2727 168 0.7881 0.1192 0.7881 0.8878
No log 15.4545 170 0.7755 0.1192 0.7755 0.8806
No log 15.6364 172 0.7819 0.1659 0.7819 0.8843
No log 15.8182 174 0.7808 0.1585 0.7808 0.8836
No log 16.0 176 0.9299 0.0856 0.9299 0.9643
No log 16.1818 178 1.0776 0.1206 1.0776 1.0381
No log 16.3636 180 1.0137 0.1077 1.0137 1.0068
No log 16.5455 182 0.9547 0.1522 0.9547 0.9771
No log 16.7273 184 0.9300 0.0805 0.9300 0.9644
No log 16.9091 186 0.8564 0.0643 0.8564 0.9254
No log 17.0909 188 0.8382 0.0771 0.8382 0.9155
No log 17.2727 190 0.8811 0.0961 0.8811 0.9387
No log 17.4545 192 1.1526 -0.0937 1.1526 1.0736
No log 17.6364 194 1.2135 -0.0411 1.2135 1.1016
No log 17.8182 196 0.9648 -0.0200 0.9648 0.9822
No log 18.0 198 0.7315 0.2096 0.7315 0.8553
No log 18.1818 200 0.7452 0.0123 0.7452 0.8632
No log 18.3636 202 0.7217 0.0123 0.7217 0.8496
No log 18.5455 204 0.6989 0.2166 0.6989 0.8360
No log 18.7273 206 0.8914 0.1316 0.8914 0.9441
No log 18.9091 208 1.1813 0.0359 1.1813 1.0869
No log 19.0909 210 1.1673 -0.0145 1.1673 1.0804
No log 19.2727 212 0.8747 0.1360 0.8747 0.9353
No log 19.4545 214 0.7309 0.2547 0.7309 0.8549
No log 19.6364 216 0.6996 0.2239 0.6996 0.8364
No log 19.8182 218 0.7096 0.2239 0.7096 0.8424
No log 20.0 220 0.7406 0.2166 0.7406 0.8606
No log 20.1818 222 0.8586 0.2534 0.8586 0.9266
No log 20.3636 224 1.1399 -0.0075 1.1399 1.0676
No log 20.5455 226 1.2514 -0.0380 1.2514 1.1187
No log 20.7273 228 0.9933 0.0065 0.9933 0.9967
No log 20.9091 230 0.7874 0.2277 0.7874 0.8874
No log 21.0909 232 0.7254 0.2466 0.7254 0.8517
No log 21.2727 234 0.7412 0.2679 0.7412 0.8610
No log 21.4545 236 0.7715 0.1775 0.7715 0.8783
No log 21.6364 238 0.7324 0.2679 0.7324 0.8558
No log 21.8182 240 0.7056 0.2112 0.7056 0.8400
No log 22.0 242 0.7223 0.1761 0.7223 0.8499
No log 22.1818 244 0.7272 0.2366 0.7272 0.8528
No log 22.3636 246 0.7875 0.1633 0.7875 0.8874
No log 22.5455 248 0.7904 0.1687 0.7904 0.8891
No log 22.7273 250 0.7425 0.2641 0.7425 0.8617
No log 22.9091 252 0.7313 0.2298 0.7313 0.8552
No log 23.0909 254 0.7256 0.2087 0.7256 0.8518
No log 23.2727 256 0.7272 0.2270 0.7272 0.8528
No log 23.4545 258 0.8213 0.1899 0.8213 0.9062
No log 23.6364 260 1.0161 0.1312 1.0161 1.0080
No log 23.8182 262 1.1601 0.0756 1.1601 1.0771
No log 24.0 264 1.1138 0.0315 1.1138 1.0554
No log 24.1818 266 1.0132 0.0741 1.0132 1.0066
No log 24.3636 268 0.8636 0.2337 0.8636 0.9293
No log 24.5455 270 0.8027 0.2401 0.8027 0.8959
No log 24.7273 272 0.7723 0.2401 0.7723 0.8788
No log 24.9091 274 0.8683 0.0486 0.8683 0.9318
No log 25.0909 276 1.0880 -0.0098 1.0880 1.0431
No log 25.2727 278 1.2624 0.0098 1.2624 1.1236
No log 25.4545 280 1.1237 0.0827 1.1237 1.0601
No log 25.6364 282 0.8927 0.1261 0.8927 0.9448
No log 25.8182 284 0.8218 0.2083 0.8218 0.9066
No log 26.0 286 0.8108 0.1862 0.8108 0.9005
No log 26.1818 288 0.8257 0.1493 0.8257 0.9087
No log 26.3636 290 1.0198 0.0416 1.0198 1.0098
No log 26.5455 292 1.1197 0.0252 1.1197 1.0582
No log 26.7273 294 0.9660 0.0470 0.9660 0.9829
No log 26.9091 296 0.8102 0.2137 0.8102 0.9001
No log 27.0909 298 0.7332 0.2791 0.7332 0.8562
No log 27.2727 300 0.7285 0.2570 0.7285 0.8535
No log 27.4545 302 0.8093 0.1026 0.8093 0.8996
No log 27.6364 304 0.8513 0.0946 0.8513 0.9226
No log 27.8182 306 0.8000 0.1437 0.8000 0.8944
No log 28.0 308 0.8788 0.1150 0.8788 0.9375
No log 28.1818 310 1.0259 -0.0374 1.0259 1.0129
No log 28.3636 312 1.0082 -0.0065 1.0082 1.0041
No log 28.5455 314 0.8584 0.0711 0.8584 0.9265
No log 28.7273 316 0.7655 0.2220 0.7655 0.8749
No log 28.9091 318 0.7579 0.2070 0.7579 0.8705
No log 29.0909 320 0.7694 0.1136 0.7694 0.8772
No log 29.2727 322 0.7953 0.2155 0.7953 0.8918
No log 29.4545 324 0.9369 0.0949 0.9369 0.9680
No log 29.6364 326 0.9461 0.1316 0.9461 0.9727
No log 29.8182 328 0.9439 0.1316 0.9439 0.9715
No log 30.0 330 0.9586 0.1148 0.9586 0.9791
No log 30.1818 332 0.9270 0.1930 0.9270 0.9628
No log 30.3636 334 0.8385 0.2070 0.8385 0.9157
No log 30.5455 336 0.7990 0.1633 0.7990 0.8939
No log 30.7273 338 0.7954 0.1646 0.7954 0.8919
No log 30.9091 340 0.7782 0.1646 0.7782 0.8821
No log 31.0909 342 0.7259 0.1800 0.7259 0.8520
No log 31.2727 344 0.7629 0.2032 0.7629 0.8734
No log 31.4545 346 0.7724 0.2077 0.7724 0.8788
No log 31.6364 348 0.8207 0.2834 0.8207 0.9059
No log 31.8182 350 0.8752 0.2958 0.8752 0.9355
No log 32.0 352 1.0010 0.1422 1.0010 1.0005
No log 32.1818 354 1.1253 0.1412 1.1253 1.0608
No log 32.3636 356 1.1509 0.1754 1.1509 1.0728
No log 32.5455 358 1.0290 0.2490 1.0290 1.0144
No log 32.7273 360 1.0163 0.2670 1.0163 1.0081
No log 32.9091 362 0.9645 0.2518 0.9645 0.9821
No log 33.0909 364 0.9644 0.2266 0.9644 0.9820
No log 33.2727 366 1.0478 0.1714 1.0478 1.0236
No log 33.4545 368 1.0313 0.1714 1.0313 1.0155
No log 33.6364 370 0.9425 0.2370 0.9425 0.9708
No log 33.8182 372 0.8483 0.2852 0.8483 0.9210
No log 34.0 374 0.8338 0.2852 0.8338 0.9131
No log 34.1818 376 0.8906 0.2790 0.8906 0.9437
No log 34.3636 378 0.9544 0.1525 0.9544 0.9769
No log 34.5455 380 0.9681 0.1402 0.9681 0.9839
No log 34.7273 382 0.9258 0.1442 0.9258 0.9622
No log 34.9091 384 0.7637 0.1758 0.7637 0.8739
No log 35.0909 386 0.7066 0.2053 0.7066 0.8406
No log 35.2727 388 0.7077 0.2431 0.7077 0.8412
No log 35.4545 390 0.7787 0.1758 0.7787 0.8825
No log 35.6364 392 0.9671 0.0089 0.9671 0.9834
No log 35.8182 394 1.0698 -0.0376 1.0698 1.0343
No log 36.0 396 1.0661 -0.0374 1.0661 1.0325
No log 36.1818 398 1.0078 -0.0359 1.0078 1.0039
No log 36.3636 400 0.9336 0.0089 0.9336 0.9662
No log 36.5455 402 0.8500 0.0711 0.8500 0.9220
No log 36.7273 404 0.7762 0.1879 0.7762 0.8810
No log 36.9091 406 0.7427 0.2078 0.7427 0.8618
No log 37.0909 408 0.7367 0.2150 0.7367 0.8583
No log 37.2727 410 0.7771 0.2424 0.7771 0.8815
No log 37.4545 412 0.8198 0.0786 0.8198 0.9054
No log 37.6364 414 0.9038 0.0111 0.9038 0.9507
No log 37.8182 416 0.9435 0.0741 0.9435 0.9714
No log 38.0 418 0.9494 -0.0269 0.9494 0.9744
No log 38.1818 420 0.9132 -0.0218 0.9132 0.9556
No log 38.3636 422 0.8145 0.1758 0.8145 0.9025
No log 38.5455 424 0.7756 0.1817 0.7756 0.8807
No log 38.7273 426 0.7953 0.1605 0.7953 0.8918
No log 38.9091 428 0.8706 -0.0200 0.8706 0.9331
No log 39.0909 430 0.9543 -0.0301 0.9543 0.9769
No log 39.2727 432 0.9532 -0.0638 0.9532 0.9763
No log 39.4545 434 0.9145 -0.0269 0.9145 0.9563
No log 39.6364 436 0.7863 0.2424 0.7863 0.8868
No log 39.8182 438 0.7052 0.1943 0.7052 0.8398
No log 40.0 440 0.6918 0.2005 0.6918 0.8317
No log 40.1818 442 0.7085 0.1646 0.7085 0.8417
No log 40.3636 444 0.7089 0.1942 0.7089 0.8419
No log 40.5455 446 0.7116 0.2431 0.7116 0.8436
No log 40.7273 448 0.7913 0.2181 0.7913 0.8896
No log 40.9091 450 0.9040 0.1360 0.9040 0.9508
No log 41.0909 452 1.1023 0.0224 1.1023 1.0499
No log 41.2727 454 1.2453 0.0599 1.2453 1.1159
No log 41.4545 456 1.1566 0.0400 1.1566 1.0755
No log 41.6364 458 0.9221 0.0873 0.9221 0.9602
No log 41.8182 460 0.7400 0.2181 0.7400 0.8602
No log 42.0 462 0.6911 0.2009 0.6911 0.8314
No log 42.1818 464 0.6833 0.2078 0.6833 0.8266
No log 42.3636 466 0.6879 0.2507 0.6879 0.8294
No log 42.5455 468 0.7140 0.2009 0.7140 0.8450
No log 42.7273 470 0.7773 0.2277 0.7773 0.8817
No log 42.9091 472 0.8692 0.1024 0.8692 0.9323
No log 43.0909 474 0.9241 0.0946 0.9241 0.9613
No log 43.2727 476 0.9605 0.0134 0.9605 0.9801
No log 43.4545 478 0.8840 0.0684 0.8840 0.9402
No log 43.6364 480 0.8251 0.1193 0.8251 0.9083
No log 43.8182 482 0.7780 0.1286 0.7780 0.8820
No log 44.0 484 0.7815 0.1660 0.7815 0.8840
No log 44.1818 486 0.7962 0.1553 0.7962 0.8923
No log 44.3636 488 0.8026 0.1453 0.8026 0.8959
No log 44.5455 490 0.7686 0.1716 0.7686 0.8767
No log 44.7273 492 0.7480 0.2248 0.7480 0.8649
No log 44.9091 494 0.7445 0.2248 0.7445 0.8628
No log 45.0909 496 0.7417 0.2318 0.7417 0.8612
No log 45.2727 498 0.7797 0.2054 0.7797 0.8830
0.2848 45.4545 500 0.8401 0.1406 0.8401 0.9166
0.2848 45.6364 502 0.8373 0.1406 0.8373 0.9151
0.2848 45.8182 504 0.8231 0.2357 0.8231 0.9072
0.2848 46.0 506 0.7922 0.2424 0.7922 0.8901
0.2848 46.1818 508 0.7659 0.2566 0.7659 0.8751
0.2848 46.3636 510 0.7962 0.2424 0.7962 0.8923
0.2848 46.5455 512 0.7843 0.2424 0.7843 0.8856
0.2848 46.7273 514 0.7790 0.1758 0.7790 0.8826
0.2848 46.9091 516 0.7671 0.2220 0.7671 0.8758
0.2848 47.0909 518 0.7767 0.2466 0.7767 0.8813
0.2848 47.2727 520 0.8436 0.1809 0.8436 0.9185
0.2848 47.4545 522 0.9765 0.0492 0.9765 0.9882
0.2848 47.6364 524 0.9896 0.0142 0.9896 0.9948
0.2848 47.8182 526 0.8622 0.1744 0.8622 0.9285
0.2848 48.0 528 0.7736 0.1636 0.7736 0.8796
0.2848 48.1818 530 0.7378 0.1904 0.7378 0.8590
0.2848 48.3636 532 0.7224 0.1179 0.7224 0.8499
0.2848 48.5455 534 0.6992 0.1630 0.6992 0.8362
0.2848 48.7273 536 0.6841 0.2096 0.6841 0.8271
0.2848 48.9091 538 0.6888 0.2009 0.6888 0.8299
0.2848 49.0909 540 0.7063 0.2502 0.7063 0.8404
0.2848 49.2727 542 0.7273 0.2424 0.7273 0.8528
0.2848 49.4545 544 0.7460 0.2288 0.7460 0.8637
0.2848 49.6364 546 0.7867 0.1286 0.7867 0.8870
0.2848 49.8182 548 0.8634 0.1879 0.8634 0.9292
0.2848 50.0 550 0.9224 0.0847 0.9224 0.9604
0.2848 50.1818 552 1.0039 0.0619 1.0039 1.0019
0.2848 50.3636 554 1.0447 0.0245 1.0447 1.0221
0.2848 50.5455 556 1.0304 0.0245 1.0304 1.0151
0.2848 50.7273 558 0.9513 0.1396 0.9513 0.9753

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
8
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k3_task3_organization

Finetuned
(4222)
this model