ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k15_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7581
  • Qwk: 0.0410
  • Mse: 0.7581
  • Rmse: 0.8707

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0488 2 3.7788 -0.0063 3.7788 1.9439
No log 0.0976 4 1.9322 0.0284 1.9322 1.3900
No log 0.1463 6 1.5523 0.0014 1.5523 1.2459
No log 0.1951 8 0.9364 -0.0504 0.9364 0.9677
No log 0.2439 10 0.9237 -0.0909 0.9237 0.9611
No log 0.2927 12 1.0702 -0.0411 1.0702 1.0345
No log 0.3415 14 1.4393 -0.0468 1.4393 1.1997
No log 0.3902 16 1.0720 -0.0359 1.0720 1.0354
No log 0.4390 18 0.7117 0.0 0.7117 0.8436
No log 0.4878 20 0.7774 -0.1236 0.7774 0.8817
No log 0.5366 22 1.1853 -0.0164 1.1853 1.0887
No log 0.5854 24 1.4004 -0.0510 1.4004 1.1834
No log 0.6341 26 1.1134 0.0298 1.1134 1.0552
No log 0.6829 28 0.9690 0.0065 0.9690 0.9844
No log 0.7317 30 0.8954 0.0137 0.8954 0.9462
No log 0.7805 32 0.7875 -0.0753 0.7875 0.8874
No log 0.8293 34 0.8254 -0.0008 0.8254 0.9085
No log 0.8780 36 0.7388 0.0374 0.7388 0.8595
No log 0.9268 38 0.6972 0.0964 0.6972 0.8350
No log 0.9756 40 0.7988 0.0409 0.7988 0.8938
No log 1.0244 42 0.9204 0.0329 0.9204 0.9593
No log 1.0732 44 0.7322 0.0909 0.7322 0.8557
No log 1.1220 46 0.7018 0.0555 0.7018 0.8378
No log 1.1707 48 0.8361 -0.0101 0.8361 0.9144
No log 1.2195 50 1.2522 0.0156 1.2522 1.1190
No log 1.2683 52 1.1400 -0.0101 1.1400 1.0677
No log 1.3171 54 0.8884 -0.0532 0.8884 0.9426
No log 1.3659 56 0.8869 -0.0163 0.8869 0.9418
No log 1.4146 58 1.0452 -0.0345 1.0452 1.0224
No log 1.4634 60 0.8780 -0.0033 0.8780 0.9370
No log 1.5122 62 0.8498 -0.0008 0.8498 0.9219
No log 1.5610 64 0.8259 -0.0790 0.8259 0.9088
No log 1.6098 66 0.9979 0.0111 0.9979 0.9990
No log 1.6585 68 0.8131 -0.0790 0.8131 0.9017
No log 1.7073 70 0.7636 -0.0766 0.7636 0.8738
No log 1.7561 72 0.7538 -0.0739 0.7538 0.8682
No log 1.8049 74 0.7140 -0.0035 0.7140 0.8450
No log 1.8537 76 0.7291 -0.0035 0.7291 0.8539
No log 1.9024 78 0.7370 0.0 0.7370 0.8585
No log 1.9512 80 0.7577 0.0 0.7577 0.8705
No log 2.0 82 0.7759 -0.0644 0.7759 0.8809
No log 2.0488 84 0.9575 -0.0425 0.9575 0.9785
No log 2.0976 86 0.9829 -0.0878 0.9829 0.9914
No log 2.1463 88 0.8344 -0.0679 0.8344 0.9134
No log 2.1951 90 0.8379 0.0068 0.8379 0.9154
No log 2.2439 92 0.8594 0.0064 0.8594 0.9270
No log 2.2927 94 0.8592 -0.0881 0.8592 0.9269
No log 2.3415 96 0.8377 0.1199 0.8377 0.9153
No log 2.3902 98 0.8064 -0.0131 0.8064 0.8980
No log 2.4390 100 0.8472 -0.1241 0.8472 0.9204
No log 2.4878 102 0.8176 -0.0499 0.8176 0.9042
No log 2.5366 104 0.8629 0.0099 0.8629 0.9289
No log 2.5854 106 0.8482 0.0068 0.8482 0.9210
No log 2.6341 108 0.8049 -0.0551 0.8049 0.8972
No log 2.6829 110 0.7819 -0.0035 0.7819 0.8842
No log 2.7317 112 0.7993 -0.0035 0.7993 0.8940
No log 2.7805 114 0.8360 -0.0451 0.8360 0.9143
No log 2.8293 116 0.8700 0.0155 0.8700 0.9327
No log 2.8780 118 0.8539 0.0 0.8539 0.9241
No log 2.9268 120 0.8905 -0.0204 0.8905 0.9437
No log 2.9756 122 0.8743 0.0528 0.8743 0.9351
No log 3.0244 124 0.9082 -0.1151 0.9082 0.9530
No log 3.0732 126 0.8928 -0.0762 0.8928 0.9449
No log 3.1220 128 0.8690 0.0680 0.8690 0.9322
No log 3.1707 130 0.8390 -0.0351 0.8390 0.9160
No log 3.2195 132 0.7852 -0.0571 0.7852 0.8861
No log 3.2683 134 0.7793 -0.0551 0.7793 0.8828
No log 3.3171 136 0.7800 -0.0571 0.7800 0.8832
No log 3.3659 138 0.7942 -0.0551 0.7942 0.8912
No log 3.4146 140 0.8002 -0.1163 0.8002 0.8945
No log 3.4634 142 0.8459 0.0759 0.8459 0.9198
No log 3.5122 144 0.8262 -0.1074 0.8262 0.9090
No log 3.5610 146 0.8898 -0.0724 0.8898 0.9433
No log 3.6098 148 0.8387 -0.1074 0.8387 0.9158
No log 3.6585 150 0.8148 -0.1153 0.8148 0.9027
No log 3.7073 152 0.8606 -0.0240 0.8606 0.9277
No log 3.7561 154 0.7891 -0.0662 0.7891 0.8883
No log 3.8049 156 0.8207 -0.0551 0.8207 0.9059
No log 3.8537 158 0.8452 -0.0571 0.8452 0.9193
No log 3.9024 160 0.7948 -0.1153 0.7948 0.8915
No log 3.9512 162 0.7670 -0.0612 0.7670 0.8758
No log 4.0 164 0.7571 -0.0550 0.7571 0.8701
No log 4.0488 166 0.7592 -0.0984 0.7592 0.8713
No log 4.0976 168 0.9513 0.0609 0.9513 0.9754
No log 4.1463 170 0.9431 -0.0056 0.9431 0.9712
No log 4.1951 172 0.8394 0.0025 0.8394 0.9162
No log 4.2439 174 0.8722 0.1032 0.8722 0.9339
No log 4.2927 176 0.9175 0.0570 0.9175 0.9579
No log 4.3415 178 0.9764 -0.0166 0.9764 0.9881
No log 4.3902 180 1.0060 -0.0076 1.0060 1.0030
No log 4.4390 182 1.0104 -0.0079 1.0104 1.0052
No log 4.4878 184 0.9656 -0.0204 0.9656 0.9827
No log 4.5366 186 0.9307 0.0173 0.9307 0.9647
No log 4.5854 188 0.8992 -0.0612 0.8992 0.9482
No log 4.6341 190 0.8643 -0.0118 0.8643 0.9297
No log 4.6829 192 0.8937 0.1029 0.8937 0.9453
No log 4.7317 194 0.9807 -0.0133 0.9807 0.9903
No log 4.7805 196 0.9343 0.0570 0.9343 0.9666
No log 4.8293 198 0.9094 0.0987 0.9094 0.9536
No log 4.8780 200 0.9345 -0.0111 0.9345 0.9667
No log 4.9268 202 0.9436 -0.0393 0.9436 0.9714
No log 4.9756 204 0.8823 0.0110 0.8823 0.9393
No log 5.0244 206 0.8868 -0.0889 0.8868 0.9417
No log 5.0732 208 0.8876 0.0240 0.8876 0.9421
No log 5.1220 210 0.8560 0.0187 0.8560 0.9252
No log 5.1707 212 0.8095 0.0622 0.8095 0.8997
No log 5.2195 214 0.8002 0.0622 0.8002 0.8945
No log 5.2683 216 0.8292 0.0724 0.8292 0.9106
No log 5.3171 218 0.7797 0.0229 0.7797 0.8830
No log 5.3659 220 0.8568 0.0310 0.8568 0.9256
No log 5.4146 222 1.0337 -0.0436 1.0337 1.0167
No log 5.4634 224 0.9879 -0.0616 0.9879 0.9939
No log 5.5122 226 0.9827 -0.0295 0.9827 0.9913
No log 5.5610 228 0.8574 0.0488 0.8574 0.9259
No log 5.6098 230 0.8373 0.0940 0.8373 0.9150
No log 5.6585 232 0.9224 0.0406 0.9224 0.9604
No log 5.7073 234 1.0217 -0.0464 1.0217 1.0108
No log 5.7561 236 0.8463 0.1425 0.8463 0.9200
No log 5.8049 238 0.8356 0.0444 0.8356 0.9141
No log 5.8537 240 0.9375 -0.0887 0.9375 0.9683
No log 5.9024 242 0.9474 -0.0528 0.9474 0.9733
No log 5.9512 244 0.8862 0.0861 0.8862 0.9414
No log 6.0 246 0.8579 0.0856 0.8579 0.9262
No log 6.0488 248 0.9409 -0.1530 0.9409 0.9700
No log 6.0976 250 0.9524 -0.1232 0.9524 0.9759
No log 6.1463 252 0.7884 0.1470 0.7884 0.8879
No log 6.1951 254 0.8531 0.0071 0.8531 0.9236
No log 6.2439 256 0.8199 0.0129 0.8199 0.9055
No log 6.2927 258 0.7764 0.1133 0.7764 0.8811
No log 6.3415 260 0.9035 -0.0040 0.9035 0.9505
No log 6.3902 262 0.9283 -0.0633 0.9283 0.9635
No log 6.4390 264 0.9087 -0.0690 0.9087 0.9532
No log 6.4878 266 1.0765 -0.1228 1.0765 1.0375
No log 6.5366 268 0.9346 -0.0778 0.9346 0.9668
No log 6.5854 270 0.8583 -0.0892 0.8583 0.9265
No log 6.6341 272 0.9072 -0.0181 0.9072 0.9524
No log 6.6829 274 0.8241 -0.0322 0.8241 0.9078
No log 6.7317 276 0.8967 -0.0801 0.8967 0.9469
No log 6.7805 278 1.0223 -0.0504 1.0223 1.0111
No log 6.8293 280 0.8376 -0.0204 0.8376 0.9152
No log 6.8780 282 0.7933 0.0148 0.7933 0.8907
No log 6.9268 284 0.9036 -0.0377 0.9036 0.9506
No log 6.9756 286 0.8949 -0.0377 0.8949 0.9460
No log 7.0244 288 0.8087 0.1080 0.8087 0.8993
No log 7.0732 290 0.8317 -0.0483 0.8317 0.9120
No log 7.1220 292 0.8361 0.0026 0.8361 0.9144
No log 7.1707 294 0.8588 0.0249 0.8588 0.9267
No log 7.2195 296 0.8817 -0.0138 0.8817 0.9390
No log 7.2683 298 0.8594 -0.0116 0.8594 0.9270
No log 7.3171 300 0.7953 0.1080 0.7953 0.8918
No log 7.3659 302 0.8085 -0.0599 0.8085 0.8992
No log 7.4146 304 0.7900 -0.1054 0.7900 0.8888
No log 7.4634 306 0.8200 -0.0999 0.8200 0.9055
No log 7.5122 308 0.8510 -0.0373 0.8510 0.9225
No log 7.5610 310 0.8799 -0.0543 0.8799 0.9380
No log 7.6098 312 0.8426 0.0249 0.8426 0.9179
No log 7.6585 314 0.8271 0.0328 0.8271 0.9094
No log 7.7073 316 0.7639 0.1080 0.7639 0.8740
No log 7.7561 318 0.7609 0.0557 0.7609 0.8723
No log 7.8049 320 0.7741 -0.0108 0.7741 0.8798
No log 7.8537 322 0.7862 0.1080 0.7862 0.8867
No log 7.9024 324 0.8512 -0.0513 0.8512 0.9226
No log 7.9512 326 0.8349 0.0291 0.8349 0.9138
No log 8.0 328 0.7852 0.0338 0.7852 0.8861
No log 8.0488 330 0.8395 -0.0295 0.8395 0.9162
No log 8.0976 332 0.7847 0.0395 0.7847 0.8858
No log 8.1463 334 0.7910 -0.0293 0.7910 0.8894
No log 8.1951 336 0.7857 -0.0293 0.7857 0.8864
No log 8.2439 338 0.7596 -0.0307 0.7596 0.8716
No log 8.2927 340 0.7310 0.0436 0.7310 0.8550
No log 8.3415 342 0.7677 0.0282 0.7677 0.8762
No log 8.3902 344 0.7357 0.0914 0.7357 0.8577
No log 8.4390 346 0.7425 0.1080 0.7425 0.8617
No log 8.4878 348 0.7421 0.1080 0.7421 0.8614
No log 8.5366 350 0.7506 0.1537 0.7506 0.8664
No log 8.5854 352 0.7913 0.1079 0.7913 0.8896
No log 8.6341 354 0.8546 0.0331 0.8546 0.9245
No log 8.6829 356 0.8331 0.0303 0.8331 0.9128
No log 8.7317 358 0.7585 0.1327 0.7585 0.8709
No log 8.7805 360 0.7631 0.0303 0.7631 0.8736
No log 8.8293 362 0.7766 0.0282 0.7766 0.8812
No log 8.8780 364 0.7686 0.0513 0.7686 0.8767
No log 8.9268 366 0.8754 0.1948 0.8754 0.9356
No log 8.9756 368 1.0043 0.0558 1.0043 1.0022
No log 9.0244 370 0.9277 0.0470 0.9277 0.9632
No log 9.0732 372 0.7308 0.0465 0.7308 0.8549
No log 9.1220 374 0.7632 0.0099 0.7632 0.8736
No log 9.1707 376 0.7115 -0.0228 0.7115 0.8435
No log 9.2195 378 0.7051 0.0571 0.7051 0.8397
No log 9.2683 380 0.7405 0.0695 0.7405 0.8605
No log 9.3171 382 0.7541 0.0218 0.7541 0.8684
No log 9.3659 384 0.8480 0.0770 0.8480 0.9209
No log 9.4146 386 0.8090 0.0702 0.8090 0.8994
No log 9.4634 388 0.8139 0.0664 0.8139 0.9022
No log 9.5122 390 0.8237 0.0664 0.8237 0.9076
No log 9.5610 392 0.7673 -0.0054 0.7673 0.8760
No log 9.6098 394 0.7797 -0.0218 0.7797 0.8830
No log 9.6585 396 0.7434 -0.0583 0.7434 0.8622
No log 9.7073 398 0.7763 0.1080 0.7763 0.8811
No log 9.7561 400 0.7927 0.1079 0.7927 0.8903
No log 9.8049 402 0.7539 -0.0027 0.7539 0.8683
No log 9.8537 404 0.7736 -0.0138 0.7736 0.8796
No log 9.9024 406 0.7761 -0.0108 0.7761 0.8809
No log 9.9512 408 0.7908 0.1079 0.7908 0.8893
No log 10.0 410 0.8563 0.0734 0.8563 0.9254
No log 10.0488 412 0.7937 0.1126 0.7937 0.8909
No log 10.0976 414 0.7622 0.1254 0.7622 0.8731
No log 10.1463 416 0.7878 0.0680 0.7878 0.8876
No log 10.1951 418 0.7789 0.0680 0.7789 0.8826
No log 10.2439 420 0.7595 0.0926 0.7595 0.8715
No log 10.2927 422 0.8245 -0.0408 0.8245 0.9080
No log 10.3415 424 0.7949 -0.0678 0.7949 0.8916
No log 10.3902 426 0.7267 0.0513 0.7267 0.8525
No log 10.4390 428 0.7478 0.0214 0.7478 0.8647
No log 10.4878 430 0.7292 -0.0658 0.7292 0.8539
No log 10.5366 432 0.7304 0.0571 0.7304 0.8546
No log 10.5854 434 0.7892 0.1185 0.7892 0.8884
No log 10.6341 436 0.8115 0.1182 0.8115 0.9008
No log 10.6829 438 0.8345 0.1176 0.8345 0.9135
No log 10.7317 440 0.9371 0.0164 0.9371 0.9680
No log 10.7805 442 0.9341 0.0164 0.9341 0.9665
No log 10.8293 444 0.8069 0.0654 0.8069 0.8983
No log 10.8780 446 0.7710 0.1031 0.7710 0.8781
No log 10.9268 448 0.7993 -0.0226 0.7993 0.8941
No log 10.9756 450 0.8490 -0.0066 0.8490 0.9214
No log 11.0244 452 0.8639 0.0776 0.8639 0.9295
No log 11.0732 454 0.8191 -0.0093 0.8191 0.9051
No log 11.1220 456 0.7614 0.0089 0.7614 0.8726
No log 11.1707 458 0.7786 0.0663 0.7786 0.8824
No log 11.2195 460 0.8257 0.0345 0.8257 0.9087
No log 11.2683 462 0.9414 0.0164 0.9414 0.9703
No log 11.3171 464 1.0801 0.0610 1.0801 1.0393
No log 11.3659 466 0.9995 -0.0079 0.9995 0.9998
No log 11.4146 468 0.8616 0.0776 0.8616 0.9282
No log 11.4634 470 0.7635 0.0 0.7635 0.8738
No log 11.5122 472 0.7612 0.0 0.7612 0.8725
No log 11.5610 474 0.7639 0.0503 0.7639 0.8740
No log 11.6098 476 0.7692 -0.0387 0.7692 0.8770
No log 11.6585 478 0.7621 -0.0407 0.7621 0.8730
No log 11.7073 480 0.7845 0.0606 0.7845 0.8857
No log 11.7561 482 0.7891 0.0644 0.7891 0.8883
No log 11.8049 484 0.7611 0.0081 0.7611 0.8724
No log 11.8537 486 0.7406 0.0874 0.7406 0.8606
No log 11.9024 488 0.7180 -0.0062 0.7180 0.8473
No log 11.9512 490 0.7383 0.1029 0.7383 0.8593
No log 12.0 492 0.7666 0.0196 0.7666 0.8755
No log 12.0488 494 0.7479 0.1974 0.7479 0.8648
No log 12.0976 496 0.6998 0.0479 0.6998 0.8365
No log 12.1463 498 0.6923 0.0967 0.6923 0.8320
0.3194 12.1951 500 0.6878 0.0967 0.6878 0.8294
0.3194 12.2439 502 0.7008 0.1024 0.7008 0.8372
0.3194 12.2927 504 0.7394 0.1590 0.7394 0.8599
0.3194 12.3415 506 0.7918 0.1909 0.7918 0.8899
0.3194 12.3902 508 0.7637 0.1480 0.7637 0.8739
0.3194 12.4390 510 0.7458 0.1906 0.7458 0.8636
0.3194 12.4878 512 0.7543 0.1340 0.7543 0.8685
0.3194 12.5366 514 0.7701 0.1803 0.7701 0.8776
0.3194 12.5854 516 0.8431 0.1785 0.8431 0.9182
0.3194 12.6341 518 0.8705 0.1117 0.8705 0.9330
0.3194 12.6829 520 0.7971 0.1451 0.7971 0.8928
0.3194 12.7317 522 0.7533 -0.0195 0.7533 0.8679
0.3194 12.7805 524 0.7853 -0.0283 0.7853 0.8862
0.3194 12.8293 526 0.7563 -0.0170 0.7563 0.8697
0.3194 12.8780 528 0.7561 -0.0138 0.7561 0.8695
0.3194 12.9268 530 0.7581 0.0410 0.7581 0.8707

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
9
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k15_task3_organization

Finetuned
(4222)
this model