ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k5_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7745
  • Qwk: 0.1352
  • Mse: 0.7745
  • Rmse: 0.8801

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.125 2 3.4094 0.0175 3.4094 1.8464
No log 0.25 4 1.6979 0.0591 1.6979 1.3030
No log 0.375 6 1.3282 0.0350 1.3283 1.1525
No log 0.5 8 2.2564 -0.0481 2.2564 1.5021
No log 0.625 10 1.0995 0.0493 1.0995 1.0486
No log 0.75 12 0.6739 0.0 0.6739 0.8209
No log 0.875 14 0.7176 0.0 0.7176 0.8471
No log 1.0 16 0.9512 -0.0982 0.9512 0.9753
No log 1.125 18 1.3529 0.0 1.3529 1.1631
No log 1.25 20 1.2340 0.0 1.2340 1.1109
No log 1.375 22 0.9524 -0.0638 0.9524 0.9759
No log 1.5 24 0.8279 -0.0790 0.8279 0.9099
No log 1.625 26 0.7571 0.0416 0.7571 0.8701
No log 1.75 28 0.7752 0.1097 0.7752 0.8805
No log 1.875 30 1.1874 0.0065 1.1874 1.0897
No log 2.0 32 1.2014 0.0065 1.2014 1.0961
No log 2.125 34 0.9022 -0.0595 0.9022 0.9498
No log 2.25 36 0.7832 0.0588 0.7832 0.8850
No log 2.375 38 0.6997 0.0506 0.6997 0.8365
No log 2.5 40 0.7386 0.0374 0.7386 0.8594
No log 2.625 42 0.9403 -0.0532 0.9403 0.9697
No log 2.75 44 0.8598 -0.0442 0.8598 0.9272
No log 2.875 46 0.7942 0.0714 0.7942 0.8912
No log 3.0 48 0.7515 0.1259 0.7515 0.8669
No log 3.125 50 0.6875 0.0555 0.6875 0.8292
No log 3.25 52 0.6846 0.0555 0.6846 0.8274
No log 3.375 54 0.8374 0.0826 0.8374 0.9151
No log 3.5 56 1.0432 0.0107 1.0432 1.0214
No log 3.625 58 1.2704 0.0642 1.2704 1.1271
No log 3.75 60 0.7770 0.0129 0.7770 0.8815
No log 3.875 62 0.6869 0.0555 0.6869 0.8288
No log 4.0 64 0.7112 0.0555 0.7112 0.8433
No log 4.125 66 0.7023 0.0555 0.7023 0.8380
No log 4.25 68 0.7800 0.0549 0.7800 0.8832
No log 4.375 70 0.9905 0.0984 0.9905 0.9953
No log 4.5 72 1.0070 0.0556 1.0070 1.0035
No log 4.625 74 0.8070 0.0129 0.8070 0.8983
No log 4.75 76 0.8591 -0.0912 0.8591 0.9268
No log 4.875 78 0.8254 0.0423 0.8254 0.9085
No log 5.0 80 0.8530 0.0660 0.8530 0.9236
No log 5.125 82 0.8781 0.0271 0.8781 0.9371
No log 5.25 84 1.1032 0.0305 1.1032 1.0503
No log 5.375 86 1.0297 0.0541 1.0297 1.0148
No log 5.5 88 0.8114 0.0 0.8114 0.9008
No log 5.625 90 0.7977 0.0327 0.7977 0.8931
No log 5.75 92 0.7685 0.0028 0.7685 0.8766
No log 5.875 94 0.8385 -0.0408 0.8385 0.9157
No log 6.0 96 0.9114 0.0446 0.9114 0.9547
No log 6.125 98 0.7543 0.0027 0.7543 0.8685
No log 6.25 100 1.1853 0.0912 1.1853 1.0887
No log 6.375 102 1.0300 0.1152 1.0300 1.0149
No log 6.5 104 0.7910 -0.0166 0.7910 0.8894
No log 6.625 106 0.9257 0.0805 0.9257 0.9622
No log 6.75 108 0.8037 0.0139 0.8037 0.8965
No log 6.875 110 0.8512 0.0966 0.8512 0.9226
No log 7.0 112 0.9341 0.0037 0.9341 0.9665
No log 7.125 114 0.8077 0.1094 0.8077 0.8987
No log 7.25 116 0.7928 0.0327 0.7928 0.8904
No log 7.375 118 0.7898 0.0359 0.7898 0.8887
No log 7.5 120 0.7922 0.0810 0.7922 0.8900
No log 7.625 122 0.7948 0.0804 0.7948 0.8915
No log 7.75 124 0.7807 0.0764 0.7807 0.8836
No log 7.875 126 0.7884 0.0679 0.7884 0.8879
No log 8.0 128 0.7647 0.1786 0.7647 0.8745
No log 8.125 130 0.7615 0.1434 0.7615 0.8727
No log 8.25 132 0.8287 0.0909 0.8287 0.9103
No log 8.375 134 0.9297 0.0651 0.9297 0.9642
No log 8.5 136 0.7913 0.0295 0.7913 0.8896
No log 8.625 138 0.8332 0.1184 0.8332 0.9128
No log 8.75 140 0.9223 0.0563 0.9223 0.9604
No log 8.875 142 0.9332 0.0713 0.9332 0.9660
No log 9.0 144 0.9356 0.1127 0.9356 0.9673
No log 9.125 146 1.0171 0.0977 1.0171 1.0085
No log 9.25 148 0.9336 0.0504 0.9336 0.9662
No log 9.375 150 0.8462 0.1519 0.8462 0.9199
No log 9.5 152 0.8167 0.2229 0.8167 0.9037
No log 9.625 154 0.8245 0.0875 0.8245 0.9080
No log 9.75 156 0.8611 0.0883 0.8611 0.9279
No log 9.875 158 0.8400 0.0870 0.8400 0.9165
No log 10.0 160 0.9490 0.1560 0.9490 0.9741
No log 10.125 162 0.8733 0.1648 0.8733 0.9345
No log 10.25 164 0.8244 0.0875 0.8244 0.9080
No log 10.375 166 0.8250 0.1569 0.8250 0.9083
No log 10.5 168 0.8084 0.0833 0.8084 0.8991
No log 10.625 170 1.0603 0.1142 1.0603 1.0297
No log 10.75 172 1.1265 0.0855 1.1265 1.0614
No log 10.875 174 0.8625 0.1039 0.8625 0.9287
No log 11.0 176 0.9701 0.0659 0.9701 0.9850
No log 11.125 178 1.0397 0.0267 1.0397 1.0197
No log 11.25 180 0.8544 0.1198 0.8544 0.9243
No log 11.375 182 1.1712 0.0865 1.1712 1.0822
No log 11.5 184 1.2211 0.1134 1.2211 1.1050
No log 11.625 186 1.0093 0.0841 1.0093 1.0046
No log 11.75 188 0.8224 0.1519 0.8224 0.9068
No log 11.875 190 0.8748 0.0497 0.8748 0.9353
No log 12.0 192 0.7722 0.0778 0.7722 0.8787
No log 12.125 194 0.8878 0.0810 0.8878 0.9422
No log 12.25 196 0.9886 0.0855 0.9886 0.9943
No log 12.375 198 0.8253 0.0392 0.8253 0.9085
No log 12.5 200 0.7268 0.1495 0.7268 0.8525
No log 12.625 202 0.9718 0.0416 0.9718 0.9858
No log 12.75 204 0.8786 0.0277 0.8786 0.9373
No log 12.875 206 0.7082 0.2142 0.7082 0.8415
No log 13.0 208 0.8993 0.0454 0.8993 0.9483
No log 13.125 210 0.8990 0.0443 0.8990 0.9481
No log 13.25 212 0.7715 0.2454 0.7715 0.8783
No log 13.375 214 0.9153 0.1794 0.9153 0.9567
No log 13.5 216 0.9934 0.0988 0.9934 0.9967
No log 13.625 218 0.8382 0.1829 0.8382 0.9155
No log 13.75 220 0.8648 0.0337 0.8648 0.9300
No log 13.875 222 0.8469 0.1040 0.8469 0.9203
No log 14.0 224 0.7890 0.0816 0.7890 0.8883
No log 14.125 226 0.8506 0.1522 0.8506 0.9223
No log 14.25 228 0.7897 0.1184 0.7897 0.8886
No log 14.375 230 0.8486 0.0747 0.8486 0.9212
No log 14.5 232 1.0045 0.0267 1.0045 1.0022
No log 14.625 234 0.9132 0.0502 0.9132 0.9556
No log 14.75 236 0.7375 0.1425 0.7375 0.8588
No log 14.875 238 0.8086 0.0959 0.8086 0.8992
No log 15.0 240 0.8144 0.1049 0.8144 0.9024
No log 15.125 242 0.8299 0.0623 0.8299 0.9110
No log 15.25 244 0.8635 0.0689 0.8635 0.9292
No log 15.375 246 0.8796 0.0721 0.8796 0.9379
No log 15.5 248 0.7948 0.0200 0.7948 0.8915
No log 15.625 250 0.7342 0.0089 0.7342 0.8569
No log 15.75 252 0.7143 -0.0449 0.7143 0.8451
No log 15.875 254 0.7288 -0.0488 0.7288 0.8537
No log 16.0 256 0.7473 0.0840 0.7473 0.8645
No log 16.125 258 0.7922 -0.0259 0.7922 0.8900
No log 16.25 260 0.7956 -0.0204 0.7956 0.8920
No log 16.375 262 0.7251 0.0432 0.7251 0.8515
No log 16.5 264 0.7192 0.0 0.7192 0.8480
No log 16.625 266 0.7776 0.1127 0.7776 0.8818
No log 16.75 268 0.7492 0.1525 0.7492 0.8655
No log 16.875 270 0.7345 0.1095 0.7345 0.8570
No log 17.0 272 0.7815 0.1451 0.7815 0.8840
No log 17.125 274 0.9235 0.0050 0.9235 0.9610
No log 17.25 276 0.8725 0.0706 0.8725 0.9341
No log 17.375 278 0.8338 0.1623 0.8338 0.9131
No log 17.5 280 0.7845 0.0709 0.7845 0.8857
No log 17.625 282 0.7704 0.0709 0.7704 0.8777
No log 17.75 284 0.7872 0.1673 0.7872 0.8872
No log 17.875 286 0.7880 0.1673 0.7880 0.8877
No log 18.0 288 0.7527 0.0709 0.7527 0.8676
No log 18.125 290 0.7374 0.0709 0.7374 0.8587
No log 18.25 292 0.7832 0.0606 0.7832 0.8850
No log 18.375 294 0.8280 0.0706 0.8280 0.9099
No log 18.5 296 0.8128 0.1509 0.8128 0.9015
No log 18.625 298 0.8490 0.1870 0.8490 0.9214
No log 18.75 300 0.8514 0.1854 0.8514 0.9227
No log 18.875 302 0.8656 0.1615 0.8656 0.9304
No log 19.0 304 0.8512 0.1442 0.8512 0.9226
No log 19.125 306 0.7950 0.0693 0.7950 0.8917
No log 19.25 308 0.7456 0.0934 0.7456 0.8635
No log 19.375 310 0.7453 0.0828 0.7453 0.8633
No log 19.5 312 0.7761 -0.0389 0.7761 0.8810
No log 19.625 314 0.8336 0.0362 0.8336 0.9130
No log 19.75 316 0.9129 0.0728 0.9129 0.9555
No log 19.875 318 0.9274 0.1041 0.9274 0.9630
No log 20.0 320 0.8526 0.1251 0.8526 0.9234
No log 20.125 322 0.8406 0.1006 0.8406 0.9169
No log 20.25 324 0.7825 0.1518 0.7825 0.8846
No log 20.375 326 0.7819 0.1035 0.7819 0.8842
No log 20.5 328 0.7459 -0.0026 0.7459 0.8637
No log 20.625 330 0.7229 0.0828 0.7229 0.8503
No log 20.75 332 0.7109 0.0918 0.7109 0.8432
No log 20.875 334 0.7155 0.0585 0.7155 0.8459
No log 21.0 336 0.7370 0.0141 0.7370 0.8585
No log 21.125 338 0.7344 0.0376 0.7344 0.8570
No log 21.25 340 0.7660 0.1192 0.7660 0.8752
No log 21.375 342 0.7850 0.1580 0.7850 0.8860
No log 21.5 344 0.8438 0.0876 0.8438 0.9186
No log 21.625 346 0.8285 0.0880 0.8285 0.9102
No log 21.75 348 0.7828 0.1580 0.7828 0.8848
No log 21.875 350 0.7785 0.2156 0.7785 0.8823
No log 22.0 352 0.7472 0.1425 0.7472 0.8644
No log 22.125 354 0.7028 0.0973 0.7028 0.8383
No log 22.25 356 0.7026 0.0680 0.7026 0.8382
No log 22.375 358 0.7410 0.0068 0.7410 0.8608
No log 22.5 360 0.7865 0.0867 0.7865 0.8869
No log 22.625 362 0.7306 0.1095 0.7306 0.8547
No log 22.75 364 0.7617 0.0157 0.7617 0.8728
No log 22.875 366 0.8269 0.1037 0.8269 0.9094
No log 23.0 368 0.8259 0.1037 0.8259 0.9088
No log 23.125 370 0.7694 -0.0051 0.7694 0.8771
No log 23.25 372 0.7742 0.0660 0.7742 0.8799
No log 23.375 374 0.7612 0.1495 0.7612 0.8724
No log 23.5 376 0.7411 0.0028 0.7411 0.8609
No log 23.625 378 0.7688 0.0683 0.7688 0.8768
No log 23.75 380 0.7904 0.1169 0.7904 0.8890
No log 23.875 382 0.7206 0.0058 0.7206 0.8489
No log 24.0 384 0.7086 0.0376 0.7086 0.8418
No log 24.125 386 0.7123 0.0732 0.7123 0.8440
No log 24.25 388 0.7134 0.1192 0.7134 0.8446
No log 24.375 390 0.7377 -0.0026 0.7377 0.8589
No log 24.5 392 0.8083 0.1036 0.8083 0.8991
No log 24.625 394 0.7793 0.1034 0.7793 0.8828
No log 24.75 396 0.7236 0.0834 0.7236 0.8506
No log 24.875 398 0.7106 0.0834 0.7106 0.8430
No log 25.0 400 0.7623 -0.0293 0.7623 0.8731
No log 25.125 402 0.7944 0.0279 0.7944 0.8913
No log 25.25 404 0.7409 -0.0350 0.7409 0.8607
No log 25.375 406 0.7383 0.1141 0.7383 0.8592
No log 25.5 408 0.7428 0.0840 0.7428 0.8619
No log 25.625 410 0.7623 -0.0762 0.7623 0.8731
No log 25.75 412 0.7968 0.0664 0.7968 0.8926
No log 25.875 414 0.8111 0.1711 0.8111 0.9006
No log 26.0 416 0.8165 0.1519 0.8165 0.9036
No log 26.125 418 0.8418 0.1746 0.8418 0.9175
No log 26.25 420 0.8453 0.0884 0.8453 0.9194
No log 26.375 422 0.8026 0.1923 0.8026 0.8959
No log 26.5 424 0.7759 0.1923 0.7759 0.8809
No log 26.625 426 0.7446 0.2070 0.7446 0.8629
No log 26.75 428 0.7295 0.2070 0.7295 0.8541
No log 26.875 430 0.7251 0.2070 0.7251 0.8515
No log 27.0 432 0.7315 0.2005 0.7315 0.8553
No log 27.125 434 0.7580 0.2015 0.7580 0.8706
No log 27.25 436 0.7935 0.2015 0.7935 0.8908
No log 27.375 438 0.7867 0.2334 0.7867 0.8870
No log 27.5 440 0.7656 0.1479 0.7656 0.8750
No log 27.625 442 0.7630 0.1823 0.7630 0.8735
No log 27.75 444 0.7489 0.1859 0.7489 0.8654
No log 27.875 446 0.7142 0.1599 0.7142 0.8451
No log 28.0 448 0.7258 -0.0293 0.7258 0.8519
No log 28.125 450 0.7309 -0.0307 0.7309 0.8550
No log 28.25 452 0.6902 -0.0032 0.6902 0.8308
No log 28.375 454 0.7025 0.1627 0.7025 0.8381
No log 28.5 456 0.7609 0.1387 0.7609 0.8723
No log 28.625 458 0.7625 0.1495 0.7625 0.8732
No log 28.75 460 0.7393 0.0327 0.7393 0.8598
No log 28.875 462 0.7435 0.0327 0.7435 0.8623
No log 29.0 464 0.7646 0.0989 0.7646 0.8744
No log 29.125 466 0.7662 0.0989 0.7662 0.8754
No log 29.25 468 0.7471 0.1465 0.7471 0.8644
No log 29.375 470 0.7707 0.1440 0.7707 0.8779
No log 29.5 472 0.7464 0.1440 0.7464 0.8639
No log 29.625 474 0.7602 0.0129 0.7602 0.8719
No log 29.75 476 0.7998 0.0257 0.7998 0.8943
No log 29.875 478 0.7763 0.0570 0.7763 0.8811
No log 30.0 480 0.7753 0.1841 0.7753 0.8805
No log 30.125 482 0.8179 0.1783 0.8179 0.9044
No log 30.25 484 0.8005 0.1415 0.8005 0.8947
No log 30.375 486 0.8012 0.1212 0.8012 0.8951
No log 30.5 488 0.8314 0.0962 0.8314 0.9118
No log 30.625 490 0.7915 0.1673 0.7915 0.8897
No log 30.75 492 0.7536 0.1094 0.7536 0.8681
No log 30.875 494 0.7708 0.1431 0.7708 0.8779
No log 31.0 496 0.7564 0.0679 0.7564 0.8697
No log 31.125 498 0.7454 0.1184 0.7454 0.8634
0.2698 31.25 500 0.7257 0.1644 0.7257 0.8519
0.2698 31.375 502 0.7103 0.1244 0.7103 0.8428
0.2698 31.5 504 0.7104 0.1354 0.7104 0.8428
0.2698 31.625 506 0.7243 0.1354 0.7243 0.8511
0.2698 31.75 508 0.7489 0.1315 0.7489 0.8654
0.2698 31.875 510 0.7745 0.1352 0.7745 0.8801

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
8
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k5_task3_organization

Finetuned
(4222)
this model