ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k13_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6381
  • Qwk: 0.5601
  • Mse: 0.6381
  • Rmse: 0.7988

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0588 2 3.9977 -0.0151 3.9977 1.9994
No log 0.1176 4 2.1675 -0.0525 2.1675 1.4722
No log 0.1765 6 1.7576 -0.0595 1.7576 1.3257
No log 0.2353 8 1.4616 0.0056 1.4616 1.2090
No log 0.2941 10 1.2390 0.0496 1.2390 1.1131
No log 0.3529 12 1.1145 0.1549 1.1145 1.0557
No log 0.4118 14 1.0946 0.2441 1.0946 1.0462
No log 0.4706 16 1.1252 0.2441 1.1252 1.0608
No log 0.5294 18 1.0632 0.1799 1.0632 1.0311
No log 0.5882 20 1.0605 0.1771 1.0605 1.0298
No log 0.6471 22 1.0797 0.0919 1.0797 1.0391
No log 0.7059 24 1.0853 0.0385 1.0853 1.0418
No log 0.7647 26 1.0887 0.2391 1.0887 1.0434
No log 0.8235 28 1.0628 0.2643 1.0628 1.0309
No log 0.8824 30 1.0715 0.1330 1.0715 1.0352
No log 0.9412 32 1.1402 0.0762 1.1402 1.0678
No log 1.0 34 1.1104 0.1203 1.1104 1.0537
No log 1.0588 36 1.0230 0.1389 1.0230 1.0114
No log 1.1176 38 0.9922 0.2291 0.9922 0.9961
No log 1.1765 40 1.0692 0.3003 1.0692 1.0340
No log 1.2353 42 1.1217 0.2880 1.1217 1.0591
No log 1.2941 44 1.0542 0.3229 1.0542 1.0267
No log 1.3529 46 0.9995 0.2716 0.9995 0.9997
No log 1.4118 48 1.0600 0.1196 1.0600 1.0296
No log 1.4706 50 1.0961 0.0981 1.0961 1.0469
No log 1.5294 52 1.0568 0.1076 1.0568 1.0280
No log 1.5882 54 0.9990 0.1545 0.9990 0.9995
No log 1.6471 56 0.9550 0.3393 0.9550 0.9772
No log 1.7059 58 1.0176 0.2842 1.0176 1.0088
No log 1.7647 60 1.0994 0.2142 1.0994 1.0485
No log 1.8235 62 1.2389 0.1281 1.2389 1.1130
No log 1.8824 64 1.3146 0.1281 1.3146 1.1466
No log 1.9412 66 1.1951 0.1968 1.1951 1.0932
No log 2.0 68 1.2402 0.1807 1.2402 1.1136
No log 2.0588 70 1.2306 0.2668 1.2306 1.1093
No log 2.1176 72 1.0799 0.3505 1.0799 1.0392
No log 2.1765 74 1.0698 0.2607 1.0698 1.0343
No log 2.2353 76 1.0569 0.2607 1.0569 1.0281
No log 2.2941 78 1.0392 0.3209 1.0392 1.0194
No log 2.3529 80 1.0204 0.3435 1.0204 1.0101
No log 2.4118 82 1.0335 0.2584 1.0335 1.0166
No log 2.4706 84 0.9953 0.3819 0.9953 0.9976
No log 2.5294 86 0.9628 0.3819 0.9628 0.9812
No log 2.5882 88 0.9194 0.3819 0.9194 0.9589
No log 2.6471 90 0.9119 0.4329 0.9119 0.9550
No log 2.7059 92 0.9086 0.4482 0.9086 0.9532
No log 2.7647 94 0.8566 0.5069 0.8566 0.9255
No log 2.8235 96 1.0294 0.4628 1.0294 1.0146
No log 2.8824 98 1.0833 0.4294 1.0833 1.0408
No log 2.9412 100 1.0096 0.4854 1.0096 1.0048
No log 3.0 102 0.7341 0.5388 0.7341 0.8568
No log 3.0588 104 0.7417 0.4456 0.7417 0.8612
No log 3.1176 106 0.7003 0.4610 0.7003 0.8369
No log 3.1765 108 0.6943 0.4862 0.6943 0.8333
No log 3.2353 110 0.6599 0.5509 0.6599 0.8123
No log 3.2941 112 0.6283 0.5640 0.6283 0.7926
No log 3.3529 114 0.6639 0.5425 0.6639 0.8148
No log 3.4118 116 0.6473 0.5185 0.6473 0.8045
No log 3.4706 118 0.5979 0.6475 0.5979 0.7732
No log 3.5294 120 0.6072 0.6349 0.6072 0.7792
No log 3.5882 122 0.6172 0.5426 0.6172 0.7856
No log 3.6471 124 0.6673 0.4554 0.6673 0.8169
No log 3.7059 126 0.6796 0.4427 0.6796 0.8244
No log 3.7647 128 0.6640 0.5063 0.6640 0.8149
No log 3.8235 130 0.6754 0.5542 0.6754 0.8218
No log 3.8824 132 0.7183 0.5516 0.7183 0.8475
No log 3.9412 134 0.9322 0.5082 0.9322 0.9655
No log 4.0 136 0.8611 0.4794 0.8611 0.9280
No log 4.0588 138 0.6888 0.5808 0.6888 0.8300
No log 4.1176 140 0.7211 0.5603 0.7211 0.8492
No log 4.1765 142 0.6864 0.5618 0.6864 0.8285
No log 4.2353 144 0.6924 0.5113 0.6924 0.8321
No log 4.2941 146 0.7472 0.4958 0.7472 0.8644
No log 4.3529 148 0.8975 0.4666 0.8975 0.9474
No log 4.4118 150 0.8570 0.3884 0.8570 0.9257
No log 4.4706 152 0.7714 0.4041 0.7714 0.8783
No log 4.5294 154 0.7861 0.4546 0.7861 0.8866
No log 4.5882 156 0.7231 0.4730 0.7231 0.8503
No log 4.6471 158 0.7057 0.5033 0.7057 0.8401
No log 4.7059 160 0.7065 0.5273 0.7065 0.8405
No log 4.7647 162 0.7062 0.5131 0.7062 0.8403
No log 4.8235 164 0.7402 0.5442 0.7402 0.8604
No log 4.8824 166 0.7727 0.5385 0.7727 0.8790
No log 4.9412 168 0.6999 0.5698 0.6999 0.8366
No log 5.0 170 0.6571 0.5370 0.6571 0.8106
No log 5.0588 172 0.6475 0.5160 0.6475 0.8047
No log 5.1176 174 0.6601 0.5517 0.6601 0.8125
No log 5.1765 176 0.6649 0.5517 0.6649 0.8154
No log 5.2353 178 0.6720 0.5500 0.6720 0.8197
No log 5.2941 180 0.6741 0.5599 0.6741 0.8210
No log 5.3529 182 0.6972 0.5934 0.6972 0.8350
No log 5.4118 184 0.8016 0.5770 0.8016 0.8953
No log 5.4706 186 0.6730 0.5830 0.6730 0.8203
No log 5.5294 188 0.6601 0.5830 0.6601 0.8125
No log 5.5882 190 0.8657 0.4851 0.8657 0.9304
No log 5.6471 192 0.9873 0.4779 0.9873 0.9936
No log 5.7059 194 1.0825 0.4503 1.0825 1.0404
No log 5.7647 196 0.7961 0.4926 0.7961 0.8922
No log 5.8235 198 0.6346 0.5978 0.6346 0.7966
No log 5.8824 200 0.7162 0.5292 0.7162 0.8463
No log 5.9412 202 0.6620 0.5686 0.6620 0.8136
No log 6.0 204 0.6297 0.5359 0.6297 0.7936
No log 6.0588 206 0.6302 0.5585 0.6302 0.7939
No log 6.1176 208 0.6184 0.5590 0.6184 0.7864
No log 6.1765 210 0.6099 0.5735 0.6099 0.7810
No log 6.2353 212 0.6562 0.6090 0.6562 0.8101
No log 6.2941 214 0.6764 0.5953 0.6764 0.8225
No log 6.3529 216 0.7216 0.6622 0.7216 0.8495
No log 6.4118 218 0.6700 0.6249 0.6700 0.8186
No log 6.4706 220 0.6667 0.5919 0.6667 0.8165
No log 6.5294 222 1.0990 0.4485 1.0990 1.0483
No log 6.5882 224 1.1607 0.3640 1.1607 1.0773
No log 6.6471 226 0.8285 0.5690 0.8285 0.9102
No log 6.7059 228 0.6311 0.5831 0.6311 0.7944
No log 6.7647 230 0.6542 0.6370 0.6542 0.8088
No log 6.8235 232 0.7367 0.5788 0.7367 0.8583
No log 6.8824 234 0.7561 0.5598 0.7561 0.8696
No log 6.9412 236 0.7098 0.4697 0.7098 0.8425
No log 7.0 238 0.6349 0.5759 0.6349 0.7968
No log 7.0588 240 0.6347 0.5747 0.6347 0.7967
No log 7.1176 242 0.6402 0.5759 0.6402 0.8001
No log 7.1765 244 0.6784 0.5363 0.6784 0.8237
No log 7.2353 246 0.7342 0.4439 0.7342 0.8568
No log 7.2941 248 0.7600 0.3921 0.7600 0.8718
No log 7.3529 250 0.8605 0.3808 0.8605 0.9276
No log 7.4118 252 0.8667 0.4065 0.8667 0.9310
No log 7.4706 254 0.7470 0.4576 0.7470 0.8643
No log 7.5294 256 0.6876 0.5274 0.6876 0.8292
No log 7.5882 258 0.7001 0.5747 0.7001 0.8367
No log 7.6471 260 0.6884 0.5510 0.6884 0.8297
No log 7.7059 262 0.6825 0.5245 0.6825 0.8262
No log 7.7647 264 0.7735 0.5178 0.7735 0.8795
No log 7.8235 266 0.7892 0.5028 0.7892 0.8884
No log 7.8824 268 0.7065 0.5088 0.7065 0.8405
No log 7.9412 270 0.6427 0.5405 0.6427 0.8017
No log 8.0 272 0.7013 0.5434 0.7013 0.8375
No log 8.0588 274 0.7500 0.5770 0.7500 0.8660
No log 8.1176 276 0.7031 0.5849 0.7031 0.8385
No log 8.1765 278 0.6271 0.5856 0.6271 0.7919
No log 8.2353 280 0.6356 0.5932 0.6356 0.7973
No log 8.2941 282 0.6749 0.5455 0.6749 0.8215
No log 8.3529 284 0.6761 0.5442 0.6761 0.8223
No log 8.4118 286 0.6308 0.5536 0.6308 0.7942
No log 8.4706 288 0.6393 0.5703 0.6393 0.7996
No log 8.5294 290 0.6170 0.5690 0.6170 0.7855
No log 8.5882 292 0.6483 0.5774 0.6483 0.8052
No log 8.6471 294 0.8026 0.5194 0.8026 0.8959
No log 8.7059 296 0.8839 0.4435 0.8839 0.9401
No log 8.7647 298 0.8508 0.5367 0.8508 0.9224
No log 8.8235 300 0.6966 0.6386 0.6966 0.8346
No log 8.8824 302 0.6047 0.6849 0.6047 0.7776
No log 8.9412 304 0.5982 0.6018 0.5982 0.7735
No log 9.0 306 0.6038 0.6664 0.6038 0.7771
No log 9.0588 308 0.7162 0.5994 0.7162 0.8463
No log 9.1176 310 0.7803 0.6201 0.7803 0.8833
No log 9.1765 312 0.7200 0.5509 0.7200 0.8485
No log 9.2353 314 0.6404 0.4980 0.6404 0.8002
No log 9.2941 316 0.6303 0.5260 0.6303 0.7939
No log 9.3529 318 0.6495 0.5798 0.6495 0.8059
No log 9.4118 320 0.7428 0.6117 0.7428 0.8618
No log 9.4706 322 0.8278 0.5272 0.8278 0.9098
No log 9.5294 324 0.7674 0.6368 0.7674 0.8760
No log 9.5882 326 0.6851 0.6053 0.6851 0.8277
No log 9.6471 328 0.6525 0.6032 0.6525 0.8077
No log 9.7059 330 0.6257 0.5734 0.6257 0.7910
No log 9.7647 332 0.6097 0.5734 0.6097 0.7808
No log 9.8235 334 0.6005 0.5622 0.6005 0.7749
No log 9.8824 336 0.5939 0.5747 0.5939 0.7706
No log 9.9412 338 0.5937 0.5635 0.5937 0.7705
No log 10.0 340 0.5754 0.5747 0.5754 0.7585
No log 10.0588 342 0.5853 0.6350 0.5853 0.7651
No log 10.1176 344 0.6017 0.6386 0.6017 0.7757
No log 10.1765 346 0.5913 0.5810 0.5913 0.7689
No log 10.2353 348 0.5885 0.6196 0.5885 0.7672
No log 10.2941 350 0.5956 0.5983 0.5956 0.7717
No log 10.3529 352 0.6088 0.5862 0.6088 0.7802
No log 10.4118 354 0.6818 0.5618 0.6818 0.8257
No log 10.4706 356 0.6772 0.5618 0.6772 0.8229
No log 10.5294 358 0.6079 0.6311 0.6079 0.7797
No log 10.5882 360 0.6056 0.5977 0.6056 0.7782
No log 10.6471 362 0.6523 0.5887 0.6523 0.8076
No log 10.7059 364 0.6210 0.5902 0.6210 0.7880
No log 10.7647 366 0.5850 0.5856 0.5850 0.7648
No log 10.8235 368 0.5924 0.6084 0.5924 0.7696
No log 10.8824 370 0.5916 0.6272 0.5916 0.7691
No log 10.9412 372 0.6015 0.5725 0.6015 0.7756
No log 11.0 374 0.6348 0.5934 0.6348 0.7968
No log 11.0588 376 0.6415 0.6078 0.6415 0.8009
No log 11.1176 378 0.6287 0.6095 0.6287 0.7929
No log 11.1765 380 0.6140 0.5725 0.6140 0.7836
No log 11.2353 382 0.6077 0.5725 0.6077 0.7795
No log 11.2941 384 0.5929 0.5606 0.5929 0.7700
No log 11.3529 386 0.6057 0.5949 0.6057 0.7783
No log 11.4118 388 0.5974 0.6133 0.5974 0.7729
No log 11.4706 390 0.5898 0.6113 0.5898 0.7680
No log 11.5294 392 0.5858 0.6084 0.5858 0.7653
No log 11.5882 394 0.5926 0.6005 0.5926 0.7698
No log 11.6471 396 0.6397 0.6135 0.6397 0.7998
No log 11.7059 398 0.6701 0.6045 0.6701 0.8186
No log 11.7647 400 0.6373 0.5862 0.6373 0.7983
No log 11.8235 402 0.5888 0.5771 0.5888 0.7674
No log 11.8824 404 0.5836 0.6252 0.5836 0.7640
No log 11.9412 406 0.6084 0.6500 0.6084 0.7800
No log 12.0 408 0.6105 0.6575 0.6105 0.7813
No log 12.0588 410 0.6086 0.6396 0.6086 0.7801
No log 12.1176 412 0.5834 0.6623 0.5834 0.7638
No log 12.1765 414 0.5797 0.5950 0.5797 0.7614
No log 12.2353 416 0.6450 0.5977 0.6450 0.8031
No log 12.2941 418 0.6917 0.5782 0.6917 0.8317
No log 12.3529 420 0.6218 0.5852 0.6218 0.7885
No log 12.4118 422 0.5781 0.6057 0.5781 0.7603
No log 12.4706 424 0.6070 0.6567 0.6070 0.7791
No log 12.5294 426 0.6377 0.6147 0.6377 0.7986
No log 12.5882 428 0.6189 0.6231 0.6189 0.7867
No log 12.6471 430 0.6017 0.5759 0.6017 0.7757
No log 12.7059 432 0.6600 0.5108 0.6600 0.8124
No log 12.7647 434 0.6518 0.5108 0.6518 0.8073
No log 12.8235 436 0.6096 0.5771 0.6096 0.7808
No log 12.8824 438 0.5850 0.5983 0.5850 0.7649
No log 12.9412 440 0.5838 0.6364 0.5838 0.7641
No log 13.0 442 0.6121 0.6049 0.6121 0.7824
No log 13.0588 444 0.6227 0.6032 0.6227 0.7891
No log 13.1176 446 0.5956 0.6124 0.5956 0.7718
No log 13.1765 448 0.5983 0.6055 0.5983 0.7735
No log 13.2353 450 0.6089 0.6063 0.6089 0.7803
No log 13.2941 452 0.5908 0.5874 0.5908 0.7686
No log 13.3529 454 0.6044 0.5561 0.6044 0.7774
No log 13.4118 456 0.6249 0.5573 0.6249 0.7905
No log 13.4706 458 0.6157 0.5782 0.6157 0.7846
No log 13.5294 460 0.6128 0.5955 0.6128 0.7828
No log 13.5882 462 0.6358 0.5940 0.6358 0.7974
No log 13.6471 464 0.6479 0.5770 0.6479 0.8049
No log 13.7059 466 0.6161 0.5955 0.6161 0.7849
No log 13.7647 468 0.6054 0.5955 0.6054 0.7780
No log 13.8235 470 0.6046 0.5955 0.6046 0.7776
No log 13.8824 472 0.6113 0.5999 0.6113 0.7819
No log 13.9412 474 0.6377 0.5581 0.6377 0.7986
No log 14.0 476 0.6413 0.5581 0.6413 0.8008
No log 14.0588 478 0.6323 0.5581 0.6323 0.7952
No log 14.1176 480 0.6134 0.5759 0.6134 0.7832
No log 14.1765 482 0.6042 0.5819 0.6042 0.7773
No log 14.2353 484 0.6137 0.5806 0.6137 0.7834
No log 14.2941 486 0.6149 0.5895 0.6149 0.7842
No log 14.3529 488 0.6066 0.6011 0.6066 0.7788
No log 14.4118 490 0.6028 0.5909 0.6028 0.7764
No log 14.4706 492 0.6110 0.5934 0.6110 0.7816
No log 14.5294 494 0.6264 0.5854 0.6264 0.7915
No log 14.5882 496 0.6152 0.5879 0.6152 0.7843
No log 14.6471 498 0.5965 0.5759 0.5965 0.7724
0.2947 14.7059 500 0.5832 0.5843 0.5832 0.7637
0.2947 14.7647 502 0.5743 0.6001 0.5743 0.7578
0.2947 14.8235 504 0.5687 0.5990 0.5687 0.7542
0.2947 14.8824 506 0.5659 0.5868 0.5659 0.7523
0.2947 14.9412 508 0.5699 0.5748 0.5699 0.7549
0.2947 15.0 510 0.5704 0.5905 0.5704 0.7553
0.2947 15.0588 512 0.5837 0.5881 0.5837 0.7640
0.2947 15.1176 514 0.5826 0.5881 0.5826 0.7633
0.2947 15.1765 516 0.5882 0.5737 0.5882 0.7669
0.2947 15.2353 518 0.6034 0.5737 0.6034 0.7768
0.2947 15.2941 520 0.6245 0.5662 0.6245 0.7903
0.2947 15.3529 522 0.6582 0.5593 0.6582 0.8113
0.2947 15.4118 524 0.6381 0.5601 0.6381 0.7988

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
8
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k13_task5_organization

Finetuned
(4223)
this model