ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k16_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9556
  • Qwk: 0.3692
  • Mse: 0.9556
  • Rmse: 0.9775

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0345 2 4.8117 -0.0020 4.8117 2.1936
No log 0.0690 4 2.7032 0.0329 2.7032 1.6442
No log 0.1034 6 1.9974 -0.0575 1.9974 1.4133
No log 0.1379 8 1.6601 0.0062 1.6601 1.2884
No log 0.1724 10 1.5266 0.0062 1.5266 1.2356
No log 0.2069 12 1.5861 0.0372 1.5861 1.2594
No log 0.2414 14 1.5667 0.0227 1.5667 1.2517
No log 0.2759 16 1.6283 0.1060 1.6283 1.2760
No log 0.3103 18 1.7799 0.1169 1.7799 1.3341
No log 0.3448 20 1.6994 0.0925 1.6994 1.3036
No log 0.3793 22 1.6897 0.1169 1.6897 1.2999
No log 0.4138 24 1.5852 0.1248 1.5852 1.2590
No log 0.4483 26 1.6672 0.1557 1.6672 1.2912
No log 0.4828 28 1.8755 0.1544 1.8755 1.3695
No log 0.5172 30 1.5571 0.1540 1.5571 1.2479
No log 0.5517 32 1.1454 0.2736 1.1454 1.0702
No log 0.5862 34 0.9746 0.3704 0.9746 0.9872
No log 0.6207 36 0.9922 0.4093 0.9922 0.9961
No log 0.6552 38 1.0215 0.4198 1.0215 1.0107
No log 0.6897 40 1.0233 0.3590 1.0233 1.0116
No log 0.7241 42 1.0497 0.3679 1.0497 1.0245
No log 0.7586 44 1.2180 0.1708 1.2180 1.1036
No log 0.7931 46 1.2879 0.1251 1.2879 1.1348
No log 0.8276 48 1.3104 0.1278 1.3104 1.1447
No log 0.8621 50 1.7577 -0.0208 1.7577 1.3258
No log 0.8966 52 1.8735 -0.0016 1.8735 1.3688
No log 0.9310 54 1.3789 0.1663 1.3789 1.1743
No log 0.9655 56 1.1933 0.3692 1.1933 1.0924
No log 1.0 58 1.0214 0.3409 1.0214 1.0106
No log 1.0345 60 0.9670 0.3409 0.9670 0.9834
No log 1.0690 62 0.9922 0.3452 0.9922 0.9961
No log 1.1034 64 1.1079 0.2915 1.1079 1.0526
No log 1.1379 66 1.1836 0.1722 1.1836 1.0879
No log 1.1724 68 1.3852 0.1553 1.3852 1.1769
No log 1.2069 70 1.2870 0.1535 1.2870 1.1344
No log 1.2414 72 1.0767 0.1655 1.0767 1.0376
No log 1.2759 74 1.1307 0.1979 1.1307 1.0633
No log 1.3103 76 1.4725 0.2521 1.4725 1.2135
No log 1.3448 78 1.8049 0.2599 1.8049 1.3434
No log 1.3793 80 1.5817 0.3092 1.5817 1.2577
No log 1.4138 82 1.3520 0.2298 1.3520 1.1627
No log 1.4483 84 1.2536 0.1979 1.2536 1.1197
No log 1.4828 86 1.1393 0.2103 1.1393 1.0674
No log 1.5172 88 1.1071 0.2870 1.1071 1.0522
No log 1.5517 90 1.1676 0.2065 1.1676 1.0806
No log 1.5862 92 0.9927 0.2984 0.9927 0.9964
No log 1.6207 94 0.9029 0.4598 0.9029 0.9502
No log 1.6552 96 1.0129 0.3030 1.0129 1.0064
No log 1.6897 98 1.1996 0.3274 1.1996 1.0953
No log 1.7241 100 1.1091 0.3274 1.1091 1.0531
No log 1.7586 102 0.9503 0.4842 0.9503 0.9748
No log 1.7931 104 0.9537 0.5054 0.9537 0.9766
No log 1.8276 106 0.9243 0.5054 0.9243 0.9614
No log 1.8621 108 0.8805 0.5479 0.8805 0.9384
No log 1.8966 110 1.0453 0.4067 1.0453 1.0224
No log 1.9310 112 1.4835 0.3389 1.4835 1.2180
No log 1.9655 114 1.7017 0.2684 1.7017 1.3045
No log 2.0 116 1.4624 0.3712 1.4624 1.2093
No log 2.0345 118 1.1817 0.3860 1.1817 1.0870
No log 2.0690 120 1.0546 0.3621 1.0546 1.0269
No log 2.1034 122 0.9041 0.4310 0.9041 0.9508
No log 2.1379 124 0.8614 0.4465 0.8614 0.9281
No log 2.1724 126 0.8435 0.4806 0.8435 0.9184
No log 2.2069 128 1.0309 0.3902 1.0309 1.0153
No log 2.2414 130 1.4667 0.2816 1.4667 1.2111
No log 2.2759 132 1.4432 0.2706 1.4432 1.2013
No log 2.3103 134 1.1583 0.3897 1.1583 1.0763
No log 2.3448 136 1.0122 0.3821 1.0122 1.0061
No log 2.3793 138 1.0020 0.4118 1.0020 1.0010
No log 2.4138 140 1.1559 0.4009 1.1559 1.0751
No log 2.4483 142 1.5140 0.3257 1.5140 1.2305
No log 2.4828 144 1.3582 0.4064 1.3582 1.1654
No log 2.5172 146 1.3621 0.3550 1.3621 1.1671
No log 2.5517 148 1.1156 0.4561 1.1156 1.0562
No log 2.5862 150 1.1557 0.4169 1.1557 1.0750
No log 2.6207 152 1.5564 0.2981 1.5564 1.2476
No log 2.6552 154 1.6650 0.2649 1.6650 1.2904
No log 2.6897 156 1.3601 0.3541 1.3601 1.1662
No log 2.7241 158 1.0749 0.4412 1.0749 1.0368
No log 2.7586 160 1.2526 0.3902 1.2526 1.1192
No log 2.7931 162 1.4439 0.3051 1.4439 1.2016
No log 2.8276 164 1.4952 0.3051 1.4952 1.2228
No log 2.8621 166 1.1453 0.3833 1.1453 1.0702
No log 2.8966 168 0.9684 0.4078 0.9684 0.9841
No log 2.9310 170 0.9191 0.3787 0.9191 0.9587
No log 2.9655 172 0.9310 0.4033 0.9310 0.9649
No log 3.0 174 0.9473 0.3557 0.9473 0.9733
No log 3.0345 176 0.9300 0.4072 0.9300 0.9644
No log 3.0690 178 0.9630 0.3457 0.9630 0.9813
No log 3.1034 180 1.0274 0.3284 1.0274 1.0136
No log 3.1379 182 1.0903 0.3009 1.0903 1.0442
No log 3.1724 184 1.1070 0.3323 1.1070 1.0521
No log 3.2069 186 1.2277 0.2589 1.2277 1.1080
No log 3.2414 188 1.0425 0.3330 1.0425 1.0210
No log 3.2759 190 0.9587 0.3344 0.9587 0.9791
No log 3.3103 192 1.0568 0.3945 1.0568 1.0280
No log 3.3448 194 1.0240 0.3854 1.0240 1.0119
No log 3.3793 196 1.0072 0.2851 1.0072 1.0036
No log 3.4138 198 1.0881 0.2963 1.0881 1.0431
No log 3.4483 200 1.0075 0.3081 1.0075 1.0037
No log 3.4828 202 1.0168 0.3979 1.0168 1.0084
No log 3.5172 204 1.0098 0.3979 1.0098 1.0049
No log 3.5517 206 0.9942 0.4013 0.9942 0.9971
No log 3.5862 208 1.0301 0.3871 1.0301 1.0149
No log 3.6207 210 1.0473 0.4369 1.0473 1.0234
No log 3.6552 212 0.9983 0.3846 0.9983 0.9992
No log 3.6897 214 0.9914 0.4405 0.9914 0.9957
No log 3.7241 216 0.9856 0.3772 0.9856 0.9928
No log 3.7586 218 0.9906 0.3849 0.9906 0.9953
No log 3.7931 220 1.0888 0.3237 1.0888 1.0435
No log 3.8276 222 1.1678 0.3005 1.1678 1.0807
No log 3.8621 224 1.0539 0.3397 1.0539 1.0266
No log 3.8966 226 0.9951 0.4105 0.9951 0.9976
No log 3.9310 228 0.9853 0.4160 0.9853 0.9926
No log 3.9655 230 0.9947 0.3243 0.9947 0.9973
No log 4.0 232 1.0162 0.3747 1.0162 1.0081
No log 4.0345 234 1.0044 0.3392 1.0044 1.0022
No log 4.0690 236 1.1089 0.3665 1.1089 1.0531
No log 4.1034 238 1.2335 0.2704 1.2335 1.1106
No log 4.1379 240 1.0780 0.3813 1.0780 1.0383
No log 4.1724 242 1.0158 0.3327 1.0158 1.0079
No log 4.2069 244 1.0175 0.3459 1.0175 1.0087
No log 4.2414 246 1.0213 0.3590 1.0213 1.0106
No log 4.2759 248 1.0710 0.2876 1.0710 1.0349
No log 4.3103 250 1.0504 0.2577 1.0504 1.0249
No log 4.3448 252 1.0232 0.2943 1.0232 1.0115
No log 4.3793 254 1.0366 0.3036 1.0366 1.0181
No log 4.4138 256 1.0712 0.3056 1.0712 1.0350
No log 4.4483 258 1.0845 0.3008 1.0845 1.0414
No log 4.4828 260 1.0099 0.2943 1.0099 1.0049
No log 4.5172 262 0.9869 0.3237 0.9869 0.9934
No log 4.5517 264 0.9958 0.3496 0.9958 0.9979
No log 4.5862 266 1.0040 0.3590 1.0040 1.0020
No log 4.6207 268 1.1690 0.4045 1.1690 1.0812
No log 4.6552 270 1.2271 0.3421 1.2271 1.1077
No log 4.6897 272 1.0495 0.3189 1.0495 1.0244
No log 4.7241 274 0.9703 0.3724 0.9703 0.9850
No log 4.7586 276 1.0219 0.4293 1.0219 1.0109
No log 4.7931 278 0.9800 0.3369 0.9800 0.9900
No log 4.8276 280 0.9508 0.3933 0.9508 0.9751
No log 4.8621 282 1.0491 0.3128 1.0491 1.0243
No log 4.8966 284 1.1764 0.3562 1.1764 1.0846
No log 4.9310 286 1.1403 0.3160 1.1403 1.0679
No log 4.9655 288 0.9790 0.4396 0.9790 0.9895
No log 5.0 290 0.9375 0.4105 0.9375 0.9682
No log 5.0345 292 0.9687 0.4388 0.9687 0.9842
No log 5.0690 294 1.0866 0.4122 1.0866 1.0424
No log 5.1034 296 1.1220 0.3064 1.1220 1.0592
No log 5.1379 298 1.0078 0.5085 1.0078 1.0039
No log 5.1724 300 0.9738 0.3562 0.9738 0.9868
No log 5.2069 302 0.9819 0.3690 0.9819 0.9909
No log 5.2414 304 0.9853 0.3463 0.9853 0.9926
No log 5.2759 306 1.0283 0.3218 1.0283 1.0140
No log 5.3103 308 1.0049 0.3637 1.0049 1.0025
No log 5.3448 310 0.9847 0.3541 0.9847 0.9923
No log 5.3793 312 0.9839 0.3042 0.9839 0.9919
No log 5.4138 314 0.9785 0.3230 0.9785 0.9892
No log 5.4483 316 0.9924 0.3108 0.9924 0.9962
No log 5.4828 318 1.0260 0.3985 1.0260 1.0129
No log 5.5172 320 1.0015 0.3218 1.0015 1.0007
No log 5.5517 322 0.9711 0.3243 0.9711 0.9854
No log 5.5862 324 0.9496 0.3811 0.9496 0.9745
No log 5.6207 326 0.9410 0.3619 0.9410 0.9701
No log 5.6552 328 0.9793 0.4305 0.9793 0.9896
No log 5.6897 330 1.1629 0.3731 1.1629 1.0784
No log 5.7241 332 1.1563 0.3640 1.1563 1.0753
No log 5.7586 334 1.0082 0.4020 1.0082 1.0041
No log 5.7931 336 0.9609 0.4036 0.9609 0.9803
No log 5.8276 338 0.9239 0.3596 0.9239 0.9612
No log 5.8621 340 0.9027 0.4256 0.9027 0.9501
No log 5.8966 342 0.9192 0.5209 0.9192 0.9587
No log 5.9310 344 1.0209 0.4119 1.0209 1.0104
No log 5.9655 346 0.9672 0.4861 0.9672 0.9835
No log 6.0 348 0.8562 0.4231 0.8562 0.9253
No log 6.0345 350 0.9171 0.3684 0.9171 0.9577
No log 6.0690 352 0.9243 0.3908 0.9243 0.9614
No log 6.1034 354 0.8875 0.4424 0.8875 0.9421
No log 6.1379 356 0.9235 0.4116 0.9235 0.9610
No log 6.1724 358 0.9888 0.4321 0.9888 0.9944
No log 6.2069 360 0.9391 0.4154 0.9391 0.9691
No log 6.2414 362 0.8871 0.3855 0.8871 0.9419
No log 6.2759 364 0.8855 0.3970 0.8855 0.9410
No log 6.3103 366 0.8951 0.4260 0.8951 0.9461
No log 6.3448 368 0.8991 0.4079 0.8991 0.9482
No log 6.3793 370 0.9021 0.3838 0.9021 0.9498
No log 6.4138 372 0.9139 0.3862 0.9139 0.9560
No log 6.4483 374 1.0087 0.3881 1.0087 1.0044
No log 6.4828 376 1.0744 0.3877 1.0744 1.0365
No log 6.5172 378 1.1908 0.3745 1.1908 1.0913
No log 6.5517 380 1.1997 0.3860 1.1997 1.0953
No log 6.5862 382 1.0537 0.3584 1.0537 1.0265
No log 6.6207 384 1.0070 0.4435 1.0070 1.0035
No log 6.6552 386 1.0496 0.3466 1.0496 1.0245
No log 6.6897 388 1.0887 0.3415 1.0887 1.0434
No log 6.7241 390 1.0665 0.3218 1.0665 1.0327
No log 6.7586 392 0.9976 0.3546 0.9976 0.9988
No log 6.7931 394 1.0080 0.3135 1.0080 1.0040
No log 6.8276 396 1.0324 0.3173 1.0324 1.0161
No log 6.8621 398 1.0505 0.3078 1.0505 1.0249
No log 6.8966 400 1.0 0.4305 1.0000 1.0
No log 6.9310 402 1.0401 0.2844 1.0401 1.0199
No log 6.9655 404 1.0173 0.2844 1.0173 1.0086
No log 7.0 406 0.9733 0.3657 0.9733 0.9866
No log 7.0345 408 0.9880 0.3321 0.9880 0.9940
No log 7.0690 410 1.0352 0.3218 1.0352 1.0174
No log 7.1034 412 1.0133 0.3521 1.0133 1.0066
No log 7.1379 414 0.9483 0.4310 0.9483 0.9738
No log 7.1724 416 0.8894 0.4772 0.8894 0.9431
No log 7.2069 418 0.8747 0.4504 0.8747 0.9352
No log 7.2414 420 0.8831 0.4404 0.8831 0.9397
No log 7.2759 422 0.9827 0.4012 0.9827 0.9913
No log 7.3103 424 1.0034 0.3835 1.0034 1.0017
No log 7.3448 426 0.9513 0.4283 0.9513 0.9753
No log 7.3793 428 0.8735 0.4494 0.8735 0.9346
No log 7.4138 430 0.8431 0.4978 0.8431 0.9182
No log 7.4483 432 0.8401 0.5226 0.8401 0.9166
No log 7.4828 434 0.8393 0.5058 0.8393 0.9161
No log 7.5172 436 0.8542 0.5125 0.8542 0.9242
No log 7.5517 438 0.8841 0.4273 0.8841 0.9403
No log 7.5862 440 0.9077 0.4578 0.9077 0.9527
No log 7.6207 442 0.9356 0.3946 0.9356 0.9673
No log 7.6552 444 0.9673 0.3884 0.9673 0.9835
No log 7.6897 446 1.0147 0.3869 1.0147 1.0073
No log 7.7241 448 0.9961 0.3869 0.9961 0.9980
No log 7.7586 450 0.9265 0.4413 0.9265 0.9626
No log 7.7931 452 0.9063 0.4321 0.9063 0.9520
No log 7.8276 454 0.9544 0.3854 0.9544 0.9769
No log 7.8621 456 0.9639 0.2784 0.9639 0.9818
No log 7.8966 458 0.9784 0.3891 0.9784 0.9891
No log 7.9310 460 1.1336 0.3640 1.1336 1.0647
No log 7.9655 462 1.2694 0.3243 1.2694 1.1267
No log 8.0 464 1.1760 0.3232 1.1760 1.0844
No log 8.0345 466 0.9843 0.3886 0.9843 0.9921
No log 8.0690 468 0.9478 0.4077 0.9478 0.9735
No log 8.1034 470 0.9655 0.3217 0.9655 0.9826
No log 8.1379 472 0.9591 0.3451 0.9591 0.9793
No log 8.1724 474 0.9908 0.3747 0.9908 0.9954
No log 8.2069 476 1.1300 0.3511 1.1300 1.0630
No log 8.2414 478 1.1679 0.3174 1.1679 1.0807
No log 8.2759 480 1.0568 0.3561 1.0568 1.0280
No log 8.3103 482 0.9436 0.4198 0.9436 0.9714
No log 8.3448 484 0.9193 0.3909 0.9193 0.9588
No log 8.3793 486 0.9377 0.4175 0.9377 0.9684
No log 8.4138 488 1.1309 0.3317 1.1309 1.0634
No log 8.4483 490 1.3273 0.2963 1.3273 1.1521
No log 8.4828 492 1.3489 0.2963 1.3489 1.1614
No log 8.5172 494 1.1372 0.3196 1.1372 1.0664
No log 8.5517 496 0.9131 0.4624 0.9131 0.9556
No log 8.5862 498 0.8920 0.5061 0.8920 0.9445
0.3389 8.6207 500 0.8984 0.4363 0.8984 0.9479
0.3389 8.6552 502 0.9038 0.4299 0.9038 0.9507
0.3389 8.6897 504 0.9491 0.4136 0.9491 0.9742
0.3389 8.7241 506 0.9580 0.4439 0.9580 0.9788
0.3389 8.7586 508 0.9381 0.4476 0.9381 0.9686
0.3389 8.7931 510 0.9449 0.4476 0.9449 0.9720
0.3389 8.8276 512 0.9478 0.4476 0.9478 0.9736
0.3389 8.8621 514 0.9536 0.4408 0.9536 0.9765
0.3389 8.8966 516 0.9631 0.4347 0.9631 0.9814
0.3389 8.9310 518 0.9616 0.3696 0.9616 0.9806
0.3389 8.9655 520 0.9930 0.3642 0.9930 0.9965
0.3389 9.0 522 0.9797 0.3642 0.9797 0.9898
0.3389 9.0345 524 0.9556 0.3692 0.9556 0.9775

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
182
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k16_task2_organization

Finetuned
(4222)
this model