ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k19_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9056
  • Qwk: 0.3847
  • Mse: 0.9056
  • Rmse: 0.9516

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0294 2 4.8810 -0.0075 4.8810 2.2093
No log 0.0588 4 3.1622 -0.0384 3.1622 1.7783
No log 0.0882 6 2.1046 -0.0164 2.1046 1.4507
No log 0.1176 8 2.0402 -0.0303 2.0402 1.4284
No log 0.1471 10 1.9710 0.0062 1.9710 1.4039
No log 0.1765 12 1.5973 0.0372 1.5973 1.2638
No log 0.2059 14 1.4861 -0.0146 1.4861 1.2190
No log 0.2353 16 1.4096 -0.0368 1.4096 1.1873
No log 0.2647 18 1.3129 0.0613 1.3129 1.1458
No log 0.2941 20 1.3341 0.0613 1.3341 1.1550
No log 0.3235 22 1.5299 0.0818 1.5299 1.2369
No log 0.3529 24 2.1341 -0.1062 2.1341 1.4609
No log 0.3824 26 2.3953 -0.0685 2.3953 1.5477
No log 0.4118 28 1.9860 -0.0156 1.9860 1.4092
No log 0.4412 30 1.5857 0.0936 1.5857 1.2592
No log 0.4706 32 1.3531 0.1530 1.3531 1.1632
No log 0.5 34 1.4390 0.0838 1.4390 1.1996
No log 0.5294 36 1.4827 0.0898 1.4827 1.2177
No log 0.5588 38 1.3633 0.1165 1.3633 1.1676
No log 0.5882 40 1.3018 0.2014 1.3018 1.1410
No log 0.6176 42 1.3913 0.0537 1.3913 1.1795
No log 0.6471 44 1.6820 0.1064 1.6820 1.2969
No log 0.6765 46 1.7718 0.0381 1.7718 1.3311
No log 0.7059 48 1.8220 0.0309 1.8220 1.3498
No log 0.7353 50 1.5279 0.0575 1.5279 1.2361
No log 0.7647 52 1.2984 0.2214 1.2984 1.1395
No log 0.7941 54 1.2493 0.1974 1.2493 1.1177
No log 0.8235 56 1.4224 0.0898 1.4224 1.1927
No log 0.8529 58 1.6074 0.0972 1.6074 1.2678
No log 0.8824 60 1.9977 0.2440 1.9977 1.4134
No log 0.9118 62 2.1767 0.2250 2.1767 1.4754
No log 0.9412 64 2.1273 0.2410 2.1273 1.4585
No log 0.9706 66 1.8999 0.2356 1.8999 1.3784
No log 1.0 68 1.6042 0.1395 1.6042 1.2666
No log 1.0294 70 1.3335 0.0838 1.3335 1.1548
No log 1.0588 72 1.2043 0.1530 1.2043 1.0974
No log 1.0882 74 1.1156 0.2395 1.1156 1.0562
No log 1.1176 76 1.1261 0.2577 1.1261 1.0612
No log 1.1471 78 1.1806 0.2243 1.1806 1.0865
No log 1.1765 80 1.2470 0.1675 1.2470 1.1167
No log 1.2059 82 1.3386 0.0784 1.3386 1.1570
No log 1.2353 84 1.3361 0.1141 1.3361 1.1559
No log 1.2647 86 1.4851 0.1535 1.4851 1.2187
No log 1.2941 88 1.7205 0.2312 1.7205 1.3117
No log 1.3235 90 1.6493 0.2655 1.6493 1.2843
No log 1.3529 92 1.2980 0.1601 1.2980 1.1393
No log 1.3824 94 1.1446 0.2191 1.1446 1.0698
No log 1.4118 96 1.1082 0.2718 1.1082 1.0527
No log 1.4412 98 1.1665 0.2191 1.1665 1.0800
No log 1.4706 100 1.3960 0.1902 1.3960 1.1815
No log 1.5 102 1.5474 0.2567 1.5474 1.2439
No log 1.5294 104 1.6052 0.3316 1.6052 1.2670
No log 1.5588 106 1.3406 0.1785 1.3406 1.1579
No log 1.5882 108 1.1422 0.2245 1.1422 1.0688
No log 1.6176 110 1.1215 0.2716 1.1215 1.0590
No log 1.6471 112 1.0297 0.3615 1.0297 1.0147
No log 1.6765 114 0.9993 0.3714 0.9993 0.9996
No log 1.7059 116 1.0533 0.3546 1.0533 1.0263
No log 1.7353 118 1.1816 0.2789 1.1816 1.0870
No log 1.7647 120 1.2163 0.2876 1.2163 1.1028
No log 1.7941 122 1.2450 0.2789 1.2450 1.1158
No log 1.8235 124 1.2303 0.2718 1.2303 1.1092
No log 1.8529 126 1.3020 0.2773 1.3020 1.1411
No log 1.8824 128 1.3664 0.2545 1.3664 1.1689
No log 1.9118 130 1.3415 0.2108 1.3415 1.1582
No log 1.9412 132 1.2316 0.3028 1.2316 1.1098
No log 1.9706 134 1.1812 0.2333 1.1812 1.0868
No log 2.0 136 1.1906 0.2280 1.1906 1.0912
No log 2.0294 138 1.1121 0.2549 1.1121 1.0545
No log 2.0588 140 1.0022 0.4 1.0022 1.0011
No log 2.0882 142 0.9726 0.3352 0.9726 0.9862
No log 2.1176 144 1.0845 0.3111 1.0845 1.0414
No log 2.1471 146 1.1426 0.3201 1.1426 1.0689
No log 2.1765 148 1.0391 0.3734 1.0391 1.0194
No log 2.2059 150 0.9303 0.4316 0.9303 0.9645
No log 2.2353 152 0.9546 0.4652 0.9546 0.9770
No log 2.2647 154 1.0319 0.3805 1.0319 1.0158
No log 2.2941 156 1.0518 0.3650 1.0518 1.0256
No log 2.3235 158 1.0001 0.3399 1.0001 1.0000
No log 2.3529 160 0.9388 0.4489 0.9388 0.9689
No log 2.3824 162 0.9530 0.4830 0.9530 0.9762
No log 2.4118 164 0.9301 0.5466 0.9301 0.9644
No log 2.4412 166 0.9336 0.4477 0.9336 0.9662
No log 2.4706 168 0.9690 0.4607 0.9690 0.9844
No log 2.5 170 0.9611 0.4731 0.9611 0.9804
No log 2.5294 172 0.9319 0.4290 0.9319 0.9653
No log 2.5588 174 0.9287 0.3155 0.9287 0.9637
No log 2.5882 176 0.9440 0.3758 0.9440 0.9716
No log 2.6176 178 0.9731 0.4304 0.9731 0.9864
No log 2.6471 180 0.9922 0.4196 0.9922 0.9961
No log 2.6765 182 1.1396 0.4186 1.1396 1.0675
No log 2.7059 184 1.2867 0.4429 1.2867 1.1343
No log 2.7353 186 1.1530 0.3967 1.1530 1.0738
No log 2.7647 188 1.0222 0.3368 1.0222 1.0110
No log 2.7941 190 0.9776 0.4808 0.9776 0.9888
No log 2.8235 192 0.9739 0.4808 0.9739 0.9869
No log 2.8529 194 0.9657 0.4493 0.9657 0.9827
No log 2.8824 196 0.9875 0.3476 0.9875 0.9937
No log 2.9118 198 0.9709 0.4618 0.9709 0.9853
No log 2.9412 200 0.9683 0.4318 0.9683 0.9840
No log 2.9706 202 0.9532 0.4618 0.9532 0.9763
No log 3.0 204 0.9458 0.3892 0.9458 0.9725
No log 3.0294 206 1.0534 0.4320 1.0534 1.0263
No log 3.0588 208 1.0890 0.3833 1.0890 1.0436
No log 3.0882 210 0.9438 0.3243 0.9438 0.9715
No log 3.1176 212 0.8939 0.4089 0.8939 0.9455
No log 3.1471 214 0.9705 0.4845 0.9705 0.9851
No log 3.1765 216 0.9089 0.4328 0.9089 0.9533
No log 3.2059 218 0.9260 0.4042 0.9260 0.9623
No log 3.2353 220 1.2044 0.4022 1.2044 1.0975
No log 3.2647 222 1.2406 0.4057 1.2406 1.1138
No log 3.2941 224 0.9922 0.4431 0.9922 0.9961
No log 3.3235 226 0.9383 0.4559 0.9383 0.9686
No log 3.3529 228 1.0431 0.4251 1.0431 1.0213
No log 3.3824 230 0.9886 0.4607 0.9886 0.9943
No log 3.4118 232 0.8903 0.4037 0.8903 0.9436
No log 3.4412 234 0.9841 0.4341 0.9841 0.9920
No log 3.4706 236 1.0144 0.4172 1.0144 1.0072
No log 3.5 238 0.9476 0.4156 0.9476 0.9735
No log 3.5294 240 0.8913 0.3692 0.8913 0.9441
No log 3.5588 242 0.8986 0.3796 0.8986 0.9479
No log 3.5882 244 0.9028 0.3938 0.9028 0.9502
No log 3.6176 246 0.8943 0.3979 0.8943 0.9456
No log 3.6471 248 0.9075 0.3773 0.9075 0.9526
No log 3.6765 250 0.8852 0.3969 0.8852 0.9409
No log 3.7059 252 0.8824 0.4626 0.8824 0.9394
No log 3.7353 254 0.8816 0.4877 0.8816 0.9389
No log 3.7647 256 0.8935 0.4829 0.8935 0.9453
No log 3.7941 258 0.9015 0.4321 0.9015 0.9495
No log 3.8235 260 0.9729 0.4068 0.9729 0.9863
No log 3.8529 262 0.9958 0.4524 0.9958 0.9979
No log 3.8824 264 0.9201 0.3820 0.9201 0.9592
No log 3.9118 266 0.8863 0.4218 0.8863 0.9414
No log 3.9412 268 0.8869 0.3650 0.8869 0.9418
No log 3.9706 270 0.9150 0.3463 0.9150 0.9565
No log 4.0 272 1.0122 0.4492 1.0122 1.0061
No log 4.0294 274 1.0499 0.4431 1.0499 1.0246
No log 4.0588 276 0.9801 0.3775 0.9801 0.9900
No log 4.0882 278 0.9294 0.3224 0.9294 0.9641
No log 4.1176 280 0.9570 0.3564 0.9570 0.9783
No log 4.1471 282 0.9924 0.3786 0.9924 0.9962
No log 4.1765 284 0.9484 0.3590 0.9484 0.9739
No log 4.2059 286 0.9174 0.3468 0.9174 0.9578
No log 4.2353 288 0.9282 0.3134 0.9282 0.9634
No log 4.2647 290 0.9414 0.3468 0.9414 0.9703
No log 4.2941 292 0.9424 0.3039 0.9424 0.9708
No log 4.3235 294 0.9482 0.3089 0.9482 0.9738
No log 4.3529 296 0.9455 0.3511 0.9455 0.9724
No log 4.3824 298 0.9539 0.3255 0.9539 0.9767
No log 4.4118 300 0.9518 0.3194 0.9518 0.9756
No log 4.4412 302 0.9418 0.3217 0.9418 0.9704
No log 4.4706 304 0.9851 0.3820 0.9851 0.9925
No log 4.5 306 0.9815 0.3564 0.9815 0.9907
No log 4.5294 308 0.9553 0.2944 0.9553 0.9774
No log 4.5588 310 0.9622 0.2728 0.9622 0.9809
No log 4.5882 312 0.9601 0.2534 0.9601 0.9799
No log 4.6176 314 0.9712 0.3564 0.9712 0.9855
No log 4.6471 316 0.9822 0.3564 0.9822 0.9910
No log 4.6765 318 0.9578 0.3854 0.9578 0.9787
No log 4.7059 320 0.9552 0.2850 0.9552 0.9774
No log 4.7353 322 0.9593 0.3036 0.9593 0.9794
No log 4.7647 324 0.9844 0.2832 0.9844 0.9922
No log 4.7941 326 1.0426 0.3453 1.0426 1.0211
No log 4.8235 328 1.0655 0.3142 1.0655 1.0322
No log 4.8529 330 1.0071 0.2896 1.0071 1.0036
No log 4.8824 332 0.9550 0.3066 0.9550 0.9772
No log 4.9118 334 0.9724 0.4062 0.9724 0.9861
No log 4.9412 336 0.9644 0.3169 0.9644 0.9820
No log 4.9706 338 0.9483 0.3196 0.9483 0.9738
No log 5.0 340 0.9703 0.3298 0.9703 0.9851
No log 5.0294 342 0.9792 0.2916 0.9792 0.9896
No log 5.0588 344 0.9885 0.3020 0.9885 0.9942
No log 5.0882 346 1.0059 0.3564 1.0059 1.0030
No log 5.1176 348 0.9742 0.3957 0.9742 0.9870
No log 5.1471 350 0.9615 0.3854 0.9615 0.9805
No log 5.1765 352 0.9437 0.3896 0.9437 0.9715
No log 5.2059 354 0.9424 0.3930 0.9424 0.9708
No log 5.2353 356 0.9417 0.4254 0.9417 0.9704
No log 5.2647 358 0.9471 0.4444 0.9471 0.9732
No log 5.2941 360 0.9515 0.4254 0.9515 0.9755
No log 5.3235 362 0.9594 0.2943 0.9594 0.9795
No log 5.3529 364 0.9593 0.3381 0.9593 0.9794
No log 5.3824 366 0.9272 0.2730 0.9272 0.9629
No log 5.4118 368 0.9179 0.3961 0.9179 0.9580
No log 5.4412 370 0.9414 0.4062 0.9414 0.9702
No log 5.4706 372 0.9098 0.3551 0.9098 0.9539
No log 5.5 374 0.8886 0.3223 0.8886 0.9426
No log 5.5294 376 0.8906 0.3223 0.8906 0.9437
No log 5.5588 378 0.8847 0.4242 0.8847 0.9406
No log 5.5882 380 0.8932 0.4439 0.8932 0.9451
No log 5.6176 382 0.8724 0.4963 0.8724 0.9340
No log 5.6471 384 0.8636 0.5415 0.8637 0.9293
No log 5.6765 386 0.8548 0.5524 0.8548 0.9246
No log 5.7059 388 0.8894 0.4429 0.8894 0.9431
No log 5.7353 390 0.9854 0.4259 0.9854 0.9927
No log 5.7647 392 0.9678 0.4622 0.9678 0.9838
No log 5.7941 394 0.8955 0.5270 0.8955 0.9463
No log 5.8235 396 0.9279 0.3813 0.9279 0.9633
No log 5.8529 398 1.0477 0.4435 1.0477 1.0236
No log 5.8824 400 1.0140 0.4435 1.0140 1.0070
No log 5.9118 402 0.9191 0.3913 0.9191 0.9587
No log 5.9412 404 0.9054 0.5322 0.9054 0.9515
No log 5.9706 406 0.9142 0.3854 0.9142 0.9561
No log 6.0 408 0.9184 0.3596 0.9184 0.9584
No log 6.0294 410 0.9248 0.3223 0.9248 0.9616
No log 6.0588 412 0.9272 0.3822 0.9272 0.9629
No log 6.0882 414 0.9272 0.3753 0.9272 0.9629
No log 6.1176 416 0.9066 0.3615 0.9066 0.9522
No log 6.1471 418 0.8964 0.3762 0.8964 0.9468
No log 6.1765 420 0.8813 0.3310 0.8813 0.9388
No log 6.2059 422 0.8653 0.4100 0.8653 0.9302
No log 6.2353 424 0.8968 0.3767 0.8968 0.9470
No log 6.2647 426 0.9364 0.3881 0.9364 0.9677
No log 6.2941 428 0.8879 0.3709 0.8879 0.9423
No log 6.3235 430 0.8880 0.3709 0.8880 0.9423
No log 6.3529 432 0.9216 0.3601 0.9216 0.9600
No log 6.3824 434 1.0199 0.4171 1.0199 1.0099
No log 6.4118 436 1.0097 0.4135 1.0097 1.0048
No log 6.4412 438 0.8899 0.3250 0.8899 0.9434
No log 6.4706 440 0.8477 0.3884 0.8477 0.9207
No log 6.5 442 0.8576 0.3744 0.8576 0.9261
No log 6.5294 444 0.8714 0.3427 0.8714 0.9335
No log 6.5588 446 0.9172 0.3989 0.9172 0.9577
No log 6.5882 448 0.9370 0.3989 0.9370 0.9680
No log 6.6176 450 0.8967 0.3829 0.8967 0.9470
No log 6.6471 452 0.9255 0.4539 0.9255 0.9620
No log 6.6765 454 0.9665 0.3841 0.9665 0.9831
No log 6.7059 456 0.9446 0.4300 0.9446 0.9719
No log 6.7353 458 0.8866 0.3998 0.8866 0.9416
No log 6.7647 460 0.8891 0.4260 0.8891 0.9429
No log 6.7941 462 0.8846 0.4012 0.8846 0.9406
No log 6.8235 464 0.8665 0.3787 0.8665 0.9309
No log 6.8529 466 0.8653 0.3998 0.8653 0.9302
No log 6.8824 468 0.8660 0.3700 0.8660 0.9306
No log 6.9118 470 0.8467 0.3890 0.8467 0.9201
No log 6.9412 472 0.8335 0.4079 0.8335 0.9130
No log 6.9706 474 0.8325 0.4181 0.8325 0.9124
No log 7.0 476 0.8413 0.4181 0.8413 0.9172
No log 7.0294 478 0.8559 0.4100 0.8559 0.9251
No log 7.0588 480 0.8994 0.4444 0.8994 0.9484
No log 7.0882 482 0.9572 0.4398 0.9572 0.9784
No log 7.1176 484 0.9426 0.3169 0.9426 0.9709
No log 7.1471 486 0.9126 0.3641 0.9126 0.9553
No log 7.1765 488 0.9773 0.3930 0.9773 0.9886
No log 7.2059 490 1.0860 0.3511 1.0860 1.0421
No log 7.2353 492 1.0774 0.3511 1.0774 1.0380
No log 7.2647 494 0.9773 0.3930 0.9773 0.9886
No log 7.2941 496 0.8984 0.3695 0.8984 0.9479
No log 7.3235 498 0.9040 0.4444 0.9040 0.9508
0.362 7.3529 500 0.9363 0.4398 0.9363 0.9676
0.362 7.3824 502 0.9065 0.3666 0.9065 0.9521
0.362 7.4118 504 0.8668 0.3896 0.8668 0.9310
0.362 7.4412 506 0.8714 0.3627 0.8714 0.9335
0.362 7.4706 508 0.8791 0.3090 0.8791 0.9376
0.362 7.5 510 0.9056 0.3847 0.9056 0.9516

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
184
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k19_task2_organization

Finetuned
(4222)
this model