ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k14_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9082
  • Qwk: 0.4624
  • Mse: 0.9082
  • Rmse: 0.9530

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0392 2 4.8579 -0.0132 4.8579 2.2041
No log 0.0784 4 3.1845 -0.0079 3.1845 1.7845
No log 0.1176 6 2.3564 -0.0608 2.3564 1.5351
No log 0.1569 8 2.0745 -0.0370 2.0745 1.4403
No log 0.1961 10 1.4409 0.0345 1.4409 1.2004
No log 0.2353 12 1.2729 0.0731 1.2729 1.1282
No log 0.2745 14 1.2906 0.0537 1.2906 1.1360
No log 0.3137 16 1.4078 0.0169 1.4078 1.1865
No log 0.3529 18 1.7255 0.0 1.7255 1.3136
No log 0.3922 20 1.7307 -0.0149 1.7307 1.3156
No log 0.4314 22 1.4704 0.0575 1.4704 1.2126
No log 0.4706 24 1.3946 0.1135 1.3946 1.1810
No log 0.5098 26 1.3715 0.0750 1.3715 1.1711
No log 0.5490 28 1.3601 0.1135 1.3601 1.1663
No log 0.5882 30 1.3956 0.0 1.3956 1.1814
No log 0.6275 32 1.5048 0.0 1.5048 1.2267
No log 0.6667 34 1.4880 0.0 1.4880 1.2198
No log 0.7059 36 1.7713 0.0844 1.7713 1.3309
No log 0.7451 38 2.1946 0.1628 2.1946 1.4814
No log 0.7843 40 2.0957 0.1457 2.0957 1.4477
No log 0.8235 42 1.7468 0.0504 1.7468 1.3217
No log 0.8627 44 1.5128 0.0766 1.5128 1.2300
No log 0.9020 46 1.6469 0.0766 1.6469 1.2833
No log 0.9412 48 1.9251 0.2132 1.9251 1.3875
No log 0.9804 50 1.8932 0.1919 1.8932 1.3759
No log 1.0196 52 1.9209 0.1470 1.9209 1.3860
No log 1.0588 54 1.8928 0.1470 1.8928 1.3758
No log 1.0980 56 1.7828 0.2054 1.7828 1.3352
No log 1.1373 58 1.4121 0.2721 1.4121 1.1883
No log 1.1765 60 1.2762 0.2202 1.2762 1.1297
No log 1.2157 62 1.0817 0.2768 1.0817 1.0400
No log 1.2549 64 1.0482 0.2891 1.0482 1.0238
No log 1.2941 66 1.0356 0.2871 1.0356 1.0177
No log 1.3333 68 1.0848 0.2467 1.0848 1.0416
No log 1.3725 70 1.1191 0.2467 1.1191 1.0579
No log 1.4118 72 1.3494 0.1884 1.3494 1.1616
No log 1.4510 74 1.9203 0.1868 1.9203 1.3857
No log 1.4902 76 2.3553 0.1404 2.3553 1.5347
No log 1.5294 78 2.3925 0.1260 2.3925 1.5468
No log 1.5686 80 2.0871 0.1967 2.0871 1.4447
No log 1.6078 82 1.6559 0.1196 1.6559 1.2868
No log 1.6471 84 1.1708 0.2424 1.1708 1.0820
No log 1.6863 86 1.0192 0.2417 1.0192 1.0096
No log 1.7255 88 0.9158 0.3960 0.9158 0.9570
No log 1.7647 90 0.9068 0.3614 0.9068 0.9523
No log 1.8039 92 1.0234 0.2890 1.0234 1.0117
No log 1.8431 94 1.3405 0.1568 1.3405 1.1578
No log 1.8824 96 1.6157 0.2158 1.6157 1.2711
No log 1.9216 98 1.7371 0.2752 1.7371 1.3180
No log 1.9608 100 1.6971 0.2497 1.6971 1.3027
No log 2.0 102 1.4267 0.2576 1.4267 1.1945
No log 2.0392 104 1.1331 0.3071 1.1331 1.0645
No log 2.0784 106 0.9501 0.3519 0.9501 0.9747
No log 2.1176 108 0.7976 0.4393 0.7976 0.8931
No log 2.1569 110 0.8009 0.4898 0.8009 0.8950
No log 2.1961 112 0.8746 0.4176 0.8746 0.9352
No log 2.2353 114 1.0472 0.4295 1.0472 1.0233
No log 2.2745 116 1.2942 0.3529 1.2942 1.1376
No log 2.3137 118 1.4514 0.3542 1.4514 1.2047
No log 2.3529 120 1.3540 0.3956 1.3540 1.1636
No log 2.3922 122 1.3021 0.3828 1.3021 1.1411
No log 2.4314 124 1.4149 0.3980 1.4149 1.1895
No log 2.4706 126 1.3855 0.3297 1.3855 1.1771
No log 2.5098 128 1.3089 0.3770 1.3089 1.1441
No log 2.5490 130 1.5980 0.3627 1.5980 1.2641
No log 2.5882 132 2.0054 0.2183 2.0054 1.4161
No log 2.6275 134 2.1628 0.1438 2.1628 1.4706
No log 2.6667 136 2.0867 0.1911 2.0867 1.4445
No log 2.7059 138 1.7773 0.2254 1.7773 1.3331
No log 2.7451 140 1.3986 0.3148 1.3986 1.1826
No log 2.7843 142 1.0669 0.2507 1.0669 1.0329
No log 2.8235 144 0.8344 0.5781 0.8344 0.9135
No log 2.8627 146 0.7671 0.5770 0.7671 0.8758
No log 2.9020 148 0.7654 0.5675 0.7654 0.8749
No log 2.9412 150 0.8101 0.5785 0.8101 0.9001
No log 2.9804 152 0.9247 0.4212 0.9247 0.9616
No log 3.0196 154 0.9695 0.3798 0.9695 0.9846
No log 3.0588 156 0.9565 0.3843 0.9565 0.9780
No log 3.0980 158 0.8786 0.5404 0.8786 0.9373
No log 3.1373 160 0.8332 0.5592 0.8332 0.9128
No log 3.1765 162 0.8546 0.5458 0.8546 0.9245
No log 3.2157 164 0.9286 0.4785 0.9286 0.9637
No log 3.2549 166 0.9160 0.4961 0.9160 0.9571
No log 3.2941 168 0.9277 0.5144 0.9277 0.9632
No log 3.3333 170 0.9059 0.5256 0.9059 0.9518
No log 3.3725 172 0.9071 0.5264 0.9071 0.9524
No log 3.4118 174 0.9083 0.5264 0.9083 0.9530
No log 3.4510 176 0.8612 0.5642 0.8612 0.9280
No log 3.4902 178 0.8790 0.5817 0.8790 0.9375
No log 3.5294 180 0.9087 0.5908 0.9087 0.9533
No log 3.5686 182 0.9047 0.5914 0.9047 0.9512
No log 3.6078 184 0.9313 0.5383 0.9313 0.9650
No log 3.6471 186 1.0624 0.4153 1.0624 1.0307
No log 3.6863 188 1.2147 0.3171 1.2147 1.1021
No log 3.7255 190 1.1910 0.3058 1.1910 1.0913
No log 3.7647 192 1.1301 0.3527 1.1301 1.0631
No log 3.8039 194 0.9367 0.5034 0.9367 0.9678
No log 3.8431 196 0.8680 0.4947 0.8680 0.9317
No log 3.8824 198 0.8759 0.5098 0.8759 0.9359
No log 3.9216 200 0.9075 0.6160 0.9075 0.9526
No log 3.9608 202 0.9048 0.5679 0.9048 0.9512
No log 4.0 204 0.9197 0.4543 0.9197 0.9590
No log 4.0392 206 0.8907 0.5127 0.8907 0.9438
No log 4.0784 208 0.8349 0.4026 0.8349 0.9137
No log 4.1176 210 0.8286 0.4343 0.8286 0.9102
No log 4.1569 212 0.8236 0.4100 0.8236 0.9075
No log 4.1961 214 0.8150 0.4662 0.8150 0.9027
No log 4.2353 216 0.8550 0.5127 0.8550 0.9246
No log 4.2745 218 0.8816 0.4976 0.8816 0.9390
No log 4.3137 220 0.8194 0.5707 0.8194 0.9052
No log 4.3529 222 0.7855 0.5230 0.7855 0.8863
No log 4.3922 224 0.8375 0.4826 0.8375 0.9152
No log 4.4314 226 0.8063 0.4998 0.8063 0.8979
No log 4.4706 228 0.8184 0.5632 0.8184 0.9047
No log 4.5098 230 0.9660 0.4627 0.9660 0.9828
No log 4.5490 232 0.9962 0.4352 0.9962 0.9981
No log 4.5882 234 0.9845 0.4850 0.9845 0.9922
No log 4.6275 236 0.9104 0.5325 0.9104 0.9542
No log 4.6667 238 0.9065 0.5404 0.9065 0.9521
No log 4.7059 240 0.9416 0.4902 0.9416 0.9703
No log 4.7451 242 0.9862 0.4739 0.9862 0.9931
No log 4.7843 244 0.9504 0.5018 0.9504 0.9749
No log 4.8235 246 0.9287 0.5150 0.9287 0.9637
No log 4.8627 248 0.8522 0.4749 0.8522 0.9231
No log 4.9020 250 0.8181 0.5011 0.8181 0.9045
No log 4.9412 252 0.8366 0.4916 0.8366 0.9147
No log 4.9804 254 0.8524 0.5283 0.8524 0.9233
No log 5.0196 256 0.8398 0.4866 0.8398 0.9164
No log 5.0588 258 0.8321 0.4995 0.8321 0.9122
No log 5.0980 260 0.8621 0.4965 0.8621 0.9285
No log 5.1373 262 0.8826 0.4757 0.8826 0.9395
No log 5.1765 264 0.8709 0.5042 0.8709 0.9332
No log 5.2157 266 0.9011 0.5433 0.9011 0.9492
No log 5.2549 268 0.9203 0.4792 0.9203 0.9593
No log 5.2941 270 0.9249 0.5236 0.9249 0.9617
No log 5.3333 272 0.8590 0.4803 0.8590 0.9268
No log 5.3725 274 0.8230 0.4318 0.8230 0.9072
No log 5.4118 276 0.8354 0.4343 0.8354 0.9140
No log 5.4510 278 0.8434 0.5230 0.8434 0.9184
No log 5.4902 280 0.8218 0.4722 0.8218 0.9065
No log 5.5294 282 0.8620 0.3571 0.8620 0.9285
No log 5.5686 284 0.9275 0.4829 0.9275 0.9631
No log 5.6078 286 0.9192 0.4596 0.9192 0.9588
No log 5.6471 288 0.8670 0.4176 0.8670 0.9311
No log 5.6863 290 0.8884 0.5361 0.8884 0.9425
No log 5.7255 292 0.8931 0.5361 0.8931 0.9450
No log 5.7647 294 0.8469 0.5582 0.8469 0.9203
No log 5.8039 296 0.8557 0.4881 0.8557 0.9250
No log 5.8431 298 0.9253 0.4431 0.9253 0.9619
No log 5.8824 300 0.8971 0.4440 0.8971 0.9472
No log 5.9216 302 0.8388 0.4982 0.8388 0.9159
No log 5.9608 304 0.8759 0.4872 0.8759 0.9359
No log 6.0 306 0.8957 0.4825 0.8957 0.9464
No log 6.0392 308 0.8575 0.4946 0.8575 0.9260
No log 6.0784 310 0.8476 0.4820 0.8476 0.9207
No log 6.1176 312 0.8689 0.4328 0.8689 0.9321
No log 6.1569 314 0.8515 0.4775 0.8515 0.9227
No log 6.1961 316 0.8025 0.4724 0.8025 0.8958
No log 6.2353 318 0.7932 0.4575 0.7932 0.8906
No log 6.2745 320 0.7976 0.4715 0.7976 0.8931
No log 6.3137 322 0.8079 0.5009 0.8079 0.8988
No log 6.3529 324 0.8218 0.5150 0.8218 0.9065
No log 6.3922 326 0.8199 0.4822 0.8199 0.9055
No log 6.4314 328 0.8342 0.4963 0.8342 0.9133
No log 6.4706 330 0.8469 0.4963 0.8469 0.9203
No log 6.5098 332 0.8629 0.4159 0.8629 0.9289
No log 6.5490 334 0.9428 0.4435 0.9428 0.9710
No log 6.5882 336 1.1477 0.4136 1.1477 1.0713
No log 6.6275 338 1.1432 0.4168 1.1432 1.0692
No log 6.6667 340 0.9902 0.4777 0.9902 0.9951
No log 6.7059 342 0.8306 0.4421 0.8306 0.9114
No log 6.7451 344 0.8004 0.5770 0.8004 0.8946
No log 6.7843 346 0.7978 0.5770 0.7978 0.8932
No log 6.8235 348 0.8098 0.5358 0.8098 0.8999
No log 6.8627 350 0.8469 0.4812 0.8469 0.9203
No log 6.9020 352 0.9262 0.4373 0.9262 0.9624
No log 6.9412 354 1.0356 0.4300 1.0356 1.0177
No log 6.9804 356 1.0046 0.3913 1.0046 1.0023
No log 7.0196 358 0.8994 0.3985 0.8994 0.9483
No log 7.0588 360 0.8589 0.4119 0.8589 0.9267
No log 7.0980 362 0.8581 0.3738 0.8581 0.9263
No log 7.1373 364 0.8743 0.4331 0.8743 0.9350
No log 7.1765 366 0.9473 0.4320 0.9473 0.9733
No log 7.2157 368 0.9674 0.4567 0.9674 0.9836
No log 7.2549 370 0.9147 0.4235 0.9147 0.9564
No log 7.2941 372 0.8363 0.5340 0.8363 0.9145
No log 7.3333 374 0.8259 0.5988 0.8259 0.9088
No log 7.3725 376 0.8418 0.5085 0.8418 0.9175
No log 7.4118 378 0.9068 0.4849 0.9068 0.9522
No log 7.4510 380 0.9644 0.4404 0.9644 0.9820
No log 7.4902 382 0.9751 0.4404 0.9751 0.9875
No log 7.5294 384 0.9626 0.4137 0.9626 0.9811
No log 7.5686 386 0.9192 0.3887 0.9192 0.9588
No log 7.6078 388 0.8772 0.4671 0.8772 0.9366
No log 7.6471 390 0.8519 0.4947 0.8519 0.9230
No log 7.6863 392 0.8582 0.5194 0.8582 0.9264
No log 7.7255 394 0.8687 0.5671 0.8687 0.9320
No log 7.7647 396 0.9265 0.4843 0.9265 0.9626
No log 7.8039 398 1.0314 0.4352 1.0314 1.0156
No log 7.8431 400 1.1510 0.3595 1.1510 1.0729
No log 7.8824 402 1.1936 0.3707 1.1936 1.0925
No log 7.9216 404 1.1248 0.3630 1.1248 1.0606
No log 7.9608 406 1.0185 0.4497 1.0185 1.0092
No log 8.0 408 0.9988 0.4273 0.9988 0.9994
No log 8.0392 410 0.9761 0.3855 0.9761 0.9880
No log 8.0784 412 0.9887 0.3855 0.9887 0.9943
No log 8.1176 414 1.0097 0.3942 1.0097 1.0048
No log 8.1569 416 0.9972 0.3942 0.9972 0.9986
No log 8.1961 418 0.9576 0.3463 0.9576 0.9786
No log 8.2353 420 0.9638 0.3983 0.9638 0.9818
No log 8.2745 422 0.9675 0.4224 0.9675 0.9836
No log 8.3137 424 0.9696 0.4011 0.9696 0.9847
No log 8.3529 426 1.0254 0.3912 1.0254 1.0126
No log 8.3922 428 1.0108 0.3874 1.0108 1.0054
No log 8.4314 430 0.9548 0.4138 0.9548 0.9772
No log 8.4706 432 0.9179 0.4575 0.9179 0.9580
No log 8.5098 434 0.8979 0.3164 0.8979 0.9476
No log 8.5490 436 0.9105 0.3868 0.9105 0.9542
No log 8.5882 438 0.9148 0.3868 0.9148 0.9564
No log 8.6275 440 0.8982 0.3263 0.8982 0.9477
No log 8.6667 442 0.9287 0.4176 0.9287 0.9637
No log 8.7059 444 0.9884 0.4518 0.9884 0.9942
No log 8.7451 446 0.9598 0.4763 0.9598 0.9797
No log 8.7843 448 0.9422 0.4425 0.9422 0.9707
No log 8.8235 450 0.8934 0.3935 0.8934 0.9452
No log 8.8627 452 0.8879 0.3868 0.8879 0.9423
No log 8.9020 454 0.8952 0.4337 0.8952 0.9461
No log 8.9412 456 0.8855 0.4337 0.8855 0.9410
No log 8.9804 458 0.8584 0.4006 0.8584 0.9265
No log 9.0196 460 0.8586 0.3948 0.8586 0.9266
No log 9.0588 462 0.8581 0.3948 0.8581 0.9264
No log 9.0980 464 0.8564 0.3762 0.8564 0.9254
No log 9.1373 466 0.8609 0.3762 0.8609 0.9278
No log 9.1765 468 0.8636 0.4104 0.8636 0.9293
No log 9.2157 470 0.8576 0.3762 0.8576 0.9261
No log 9.2549 472 0.8588 0.4084 0.8588 0.9267
No log 9.2941 474 0.9010 0.5327 0.9010 0.9492
No log 9.3333 476 0.9391 0.5568 0.9391 0.9691
No log 9.3725 478 0.9425 0.5228 0.9425 0.9708
No log 9.4118 480 0.8842 0.5127 0.8842 0.9403
No log 9.4510 482 0.8399 0.4479 0.8399 0.9165
No log 9.4902 484 0.8251 0.4218 0.8251 0.9083
No log 9.5294 486 0.8240 0.4444 0.8240 0.9078
No log 9.5686 488 0.8245 0.4048 0.8245 0.9080
No log 9.6078 490 0.8407 0.4465 0.8407 0.9169
No log 9.6471 492 0.8672 0.5046 0.8672 0.9313
No log 9.6863 494 0.8623 0.5094 0.8623 0.9286
No log 9.7255 496 0.8259 0.4240 0.8259 0.9088
No log 9.7647 498 0.8153 0.4444 0.8153 0.9029
0.3757 9.8039 500 0.8413 0.4866 0.8413 0.9172
0.3757 9.8431 502 0.8436 0.4885 0.8436 0.9185
0.3757 9.8824 504 0.8370 0.4411 0.8370 0.9149
0.3757 9.9216 506 0.8466 0.4411 0.8466 0.9201
0.3757 9.9608 508 0.8620 0.4196 0.8620 0.9284
0.3757 10.0 510 0.8963 0.4672 0.8963 0.9467
0.3757 10.0392 512 0.9082 0.4624 0.9082 0.9530

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
183
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k14_task2_organization

Finetuned
(4222)
this model