ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k15_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5827
  • Qwk: 0.3354
  • Mse: 0.5827
  • Rmse: 0.7633

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0526 2 2.4742 -0.0262 2.4742 1.5730
No log 0.1053 4 1.3467 0.0423 1.3467 1.1605
No log 0.1579 6 0.9517 -0.0622 0.9517 0.9756
No log 0.2105 8 0.9130 0.1508 0.9130 0.9555
No log 0.2632 10 0.9536 0.1180 0.9536 0.9765
No log 0.3158 12 0.9793 0.1584 0.9793 0.9896
No log 0.3684 14 0.7882 -0.0149 0.7882 0.8878
No log 0.4211 16 0.7247 0.0 0.7247 0.8513
No log 0.4737 18 0.8137 0.2156 0.8137 0.9020
No log 0.5263 20 0.8123 0.1786 0.8123 0.9013
No log 0.5789 22 0.8021 0.0 0.8021 0.8956
No log 0.6316 24 0.7675 -0.0027 0.7675 0.8761
No log 0.6842 26 0.6940 0.1287 0.6940 0.8331
No log 0.7368 28 0.7868 0.1372 0.7868 0.8870
No log 0.7895 30 0.8928 0.3444 0.8928 0.9449
No log 0.8421 32 0.7854 0.0944 0.7854 0.8862
No log 0.8947 34 0.7197 0.1327 0.7197 0.8483
No log 0.9474 36 0.7623 0.1327 0.7623 0.8731
No log 1.0 38 0.7829 0.2128 0.7829 0.8848
No log 1.0526 40 0.7623 0.2412 0.7623 0.8731
No log 1.1053 42 0.7603 0.1942 0.7603 0.8720
No log 1.1579 44 0.7794 0.3019 0.7794 0.8829
No log 1.2105 46 0.9561 0.1300 0.9561 0.9778
No log 1.2632 48 1.0133 0.0390 1.0133 1.0066
No log 1.3158 50 0.8659 0.2642 0.8659 0.9306
No log 1.3684 52 0.7905 0.3050 0.7905 0.8891
No log 1.4211 54 0.7617 0.1710 0.7617 0.8728
No log 1.4737 56 0.7881 0.3477 0.7881 0.8878
No log 1.5263 58 0.7931 0.3125 0.7931 0.8906
No log 1.5789 60 0.7634 0.2815 0.7634 0.8737
No log 1.6316 62 0.7977 0.1988 0.7977 0.8931
No log 1.6842 64 0.7782 0.2540 0.7782 0.8822
No log 1.7368 66 0.7562 0.2711 0.7562 0.8696
No log 1.7895 68 0.7734 0.1935 0.7734 0.8794
No log 1.8421 70 0.7374 0.1935 0.7374 0.8587
No log 1.8947 72 0.6690 0.2325 0.6690 0.8179
No log 1.9474 74 0.7951 0.2233 0.7951 0.8917
No log 2.0 76 0.7147 0.3299 0.7147 0.8454
No log 2.0526 78 0.6206 0.4111 0.6206 0.7878
No log 2.1053 80 0.8465 0.3134 0.8465 0.9201
No log 2.1579 82 1.0390 0.4186 1.0390 1.0193
No log 2.2105 84 0.9510 0.3782 0.9510 0.9752
No log 2.2632 86 0.6268 0.3524 0.6268 0.7917
No log 2.3158 88 0.6908 0.4677 0.6908 0.8312
No log 2.3684 90 0.7447 0.4018 0.7447 0.8630
No log 2.4211 92 0.6870 0.3125 0.6870 0.8288
No log 2.4737 94 0.6383 0.3228 0.6383 0.7989
No log 2.5263 96 0.7360 0.3099 0.7360 0.8579
No log 2.5789 98 0.8811 0.3234 0.8811 0.9386
No log 2.6316 100 0.8034 0.3099 0.8034 0.8963
No log 2.6842 102 0.7284 0.1752 0.7284 0.8535
No log 2.7368 104 0.7980 0.2846 0.7980 0.8933
No log 2.7895 106 0.8182 0.2435 0.8182 0.9045
No log 2.8421 108 0.7762 0.2738 0.7762 0.8810
No log 2.8947 110 0.7566 0.2738 0.7566 0.8698
No log 2.9474 112 0.7072 0.3001 0.7072 0.8410
No log 3.0 114 0.6775 0.2970 0.6775 0.8231
No log 3.0526 116 0.6462 0.3215 0.6462 0.8038
No log 3.1053 118 0.5811 0.3633 0.5811 0.7623
No log 3.1579 120 0.6644 0.4190 0.6644 0.8151
No log 3.2105 122 0.6913 0.3942 0.6913 0.8314
No log 3.2632 124 0.5908 0.3809 0.5908 0.7686
No log 3.3158 126 0.7371 0.4652 0.7371 0.8585
No log 3.3684 128 0.8727 0.3425 0.8727 0.9342
No log 3.4211 130 0.7783 0.3827 0.7783 0.8822
No log 3.4737 132 0.6236 0.4817 0.6236 0.7897
No log 3.5263 134 0.7051 0.4270 0.7051 0.8397
No log 3.5789 136 0.9443 0.3052 0.9443 0.9718
No log 3.6316 138 0.9926 0.3052 0.9926 0.9963
No log 3.6842 140 0.9194 0.2830 0.9194 0.9589
No log 3.7368 142 0.8147 0.3822 0.8147 0.9026
No log 3.7895 144 0.6964 0.4182 0.6964 0.8345
No log 3.8421 146 0.6642 0.3530 0.6642 0.8150
No log 3.8947 148 0.6608 0.3728 0.6608 0.8129
No log 3.9474 150 0.6710 0.3754 0.6710 0.8191
No log 4.0 152 0.7809 0.3653 0.7809 0.8837
No log 4.0526 154 1.0638 0.2872 1.0638 1.0314
No log 4.1053 156 1.1561 0.2392 1.1561 1.0752
No log 4.1579 158 1.0542 0.2756 1.0542 1.0268
No log 4.2105 160 0.8145 0.2904 0.8145 0.9025
No log 4.2632 162 0.6302 0.4111 0.6302 0.7938
No log 4.3158 164 0.6788 0.3902 0.6788 0.8239
No log 4.3684 166 0.7493 0.4708 0.7493 0.8656
No log 4.4211 168 0.7314 0.4690 0.7314 0.8552
No log 4.4737 170 0.6900 0.4478 0.6900 0.8307
No log 4.5263 172 0.7100 0.4413 0.7100 0.8426
No log 4.5789 174 0.7475 0.3503 0.7475 0.8646
No log 4.6316 176 0.7553 0.3503 0.7553 0.8691
No log 4.6842 178 0.7432 0.4270 0.7432 0.8621
No log 4.7368 180 0.7022 0.3116 0.7022 0.8380
No log 4.7895 182 0.7344 0.4190 0.7344 0.8570
No log 4.8421 184 0.8381 0.2958 0.8381 0.9155
No log 4.8947 186 0.8743 0.3097 0.8743 0.9350
No log 4.9474 188 0.8019 0.3036 0.8019 0.8955
No log 5.0 190 0.6537 0.5032 0.6537 0.8085
No log 5.0526 192 0.7095 0.4652 0.7095 0.8423
No log 5.1053 194 0.8942 0.3185 0.8942 0.9456
No log 5.1579 196 0.8896 0.3185 0.8896 0.9432
No log 5.2105 198 0.6909 0.4429 0.6909 0.8312
No log 5.2632 200 0.6416 0.4523 0.6416 0.8010
No log 5.3158 202 0.7933 0.3219 0.7933 0.8907
No log 5.3684 204 0.8430 0.3012 0.8430 0.9182
No log 5.4211 206 0.7460 0.3480 0.7460 0.8637
No log 5.4737 208 0.6201 0.4322 0.6201 0.7874
No log 5.5263 210 0.6676 0.4949 0.6676 0.8171
No log 5.5789 212 0.6688 0.4713 0.6688 0.8178
No log 5.6316 214 0.6173 0.4484 0.6173 0.7857
No log 5.6842 216 0.6435 0.4081 0.6435 0.8022
No log 5.7368 218 0.6934 0.3609 0.6934 0.8327
No log 5.7895 220 0.6635 0.3677 0.6635 0.8146
No log 5.8421 222 0.6301 0.5095 0.6301 0.7938
No log 5.8947 224 0.6350 0.4590 0.6350 0.7969
No log 5.9474 226 0.6768 0.4199 0.6768 0.8227
No log 6.0 228 0.7713 0.3432 0.7713 0.8782
No log 6.0526 230 0.7957 0.3827 0.7957 0.8920
No log 6.1053 232 0.7240 0.3615 0.7240 0.8509
No log 6.1579 234 0.6286 0.5110 0.6286 0.7929
No log 6.2105 236 0.6117 0.4923 0.6117 0.7821
No log 6.2632 238 0.6096 0.4762 0.6096 0.7808
No log 6.3158 240 0.6084 0.4762 0.6084 0.7800
No log 6.3684 242 0.6253 0.4611 0.6253 0.7908
No log 6.4211 244 0.6681 0.4468 0.6681 0.8174
No log 6.4737 246 0.6616 0.4351 0.6616 0.8134
No log 6.5263 248 0.6492 0.4161 0.6492 0.8057
No log 6.5789 250 0.6595 0.3915 0.6595 0.8121
No log 6.6316 252 0.6643 0.4656 0.6643 0.8151
No log 6.6842 254 0.6725 0.4285 0.6725 0.8201
No log 6.7368 256 0.6730 0.4285 0.6730 0.8204
No log 6.7895 258 0.6613 0.4516 0.6613 0.8132
No log 6.8421 260 0.6864 0.3640 0.6864 0.8285
No log 6.8947 262 0.7230 0.3653 0.7230 0.8503
No log 6.9474 264 0.8123 0.3394 0.8123 0.9013
No log 7.0 266 0.8276 0.3394 0.8276 0.9097
No log 7.0526 268 0.7615 0.3394 0.7615 0.8726
No log 7.1053 270 0.6895 0.4134 0.6895 0.8304
No log 7.1579 272 0.6898 0.4134 0.6898 0.8305
No log 7.2105 274 0.7241 0.4134 0.7241 0.8510
No log 7.2632 276 0.7271 0.4029 0.7271 0.8527
No log 7.3158 278 0.7505 0.3183 0.7505 0.8663
No log 7.3684 280 0.7760 0.3394 0.7760 0.8809
No log 7.4211 282 0.7468 0.3319 0.7468 0.8642
No log 7.4737 284 0.7255 0.3699 0.7255 0.8517
No log 7.5263 286 0.6792 0.3425 0.6792 0.8241
No log 7.5789 288 0.6720 0.4514 0.6720 0.8197
No log 7.6316 290 0.6707 0.4455 0.6707 0.8190
No log 7.6842 292 0.7229 0.5098 0.7229 0.8502
No log 7.7368 294 0.8001 0.3239 0.8001 0.8945
No log 7.7895 296 0.7768 0.4077 0.7768 0.8813
No log 7.8421 298 0.6694 0.5937 0.6694 0.8182
No log 7.8947 300 0.6446 0.4334 0.6446 0.8029
No log 7.9474 302 0.7535 0.3892 0.7535 0.8681
No log 8.0 304 0.8514 0.4250 0.8514 0.9227
No log 8.0526 306 0.8075 0.4032 0.8075 0.8986
No log 8.1053 308 0.6964 0.4825 0.6964 0.8345
No log 8.1579 310 0.6299 0.4536 0.6299 0.7937
No log 8.2105 312 0.6410 0.5003 0.6410 0.8007
No log 8.2632 314 0.6426 0.5167 0.6426 0.8016
No log 8.3158 316 0.6348 0.4137 0.6348 0.7967
No log 8.3684 318 0.6200 0.4253 0.6200 0.7874
No log 8.4211 320 0.6143 0.4308 0.6143 0.7838
No log 8.4737 322 0.6103 0.4495 0.6103 0.7812
No log 8.5263 324 0.6705 0.4315 0.6705 0.8188
No log 8.5789 326 0.8006 0.3312 0.8006 0.8948
No log 8.6316 328 0.8928 0.3735 0.8928 0.9449
No log 8.6842 330 0.8453 0.3632 0.8453 0.9194
No log 8.7368 332 0.7150 0.3754 0.7150 0.8456
No log 8.7895 334 0.6515 0.4218 0.6515 0.8072
No log 8.8421 336 0.5915 0.5272 0.5915 0.7691
No log 8.8947 338 0.5672 0.5177 0.5672 0.7531
No log 8.9474 340 0.5655 0.4722 0.5655 0.7520
No log 9.0 342 0.5624 0.4776 0.5624 0.7499
No log 9.0526 344 0.5802 0.5692 0.5802 0.7617
No log 9.1053 346 0.6370 0.5500 0.6370 0.7981
No log 9.1579 348 0.6974 0.4686 0.6974 0.8351
No log 9.2105 350 0.7002 0.4686 0.7002 0.8368
No log 9.2632 352 0.6332 0.5696 0.6332 0.7958
No log 9.3158 354 0.5675 0.5678 0.5675 0.7533
No log 9.3684 356 0.5481 0.4937 0.5481 0.7404
No log 9.4211 358 0.5545 0.4314 0.5545 0.7447
No log 9.4737 360 0.5375 0.4768 0.5375 0.7331
No log 9.5263 362 0.5279 0.4484 0.5279 0.7265
No log 9.5789 364 0.5283 0.4127 0.5283 0.7268
No log 9.6316 366 0.5288 0.4217 0.5288 0.7272
No log 9.6842 368 0.5364 0.4019 0.5364 0.7324
No log 9.7368 370 0.5599 0.4027 0.5599 0.7483
No log 9.7895 372 0.6190 0.4597 0.6190 0.7868
No log 9.8421 374 0.6763 0.4272 0.6763 0.8224
No log 9.8947 376 0.6970 0.4272 0.6970 0.8349
No log 9.9474 378 0.6467 0.4597 0.6467 0.8042
No log 10.0 380 0.5988 0.4013 0.5988 0.7738
No log 10.0526 382 0.5846 0.4484 0.5846 0.7646
No log 10.1053 384 0.5987 0.4463 0.5987 0.7738
No log 10.1579 386 0.6253 0.4201 0.6253 0.7907
No log 10.2105 388 0.7074 0.4738 0.7074 0.8411
No log 10.2632 390 0.8107 0.4018 0.8107 0.9004
No log 10.3158 392 0.7848 0.3913 0.7848 0.8859
No log 10.3684 394 0.7276 0.4349 0.7276 0.8530
No log 10.4211 396 0.6852 0.4819 0.6852 0.8278
No log 10.4737 398 0.6185 0.3933 0.6185 0.7865
No log 10.5263 400 0.5995 0.4262 0.5995 0.7743
No log 10.5789 402 0.5925 0.4423 0.5925 0.7697
No log 10.6316 404 0.6125 0.3974 0.6125 0.7827
No log 10.6842 406 0.6715 0.4819 0.6715 0.8195
No log 10.7368 408 0.6650 0.4144 0.6650 0.8155
No log 10.7895 410 0.6117 0.4618 0.6117 0.7821
No log 10.8421 412 0.5883 0.5131 0.5883 0.7670
No log 10.8947 414 0.5779 0.4681 0.5779 0.7602
No log 10.9474 416 0.5832 0.4569 0.5832 0.7637
No log 11.0 418 0.5743 0.4788 0.5743 0.7578
No log 11.0526 420 0.5643 0.5095 0.5643 0.7512
No log 11.1053 422 0.5686 0.4722 0.5686 0.7541
No log 11.1579 424 0.5787 0.4972 0.5787 0.7607
No log 11.2105 426 0.5892 0.4738 0.5892 0.7676
No log 11.2632 428 0.6094 0.4717 0.6094 0.7806
No log 11.3158 430 0.6527 0.5276 0.6527 0.8079
No log 11.3684 432 0.6550 0.4855 0.6550 0.8093
No log 11.4211 434 0.6210 0.4660 0.6210 0.7880
No log 11.4737 436 0.5874 0.4343 0.5874 0.7664
No log 11.5263 438 0.5743 0.4322 0.5743 0.7578
No log 11.5789 440 0.5751 0.4044 0.5751 0.7583
No log 11.6316 442 0.6119 0.3197 0.6119 0.7822
No log 11.6842 444 0.6490 0.3267 0.6490 0.8056
No log 11.7368 446 0.6497 0.3267 0.6497 0.8060
No log 11.7895 448 0.6279 0.3572 0.6279 0.7924
No log 11.8421 450 0.6135 0.4036 0.6135 0.7833
No log 11.8947 452 0.6097 0.4637 0.6097 0.7808
No log 11.9474 454 0.6138 0.5009 0.6138 0.7834
No log 12.0 456 0.6141 0.4737 0.6141 0.7837
No log 12.0526 458 0.6055 0.5009 0.6055 0.7781
No log 12.1053 460 0.6026 0.4820 0.6026 0.7763
No log 12.1579 462 0.5875 0.4575 0.5875 0.7665
No log 12.2105 464 0.5648 0.4569 0.5648 0.7516
No log 12.2632 466 0.5528 0.4322 0.5528 0.7435
No log 12.3158 468 0.5506 0.4044 0.5506 0.7420
No log 12.3684 470 0.5579 0.3228 0.5579 0.7469
No log 12.4211 472 0.5705 0.3523 0.5705 0.7553
No log 12.4737 474 0.6239 0.3728 0.6239 0.7899
No log 12.5263 476 0.6757 0.3224 0.6757 0.8220
No log 12.5789 478 0.7165 0.4212 0.7165 0.8464
No log 12.6316 480 0.6811 0.3966 0.6811 0.8253
No log 12.6842 482 0.6402 0.3524 0.6402 0.8001
No log 12.7368 484 0.6091 0.3445 0.6091 0.7804
No log 12.7895 486 0.5898 0.3552 0.5898 0.7680
No log 12.8421 488 0.5820 0.3863 0.5820 0.7629
No log 12.8947 490 0.5926 0.4161 0.5926 0.7698
No log 12.9474 492 0.6111 0.3728 0.6111 0.7817
No log 13.0 494 0.6315 0.4354 0.6315 0.7947
No log 13.0526 496 0.6816 0.4602 0.6816 0.8256
No log 13.1053 498 0.7210 0.4404 0.7210 0.8491
0.3837 13.1579 500 0.6919 0.4404 0.6919 0.8318
0.3837 13.2105 502 0.6139 0.4782 0.6139 0.7835
0.3837 13.2632 504 0.5931 0.4678 0.5931 0.7701
0.3837 13.3158 506 0.5840 0.5184 0.5840 0.7642
0.3837 13.3684 508 0.5733 0.5021 0.5733 0.7572
0.3837 13.4211 510 0.5664 0.5021 0.5664 0.7526
0.3837 13.4737 512 0.5620 0.5021 0.5620 0.7497
0.3837 13.5263 514 0.5541 0.5021 0.5541 0.7444
0.3837 13.5789 516 0.5593 0.5021 0.5593 0.7479
0.3837 13.6316 518 0.6003 0.4576 0.6003 0.7748
0.3837 13.6842 520 0.6271 0.4534 0.6271 0.7919
0.3837 13.7368 522 0.6206 0.3840 0.6206 0.7878
0.3837 13.7895 524 0.5944 0.3224 0.5944 0.7710
0.3837 13.8421 526 0.5817 0.3575 0.5817 0.7627
0.3837 13.8947 528 0.5815 0.3354 0.5815 0.7626
0.3837 13.9474 530 0.5827 0.3354 0.5827 0.7633

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
7
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k15_task7_organization

Finetuned
(4222)
this model