ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k13_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5551
  • Qwk: 0.4212
  • Mse: 0.5551
  • Rmse: 0.7451

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0303 2 4.4808 -0.0241 4.4808 2.1168
No log 0.0606 4 2.5361 0.0292 2.5361 1.5925
No log 0.0909 6 1.3618 0.0251 1.3618 1.1670
No log 0.1212 8 1.4844 -0.0190 1.4844 1.2184
No log 0.1515 10 1.6436 0.0005 1.6436 1.2820
No log 0.1818 12 1.8597 0.0599 1.8597 1.3637
No log 0.2121 14 1.0017 0.1537 1.0017 1.0008
No log 0.2424 16 0.7737 0.1747 0.7737 0.8796
No log 0.2727 18 0.7482 0.2470 0.7482 0.8650
No log 0.3030 20 0.7895 0.2655 0.7895 0.8886
No log 0.3333 22 0.7725 0.3027 0.7725 0.8789
No log 0.3636 24 0.6700 0.3747 0.6700 0.8185
No log 0.3939 26 0.6799 0.4174 0.6799 0.8246
No log 0.4242 28 0.6332 0.3845 0.6332 0.7957
No log 0.4545 30 0.6474 0.4452 0.6474 0.8046
No log 0.4848 32 0.7149 0.4510 0.7149 0.8455
No log 0.5152 34 0.9503 0.4133 0.9503 0.9748
No log 0.5455 36 1.1799 0.3789 1.1799 1.0862
No log 0.5758 38 1.3263 0.3736 1.3263 1.1517
No log 0.6061 40 1.1906 0.4102 1.1906 1.0911
No log 0.6364 42 1.0628 0.4063 1.0628 1.0309
No log 0.6667 44 0.9567 0.3767 0.9567 0.9781
No log 0.6970 46 0.9060 0.4121 0.9060 0.9519
No log 0.7273 48 0.9660 0.3717 0.9660 0.9829
No log 0.7576 50 0.8919 0.4006 0.8919 0.9444
No log 0.7879 52 0.6718 0.4172 0.6718 0.8196
No log 0.8182 54 0.6396 0.4353 0.6396 0.7998
No log 0.8485 56 0.6812 0.4688 0.6812 0.8254
No log 0.8788 58 0.7271 0.4144 0.7271 0.8527
No log 0.9091 60 1.0504 0.3751 1.0504 1.0249
No log 0.9394 62 0.9591 0.4176 0.9591 0.9793
No log 0.9697 64 0.9006 0.4011 0.9006 0.9490
No log 1.0 66 0.7286 0.3715 0.7286 0.8536
No log 1.0303 68 0.6304 0.4564 0.6304 0.7940
No log 1.0606 70 0.6965 0.5445 0.6965 0.8346
No log 1.0909 72 0.6411 0.5599 0.6411 0.8007
No log 1.1212 74 0.6210 0.5291 0.6210 0.7881
No log 1.1515 76 0.7829 0.4322 0.7829 0.8848
No log 1.1818 78 0.9439 0.4345 0.9439 0.9715
No log 1.2121 80 0.7723 0.4542 0.7723 0.8788
No log 1.2424 82 0.6813 0.4906 0.6813 0.8254
No log 1.2727 84 0.6848 0.4979 0.6848 0.8275
No log 1.3030 86 0.7594 0.4587 0.7594 0.8714
No log 1.3333 88 0.9137 0.4365 0.9137 0.9559
No log 1.3636 90 0.9900 0.4382 0.9900 0.9950
No log 1.3939 92 0.8043 0.4900 0.8043 0.8968
No log 1.4242 94 0.7222 0.4661 0.7222 0.8498
No log 1.4545 96 0.7857 0.5066 0.7857 0.8864
No log 1.4848 98 0.9865 0.4715 0.9865 0.9932
No log 1.5152 100 1.1158 0.4844 1.1158 1.0563
No log 1.5455 102 0.8297 0.4703 0.8297 0.9109
No log 1.5758 104 0.6863 0.5250 0.6863 0.8284
No log 1.6061 106 0.6847 0.5023 0.6847 0.8275
No log 1.6364 108 0.7205 0.4312 0.7205 0.8488
No log 1.6667 110 0.7952 0.4453 0.7952 0.8917
No log 1.6970 112 0.7546 0.4862 0.7546 0.8687
No log 1.7273 114 0.8153 0.4875 0.8153 0.9029
No log 1.7576 116 0.7346 0.4767 0.7346 0.8571
No log 1.7879 118 0.7633 0.5047 0.7633 0.8737
No log 1.8182 120 0.7097 0.4964 0.7097 0.8424
No log 1.8485 122 0.7607 0.5132 0.7607 0.8722
No log 1.8788 124 1.1309 0.3855 1.1309 1.0634
No log 1.9091 126 1.0725 0.4128 1.0725 1.0356
No log 1.9394 128 0.7209 0.4566 0.7209 0.8490
No log 1.9697 130 0.6958 0.4584 0.6958 0.8342
No log 2.0 132 0.8342 0.3164 0.8342 0.9133
No log 2.0303 134 0.7690 0.4088 0.7690 0.8770
No log 2.0606 136 0.6161 0.3859 0.6161 0.7849
No log 2.0909 138 0.7461 0.4110 0.7461 0.8638
No log 2.1212 140 0.7965 0.4481 0.7965 0.8925
No log 2.1515 142 0.6596 0.4374 0.6596 0.8122
No log 2.1818 144 0.6677 0.4601 0.6677 0.8171
No log 2.2121 146 0.6872 0.4878 0.6872 0.8290
No log 2.2424 148 0.6989 0.4789 0.6989 0.8360
No log 2.2727 150 0.7044 0.5152 0.7044 0.8393
No log 2.3030 152 0.7044 0.5177 0.7044 0.8393
No log 2.3333 154 0.7024 0.5190 0.7024 0.8381
No log 2.3636 156 0.6694 0.4866 0.6694 0.8182
No log 2.3939 158 0.6563 0.4483 0.6563 0.8101
No log 2.4242 160 0.6480 0.4429 0.6480 0.8050
No log 2.4545 162 0.6698 0.4511 0.6698 0.8184
No log 2.4848 164 1.0419 0.4183 1.0419 1.0207
No log 2.5152 166 1.3435 0.2958 1.3435 1.1591
No log 2.5455 168 1.1364 0.3778 1.1364 1.0660
No log 2.5758 170 0.7768 0.4366 0.7768 0.8813
No log 2.6061 172 0.6677 0.4968 0.6677 0.8171
No log 2.6364 174 0.6627 0.5041 0.6627 0.8141
No log 2.6667 176 0.6814 0.5120 0.6814 0.8254
No log 2.6970 178 0.8206 0.4677 0.8206 0.9058
No log 2.7273 180 1.1941 0.3877 1.1941 1.0928
No log 2.7576 182 1.2552 0.3325 1.2552 1.1204
No log 2.7879 184 0.9366 0.4667 0.9366 0.9678
No log 2.8182 186 0.7207 0.5060 0.7207 0.8490
No log 2.8485 188 0.6493 0.4484 0.6493 0.8058
No log 2.8788 190 0.6411 0.4381 0.6411 0.8007
No log 2.9091 192 0.6513 0.4219 0.6513 0.8070
No log 2.9394 194 0.6972 0.5010 0.6972 0.8350
No log 2.9697 196 0.7030 0.5060 0.7030 0.8385
No log 3.0 198 0.6519 0.4777 0.6519 0.8074
No log 3.0303 200 0.6801 0.4890 0.6801 0.8247
No log 3.0606 202 0.7258 0.4850 0.7258 0.8520
No log 3.0909 204 0.7016 0.4962 0.7016 0.8376
No log 3.1212 206 0.7521 0.4658 0.7521 0.8672
No log 3.1515 208 0.7809 0.4506 0.7809 0.8837
No log 3.1818 210 0.8575 0.4638 0.8575 0.9260
No log 3.2121 212 0.8918 0.4352 0.8918 0.9443
No log 3.2424 214 0.7952 0.4743 0.7952 0.8917
No log 3.2727 216 0.7254 0.4500 0.7254 0.8517
No log 3.3030 218 0.6812 0.3975 0.6812 0.8254
No log 3.3333 220 0.6532 0.3823 0.6532 0.8082
No log 3.3636 222 0.6358 0.4220 0.6358 0.7973
No log 3.3939 224 0.6385 0.3984 0.6385 0.7991
No log 3.4242 226 0.6443 0.4320 0.6443 0.8027
No log 3.4545 228 0.6678 0.4728 0.6678 0.8172
No log 3.4848 230 0.6437 0.4869 0.6437 0.8023
No log 3.5152 232 0.6448 0.4815 0.6448 0.8030
No log 3.5455 234 0.7799 0.5298 0.7799 0.8831
No log 3.5758 236 0.9535 0.4777 0.9535 0.9765
No log 3.6061 238 0.8486 0.5055 0.8486 0.9212
No log 3.6364 240 0.6148 0.4733 0.6148 0.7841
No log 3.6667 242 0.6091 0.4652 0.6091 0.7805
No log 3.6970 244 0.6281 0.4744 0.6281 0.7925
No log 3.7273 246 0.5800 0.4373 0.5800 0.7615
No log 3.7576 248 0.6383 0.4943 0.6383 0.7989
No log 3.7879 250 0.7672 0.5089 0.7672 0.8759
No log 3.8182 252 0.7014 0.4921 0.7014 0.8375
No log 3.8485 254 0.6049 0.4999 0.6049 0.7777
No log 3.8788 256 0.7274 0.5466 0.7274 0.8529
No log 3.9091 258 0.7667 0.5246 0.7667 0.8756
No log 3.9394 260 0.6503 0.5595 0.6503 0.8064
No log 3.9697 262 0.6855 0.5239 0.6855 0.8279
No log 4.0 264 1.1865 0.3740 1.1865 1.0893
No log 4.0303 266 1.5937 0.2066 1.5937 1.2624
No log 4.0606 268 1.5409 0.2439 1.5409 1.2413
No log 4.0909 270 1.1530 0.3918 1.1530 1.0738
No log 4.1212 272 0.7346 0.5468 0.7346 0.8571
No log 4.1515 274 0.5872 0.4874 0.5872 0.7663
No log 4.1818 276 0.5824 0.5482 0.5824 0.7632
No log 4.2121 278 0.5960 0.5504 0.5960 0.7720
No log 4.2424 280 0.6075 0.5420 0.6075 0.7794
No log 4.2727 282 0.6644 0.4601 0.6644 0.8151
No log 4.3030 284 0.7824 0.4854 0.7824 0.8846
No log 4.3333 286 0.7903 0.4707 0.7903 0.8890
No log 4.3636 288 0.6925 0.4436 0.6925 0.8322
No log 4.3939 290 0.6296 0.4759 0.6296 0.7935
No log 4.4242 292 0.5991 0.4770 0.5991 0.7740
No log 4.4545 294 0.5891 0.4350 0.5891 0.7675
No log 4.4848 296 0.6042 0.4274 0.6042 0.7773
No log 4.5152 298 0.6956 0.4510 0.6956 0.8340
No log 4.5455 300 0.8273 0.4312 0.8273 0.9096
No log 4.5758 302 0.7970 0.4565 0.7970 0.8928
No log 4.6061 304 0.7430 0.4411 0.7430 0.8620
No log 4.6364 306 0.6427 0.4757 0.6427 0.8017
No log 4.6667 308 0.6557 0.4955 0.6557 0.8098
No log 4.6970 310 0.6590 0.4908 0.6590 0.8118
No log 4.7273 312 0.6586 0.4804 0.6586 0.8116
No log 4.7576 314 0.7156 0.4420 0.7156 0.8459
No log 4.7879 316 0.7504 0.4813 0.7504 0.8662
No log 4.8182 318 0.6743 0.4343 0.6743 0.8211
No log 4.8485 320 0.6258 0.4381 0.6258 0.7911
No log 4.8788 322 0.6246 0.4196 0.6246 0.7903
No log 4.9091 324 0.6527 0.4258 0.6527 0.8079
No log 4.9394 326 0.7451 0.4482 0.7451 0.8632
No log 4.9697 328 0.7615 0.4467 0.7615 0.8726
No log 5.0 330 0.6440 0.3989 0.6440 0.8025
No log 5.0303 332 0.5931 0.4569 0.5931 0.7702
No log 5.0606 334 0.5793 0.4465 0.5793 0.7611
No log 5.0909 336 0.5706 0.4540 0.5706 0.7554
No log 5.1212 338 0.5692 0.4185 0.5692 0.7545
No log 5.1515 340 0.5719 0.4642 0.5719 0.7563
No log 5.1818 342 0.6133 0.4855 0.6133 0.7831
No log 5.2121 344 0.6375 0.4559 0.6375 0.7984
No log 5.2424 346 0.5787 0.4512 0.5787 0.7607
No log 5.2727 348 0.5841 0.4536 0.5841 0.7643
No log 5.3030 350 0.6138 0.4875 0.6138 0.7834
No log 5.3333 352 0.6032 0.5041 0.6032 0.7767
No log 5.3636 354 0.5917 0.4863 0.5917 0.7693
No log 5.3939 356 0.5821 0.4920 0.5821 0.7630
No log 5.4242 358 0.5692 0.5133 0.5692 0.7544
No log 5.4545 360 0.5492 0.4166 0.5492 0.7411
No log 5.4848 362 0.5609 0.4020 0.5609 0.7489
No log 5.5152 364 0.5727 0.4568 0.5727 0.7567
No log 5.5455 366 0.6458 0.5145 0.6458 0.8036
No log 5.5758 368 0.7154 0.5334 0.7154 0.8458
No log 5.6061 370 0.6886 0.5423 0.6886 0.8298
No log 5.6364 372 0.6272 0.5433 0.6272 0.7919
No log 5.6667 374 0.5995 0.4928 0.5995 0.7743
No log 5.6970 376 0.6158 0.4887 0.6158 0.7847
No log 5.7273 378 0.7061 0.5322 0.7061 0.8403
No log 5.7576 380 0.6838 0.5182 0.6838 0.8269
No log 5.7879 382 0.5873 0.4948 0.5873 0.7664
No log 5.8182 384 0.5674 0.5096 0.5674 0.7533
No log 5.8485 386 0.5745 0.5096 0.5745 0.7579
No log 5.8788 388 0.5728 0.5057 0.5728 0.7568
No log 5.9091 390 0.6111 0.5355 0.6111 0.7817
No log 5.9394 392 0.7185 0.4618 0.7185 0.8477
No log 5.9697 394 0.7271 0.4745 0.7271 0.8527
No log 6.0 396 0.6373 0.4898 0.6373 0.7983
No log 6.0303 398 0.5958 0.4938 0.5958 0.7719
No log 6.0606 400 0.5834 0.4559 0.5834 0.7638
No log 6.0909 402 0.5930 0.3894 0.5930 0.7700
No log 6.1212 404 0.6477 0.4769 0.6477 0.8048
No log 6.1515 406 0.6430 0.4646 0.6430 0.8019
No log 6.1818 408 0.5802 0.4307 0.5802 0.7617
No log 6.2121 410 0.5687 0.4719 0.5687 0.7541
No log 6.2424 412 0.5708 0.4666 0.5708 0.7555
No log 6.2727 414 0.5852 0.4526 0.5852 0.7650
No log 6.3030 416 0.6753 0.4927 0.6753 0.8218
No log 6.3333 418 0.6999 0.4941 0.6999 0.8366
No log 6.3636 420 0.6612 0.4771 0.6612 0.8131
No log 6.3939 422 0.5875 0.4751 0.5875 0.7665
No log 6.4242 424 0.5558 0.4321 0.5558 0.7455
No log 6.4545 426 0.5510 0.4398 0.5510 0.7423
No log 6.4848 428 0.5511 0.3973 0.5511 0.7424
No log 6.5152 430 0.5865 0.4553 0.5865 0.7658
No log 6.5455 432 0.6510 0.5078 0.6510 0.8069
No log 6.5758 434 0.6614 0.4764 0.6614 0.8133
No log 6.6061 436 0.5964 0.4336 0.5964 0.7722
No log 6.6364 438 0.5731 0.4265 0.5731 0.7570
No log 6.6667 440 0.6180 0.5053 0.6180 0.7861
No log 6.6970 442 0.6422 0.5589 0.6422 0.8014
No log 6.7273 444 0.6178 0.4827 0.6178 0.7860
No log 6.7576 446 0.6027 0.5048 0.6027 0.7763
No log 6.7879 448 0.6377 0.5156 0.6377 0.7986
No log 6.8182 450 0.6358 0.5339 0.6358 0.7974
No log 6.8485 452 0.6185 0.5195 0.6185 0.7865
No log 6.8788 454 0.6161 0.5155 0.6161 0.7849
No log 6.9091 456 0.6091 0.4766 0.6091 0.7804
No log 6.9394 458 0.6441 0.4935 0.6441 0.8026
No log 6.9697 460 0.6545 0.5023 0.6545 0.8090
No log 7.0 462 0.5925 0.4463 0.5925 0.7698
No log 7.0303 464 0.5732 0.3587 0.5732 0.7571
No log 7.0606 466 0.6089 0.4925 0.6089 0.7803
No log 7.0909 468 0.6028 0.4876 0.6028 0.7764
No log 7.1212 470 0.5900 0.4278 0.5900 0.7681
No log 7.1515 472 0.5834 0.4460 0.5834 0.7638
No log 7.1818 474 0.5763 0.3968 0.5763 0.7591
No log 7.2121 476 0.5728 0.4291 0.5728 0.7568
No log 7.2424 478 0.5845 0.5137 0.5845 0.7645
No log 7.2727 480 0.5687 0.5270 0.5687 0.7542
No log 7.3030 482 0.5721 0.5105 0.5721 0.7563
No log 7.3333 484 0.5821 0.5016 0.5821 0.7629
No log 7.3636 486 0.5978 0.5205 0.5978 0.7732
No log 7.3939 488 0.6513 0.4976 0.6513 0.8070
No log 7.4242 490 0.6193 0.4842 0.6193 0.7869
No log 7.4545 492 0.5926 0.4975 0.5926 0.7698
No log 7.4848 494 0.5685 0.5184 0.5685 0.7540
No log 7.5152 496 0.5642 0.4591 0.5642 0.7511
No log 7.5455 498 0.5645 0.4809 0.5645 0.7514
0.365 7.5758 500 0.5696 0.4809 0.5696 0.7547
0.365 7.6061 502 0.5845 0.5265 0.5845 0.7645
0.365 7.6364 504 0.5997 0.4867 0.5997 0.7744
0.365 7.6667 506 0.5995 0.5138 0.5995 0.7743
0.365 7.6970 508 0.5957 0.5004 0.5957 0.7718
0.365 7.7273 510 0.5819 0.4653 0.5819 0.7628
0.365 7.7576 512 0.6377 0.4761 0.6377 0.7986
0.365 7.7879 514 0.6919 0.4495 0.6919 0.8318
0.365 7.8182 516 0.6691 0.4929 0.6691 0.8180
0.365 7.8485 518 0.6075 0.4847 0.6075 0.7794
0.365 7.8788 520 0.5949 0.4640 0.5949 0.7713
0.365 7.9091 522 0.6123 0.4954 0.6123 0.7825
0.365 7.9394 524 0.6110 0.5043 0.6110 0.7817
0.365 7.9697 526 0.5803 0.4579 0.5803 0.7617
0.365 8.0 528 0.5692 0.4192 0.5692 0.7544
0.365 8.0303 530 0.5701 0.4132 0.5701 0.7551
0.365 8.0606 532 0.5876 0.4628 0.5876 0.7665
0.365 8.0909 534 0.5948 0.5031 0.5948 0.7712
0.365 8.1212 536 0.5994 0.4910 0.5994 0.7742
0.365 8.1515 538 0.5788 0.4608 0.5788 0.7608
0.365 8.1818 540 0.5690 0.4559 0.5690 0.7543
0.365 8.2121 542 0.5620 0.4356 0.5620 0.7497
0.365 8.2424 544 0.5573 0.4368 0.5573 0.7465
0.365 8.2727 546 0.5551 0.4212 0.5551 0.7451

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
6
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k13_task2_organization

Finetuned
(4222)
this model