ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k19_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8953
  • Qwk: 0.6310
  • Mse: 0.8953
  • Rmse: 0.9462

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0217 2 5.4499 0.0060 5.4499 2.3345
No log 0.0435 4 3.3724 0.0629 3.3724 1.8364
No log 0.0652 6 2.3352 -0.0712 2.3352 1.5281
No log 0.0870 8 2.6157 -0.0878 2.6157 1.6173
No log 0.1087 10 1.7444 0.0139 1.7444 1.3208
No log 0.1304 12 1.5661 -0.0045 1.5661 1.2515
No log 0.1522 14 1.3082 0.1369 1.3082 1.1438
No log 0.1739 16 1.1941 0.1941 1.1941 1.0927
No log 0.1957 18 1.3025 0.1161 1.3025 1.1413
No log 0.2174 20 1.4922 0.0302 1.4922 1.2216
No log 0.2391 22 1.5103 0.0440 1.5103 1.2289
No log 0.2609 24 1.4271 0.1372 1.4271 1.1946
No log 0.2826 26 1.5068 0.1406 1.5068 1.2275
No log 0.3043 28 1.4679 0.1336 1.4679 1.2116
No log 0.3261 30 1.2553 0.2568 1.2553 1.1204
No log 0.3478 32 1.1750 0.3199 1.1750 1.0840
No log 0.3696 34 1.2857 0.2838 1.2857 1.1339
No log 0.3913 36 1.2365 0.2866 1.2365 1.1120
No log 0.4130 38 1.1940 0.3174 1.1940 1.0927
No log 0.4348 40 1.2162 0.3315 1.2162 1.1028
No log 0.4565 42 1.3430 0.3476 1.3430 1.1589
No log 0.4783 44 1.0822 0.3536 1.0822 1.0403
No log 0.5 46 1.0232 0.3984 1.0232 1.0115
No log 0.5217 48 1.0265 0.3645 1.0265 1.0132
No log 0.5435 50 1.0117 0.3816 1.0117 1.0058
No log 0.5652 52 1.0444 0.4140 1.0444 1.0219
No log 0.5870 54 1.1067 0.3885 1.1067 1.0520
No log 0.6087 56 1.0896 0.4041 1.0896 1.0438
No log 0.6304 58 1.0558 0.3905 1.0558 1.0275
No log 0.6522 60 1.0863 0.3947 1.0863 1.0423
No log 0.6739 62 1.2683 0.3589 1.2683 1.1262
No log 0.6957 64 1.2027 0.3903 1.2027 1.0967
No log 0.7174 66 0.9521 0.4298 0.9521 0.9758
No log 0.7391 68 0.8959 0.4030 0.8959 0.9465
No log 0.7609 70 0.8674 0.4861 0.8674 0.9313
No log 0.7826 72 1.1526 0.4808 1.1526 1.0736
No log 0.8043 74 1.3034 0.375 1.3034 1.1417
No log 0.8261 76 0.9520 0.5238 0.9520 0.9757
No log 0.8478 78 0.8189 0.5340 0.8189 0.9049
No log 0.8696 80 0.8792 0.5560 0.8792 0.9377
No log 0.8913 82 0.7730 0.5707 0.7730 0.8792
No log 0.9130 84 0.8352 0.5901 0.8352 0.9139
No log 0.9348 86 0.8918 0.5809 0.8918 0.9443
No log 0.9565 88 0.7612 0.5962 0.7612 0.8725
No log 0.9783 90 0.7655 0.4979 0.7655 0.8750
No log 1.0 92 0.8264 0.5273 0.8264 0.9091
No log 1.0217 94 0.9179 0.5021 0.9179 0.9581
No log 1.0435 96 0.8525 0.5393 0.8525 0.9233
No log 1.0652 98 0.8242 0.5764 0.8242 0.9079
No log 1.0870 100 0.8623 0.6398 0.8623 0.9286
No log 1.1087 102 0.8603 0.6415 0.8603 0.9275
No log 1.1304 104 0.8803 0.6391 0.8803 0.9383
No log 1.1522 106 0.9499 0.6323 0.9499 0.9746
No log 1.1739 108 1.0386 0.5921 1.0386 1.0191
No log 1.1957 110 0.9862 0.5561 0.9862 0.9931
No log 1.2174 112 0.8603 0.6126 0.8603 0.9275
No log 1.2391 114 0.8023 0.6345 0.8023 0.8957
No log 1.2609 116 0.7986 0.6422 0.7986 0.8937
No log 1.2826 118 0.8948 0.5943 0.8948 0.9459
No log 1.3043 120 0.9439 0.5638 0.9439 0.9715
No log 1.3261 122 0.7703 0.6358 0.7703 0.8777
No log 1.3478 124 0.7782 0.6222 0.7782 0.8822
No log 1.3696 126 0.7700 0.6291 0.7700 0.8775
No log 1.3913 128 0.7154 0.5734 0.7154 0.8458
No log 1.4130 130 0.9798 0.5354 0.9798 0.9898
No log 1.4348 132 1.0229 0.5437 1.0229 1.0114
No log 1.4565 134 0.9301 0.5990 0.9301 0.9644
No log 1.4783 136 0.9355 0.6111 0.9355 0.9672
No log 1.5 138 1.0716 0.5424 1.0716 1.0352
No log 1.5217 140 1.0307 0.5570 1.0307 1.0153
No log 1.5435 142 1.0169 0.5716 1.0169 1.0084
No log 1.5652 144 0.9038 0.6087 0.9038 0.9507
No log 1.5870 146 0.8592 0.6094 0.8592 0.9269
No log 1.6087 148 0.9373 0.5714 0.9373 0.9682
No log 1.6304 150 0.8392 0.5997 0.8392 0.9161
No log 1.6522 152 0.7727 0.5931 0.7727 0.8791
No log 1.6739 154 0.7300 0.6285 0.7300 0.8544
No log 1.6957 156 0.7345 0.6139 0.7345 0.8570
No log 1.7174 158 0.7574 0.6250 0.7574 0.8703
No log 1.7391 160 0.7274 0.6498 0.7274 0.8529
No log 1.7609 162 0.7268 0.6357 0.7268 0.8525
No log 1.7826 164 0.7611 0.5890 0.7611 0.8724
No log 1.8043 166 0.8420 0.5862 0.8420 0.9176
No log 1.8261 168 0.8124 0.6086 0.8124 0.9013
No log 1.8478 170 0.8741 0.6053 0.8741 0.9349
No log 1.8696 172 0.7075 0.6537 0.7075 0.8411
No log 1.8913 174 0.6805 0.6501 0.6805 0.8249
No log 1.9130 176 0.6778 0.6687 0.6778 0.8233
No log 1.9348 178 0.6681 0.6671 0.6681 0.8173
No log 1.9565 180 0.8725 0.6178 0.8725 0.9341
No log 1.9783 182 0.8908 0.5820 0.8908 0.9438
No log 2.0 184 0.7654 0.6784 0.7654 0.8749
No log 2.0217 186 0.7054 0.6896 0.7054 0.8399
No log 2.0435 188 0.7355 0.7011 0.7355 0.8576
No log 2.0652 190 0.7456 0.6873 0.7456 0.8635
No log 2.0870 192 0.8244 0.6427 0.8244 0.9080
No log 2.1087 194 0.8502 0.6315 0.8502 0.9221
No log 2.1304 196 0.9401 0.5909 0.9401 0.9696
No log 2.1522 198 0.7754 0.6377 0.7754 0.8805
No log 2.1739 200 0.6460 0.6598 0.6460 0.8037
No log 2.1957 202 0.6850 0.6604 0.6850 0.8277
No log 2.2174 204 0.6491 0.6563 0.6491 0.8057
No log 2.2391 206 0.7570 0.6604 0.7570 0.8700
No log 2.2609 208 1.0338 0.5485 1.0338 1.0168
No log 2.2826 210 1.0287 0.5485 1.0287 1.0143
No log 2.3043 212 0.8185 0.6371 0.8185 0.9047
No log 2.3261 214 0.7027 0.6634 0.7027 0.8383
No log 2.3478 216 0.7277 0.6704 0.7277 0.8530
No log 2.3696 218 0.8580 0.6456 0.8580 0.9263
No log 2.3913 220 0.9889 0.5878 0.9889 0.9944
No log 2.4130 222 1.1221 0.5143 1.1221 1.0593
No log 2.4348 224 0.9647 0.5731 0.9647 0.9822
No log 2.4565 226 0.7595 0.6366 0.7595 0.8715
No log 2.4783 228 0.6814 0.6423 0.6814 0.8255
No log 2.5 230 0.7302 0.5653 0.7302 0.8545
No log 2.5217 232 0.8086 0.6062 0.8086 0.8992
No log 2.5435 234 0.8034 0.6042 0.8034 0.8964
No log 2.5652 236 0.7471 0.6311 0.7471 0.8643
No log 2.5870 238 0.7299 0.6418 0.7299 0.8543
No log 2.6087 240 0.7589 0.6131 0.7589 0.8712
No log 2.6304 242 0.8637 0.6073 0.8637 0.9293
No log 2.6522 244 0.8340 0.6012 0.8340 0.9132
No log 2.6739 246 0.7641 0.6230 0.7641 0.8741
No log 2.6957 248 0.7144 0.6162 0.7144 0.8452
No log 2.7174 250 0.7061 0.6581 0.7061 0.8403
No log 2.7391 252 0.7141 0.6537 0.7141 0.8451
No log 2.7609 254 0.6944 0.6659 0.6944 0.8333
No log 2.7826 256 0.8054 0.6164 0.8054 0.8974
No log 2.8043 258 1.0727 0.5315 1.0727 1.0357
No log 2.8261 260 1.0744 0.5212 1.0744 1.0365
No log 2.8478 262 0.8808 0.5924 0.8808 0.9385
No log 2.8696 264 0.6663 0.6632 0.6663 0.8163
No log 2.8913 266 0.6414 0.6951 0.6414 0.8009
No log 2.9130 268 0.6374 0.7282 0.6374 0.7984
No log 2.9348 270 0.6644 0.6730 0.6644 0.8151
No log 2.9565 272 0.7451 0.6655 0.7451 0.8632
No log 2.9783 274 0.8111 0.6305 0.8111 0.9006
No log 3.0 276 0.7530 0.6525 0.7530 0.8677
No log 3.0217 278 0.6330 0.7001 0.6330 0.7956
No log 3.0435 280 0.6051 0.6747 0.6051 0.7779
No log 3.0652 282 0.6103 0.6858 0.6103 0.7812
No log 3.0870 284 0.7022 0.7015 0.7022 0.8380
No log 3.1087 286 0.7630 0.6579 0.7630 0.8735
No log 3.1304 288 0.7207 0.6791 0.7207 0.8489
No log 3.1522 290 0.6287 0.6772 0.6287 0.7929
No log 3.1739 292 0.6596 0.6804 0.6596 0.8121
No log 3.1957 294 0.8196 0.6245 0.8196 0.9053
No log 3.2174 296 1.0051 0.5683 1.0051 1.0025
No log 3.2391 298 0.9261 0.5580 0.9261 0.9623
No log 3.2609 300 0.7892 0.6249 0.7892 0.8884
No log 3.2826 302 0.7905 0.6116 0.7905 0.8891
No log 3.3043 304 0.9098 0.5767 0.9097 0.9538
No log 3.3261 306 0.9529 0.5553 0.9529 0.9762
No log 3.3478 308 0.8601 0.5833 0.8601 0.9274
No log 3.3696 310 0.7473 0.6111 0.7473 0.8645
No log 3.3913 312 0.7275 0.6619 0.7275 0.8529
No log 3.4130 314 0.7784 0.6144 0.7784 0.8823
No log 3.4348 316 0.7679 0.6369 0.7679 0.8763
No log 3.4565 318 0.7437 0.6597 0.7437 0.8624
No log 3.4783 320 0.6976 0.6857 0.6976 0.8352
No log 3.5 322 0.6649 0.6995 0.6649 0.8154
No log 3.5217 324 0.6994 0.6757 0.6994 0.8363
No log 3.5435 326 0.7796 0.6087 0.7796 0.8829
No log 3.5652 328 0.8828 0.5910 0.8828 0.9396
No log 3.5870 330 0.8984 0.6009 0.8984 0.9479
No log 3.6087 332 0.7608 0.6378 0.7608 0.8722
No log 3.6304 334 0.7135 0.6788 0.7135 0.8447
No log 3.6522 336 0.7968 0.6482 0.7968 0.8927
No log 3.6739 338 0.7680 0.6436 0.7680 0.8764
No log 3.6957 340 0.7202 0.6498 0.7202 0.8487
No log 3.7174 342 0.7037 0.6889 0.7037 0.8389
No log 3.7391 344 0.7201 0.6800 0.7201 0.8486
No log 3.7609 346 0.7710 0.6471 0.7710 0.8781
No log 3.7826 348 0.7863 0.6317 0.7863 0.8867
No log 3.8043 350 0.7231 0.7018 0.7231 0.8504
No log 3.8261 352 0.6887 0.7068 0.6887 0.8299
No log 3.8478 354 0.6534 0.6994 0.6534 0.8084
No log 3.8696 356 0.6647 0.6924 0.6647 0.8153
No log 3.8913 358 0.6703 0.6670 0.6703 0.8187
No log 3.9130 360 0.6566 0.6508 0.6566 0.8103
No log 3.9348 362 0.6469 0.6656 0.6469 0.8043
No log 3.9565 364 0.6454 0.6556 0.6454 0.8034
No log 3.9783 366 0.6360 0.6740 0.6360 0.7975
No log 4.0 368 0.6617 0.7018 0.6617 0.8134
No log 4.0217 370 0.7841 0.6421 0.7841 0.8855
No log 4.0435 372 0.8129 0.6321 0.8129 0.9016
No log 4.0652 374 0.7676 0.6365 0.7676 0.8761
No log 4.0870 376 0.7089 0.6785 0.7089 0.8419
No log 4.1087 378 0.6878 0.6695 0.6878 0.8294
No log 4.1304 380 0.7413 0.6665 0.7413 0.8610
No log 4.1522 382 0.7783 0.6447 0.7783 0.8822
No log 4.1739 384 0.7353 0.6676 0.7353 0.8575
No log 4.1957 386 0.6939 0.6951 0.6939 0.8330
No log 4.2174 388 0.6890 0.7020 0.6890 0.8300
No log 4.2391 390 0.7320 0.6794 0.7320 0.8555
No log 4.2609 392 0.8067 0.6456 0.8067 0.8982
No log 4.2826 394 0.7612 0.6787 0.7612 0.8725
No log 4.3043 396 0.6956 0.6746 0.6956 0.8340
No log 4.3261 398 0.6728 0.6640 0.6728 0.8202
No log 4.3478 400 0.6723 0.6490 0.6723 0.8200
No log 4.3696 402 0.6712 0.6391 0.6712 0.8193
No log 4.3913 404 0.6921 0.6271 0.6921 0.8319
No log 4.4130 406 0.7292 0.6273 0.7292 0.8539
No log 4.4348 408 0.7536 0.6381 0.7536 0.8681
No log 4.4565 410 0.8094 0.6162 0.8094 0.8997
No log 4.4783 412 0.8771 0.6032 0.8771 0.9365
No log 4.5 414 0.8909 0.6035 0.8909 0.9439
No log 4.5217 416 0.8838 0.5985 0.8838 0.9401
No log 4.5435 418 0.7607 0.625 0.7607 0.8722
No log 4.5652 420 0.6551 0.7020 0.6551 0.8094
No log 4.5870 422 0.6563 0.6790 0.6563 0.8101
No log 4.6087 424 0.7263 0.6570 0.7263 0.8523
No log 4.6304 426 0.9044 0.5808 0.9044 0.9510
No log 4.6522 428 1.0306 0.5669 1.0306 1.0152
No log 4.6739 430 0.9503 0.5782 0.9503 0.9748
No log 4.6957 432 0.7442 0.6523 0.7442 0.8626
No log 4.7174 434 0.6784 0.6892 0.6784 0.8237
No log 4.7391 436 0.6806 0.7034 0.6806 0.8250
No log 4.7609 438 0.7770 0.6667 0.7770 0.8815
No log 4.7826 440 0.8206 0.6456 0.8206 0.9059
No log 4.8043 442 0.7492 0.6753 0.7492 0.8655
No log 4.8261 444 0.6413 0.6987 0.6413 0.8008
No log 4.8478 446 0.6303 0.6892 0.6303 0.7939
No log 4.8696 448 0.6688 0.7088 0.6688 0.8178
No log 4.8913 450 0.7506 0.6635 0.7506 0.8664
No log 4.9130 452 0.8525 0.6551 0.8525 0.9233
No log 4.9348 454 0.8235 0.6726 0.8235 0.9075
No log 4.9565 456 0.7254 0.6942 0.7254 0.8517
No log 4.9783 458 0.6975 0.6757 0.6975 0.8351
No log 5.0 460 0.7264 0.6961 0.7264 0.8523
No log 5.0217 462 0.7479 0.6546 0.7479 0.8648
No log 5.0435 464 0.8024 0.6205 0.8024 0.8958
No log 5.0652 466 0.8208 0.6084 0.8208 0.9060
No log 5.0870 468 0.8298 0.6023 0.8298 0.9109
No log 5.1087 470 0.7818 0.6342 0.7818 0.8842
No log 5.1304 472 0.7034 0.6534 0.7034 0.8387
No log 5.1522 474 0.6958 0.6593 0.6958 0.8341
No log 5.1739 476 0.7565 0.6666 0.7565 0.8698
No log 5.1957 478 0.9472 0.5742 0.9472 0.9733
No log 5.2174 480 1.0367 0.5636 1.0367 1.0182
No log 5.2391 482 0.9019 0.6314 0.9019 0.9497
No log 5.2609 484 0.7271 0.6655 0.7271 0.8527
No log 5.2826 486 0.7069 0.6693 0.7069 0.8408
No log 5.3043 488 0.6944 0.6909 0.6944 0.8333
No log 5.3261 490 0.7369 0.6826 0.7369 0.8584
No log 5.3478 492 0.7246 0.6801 0.7246 0.8513
No log 5.3696 494 0.6986 0.6801 0.6986 0.8358
No log 5.3913 496 0.7243 0.6645 0.7243 0.8511
No log 5.4130 498 0.8238 0.6118 0.8238 0.9076
0.4481 5.4348 500 0.7643 0.6441 0.7643 0.8742
0.4481 5.4565 502 0.7213 0.6730 0.7213 0.8493
0.4481 5.4783 504 0.7640 0.6354 0.7640 0.8741
0.4481 5.5 506 0.8349 0.6435 0.8349 0.9137
0.4481 5.5217 508 0.7457 0.6590 0.7457 0.8636
0.4481 5.5435 510 0.7424 0.6530 0.7424 0.8616
0.4481 5.5652 512 0.7962 0.6123 0.7962 0.8923
0.4481 5.5870 514 0.8656 0.6015 0.8656 0.9304
0.4481 5.6087 516 0.8543 0.6058 0.8543 0.9243
0.4481 5.6304 518 0.9103 0.6331 0.9103 0.9541
0.4481 5.6522 520 0.8953 0.6310 0.8953 0.9462

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
5
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k19_task1_organization

Finetuned
(4222)
this model