ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k11_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7249
  • Qwk: 0.7310
  • Mse: 0.7249
  • Rmse: 0.8514

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0392 2 6.8798 0.0242 6.8798 2.6229
No log 0.0784 4 4.6619 0.1040 4.6619 2.1591
No log 0.1176 6 3.5305 0.0103 3.5305 1.8790
No log 0.1569 8 2.4483 0.1781 2.4483 1.5647
No log 0.1961 10 1.9638 0.1538 1.9638 1.4013
No log 0.2353 12 1.9751 0.1587 1.9751 1.4054
No log 0.2745 14 2.2191 0.2222 2.2191 1.4897
No log 0.3137 16 2.1624 0.2416 2.1624 1.4705
No log 0.3529 18 1.6220 0.4627 1.6220 1.2736
No log 0.3922 20 1.4015 0.5113 1.4015 1.1839
No log 0.4314 22 1.6461 0.4895 1.6461 1.2830
No log 0.4706 24 1.6610 0.4533 1.6610 1.2888
No log 0.5098 26 1.3829 0.4662 1.3829 1.1759
No log 0.5490 28 1.2499 0.4733 1.2499 1.1180
No log 0.5882 30 1.1612 0.4651 1.1612 1.0776
No log 0.6275 32 1.0874 0.5455 1.0874 1.0428
No log 0.6667 34 1.1640 0.5547 1.1640 1.0789
No log 0.7059 36 1.3800 0.5278 1.3800 1.1748
No log 0.7451 38 2.4638 0.3037 2.4638 1.5697
No log 0.7843 40 2.9469 0.2816 2.9469 1.7167
No log 0.8235 42 2.4912 0.2872 2.4912 1.5784
No log 0.8627 44 1.5294 0.4575 1.5294 1.2367
No log 0.9020 46 1.2339 0.5714 1.2339 1.1108
No log 0.9412 48 1.3406 0.5070 1.3406 1.1578
No log 0.9804 50 1.5152 0.4626 1.5152 1.2309
No log 1.0196 52 1.4897 0.4832 1.4897 1.2205
No log 1.0588 54 1.3197 0.5526 1.3197 1.1488
No log 1.0980 56 1.2439 0.5674 1.2439 1.1153
No log 1.1373 58 1.1463 0.6345 1.1463 1.0706
No log 1.1765 60 1.1493 0.6164 1.1493 1.0720
No log 1.2157 62 1.2549 0.5235 1.2549 1.1202
No log 1.2549 64 1.3221 0.4552 1.3221 1.1498
No log 1.2941 66 1.1360 0.5797 1.1360 1.0658
No log 1.3333 68 1.0332 0.6043 1.0332 1.0164
No log 1.3725 70 0.9719 0.5865 0.9719 0.9858
No log 1.4118 72 0.9902 0.6131 0.9902 0.9951
No log 1.4510 74 1.2440 0.5676 1.2440 1.1154
No log 1.4902 76 1.4434 0.5526 1.4434 1.2014
No log 1.5294 78 1.3165 0.5860 1.3165 1.1474
No log 1.5686 80 0.9547 0.6795 0.9547 0.9771
No log 1.6078 82 0.9131 0.6879 0.9131 0.9556
No log 1.6471 84 0.9033 0.6871 0.9033 0.9504
No log 1.6863 86 0.7536 0.75 0.7536 0.8681
No log 1.7255 88 0.6854 0.8 0.6854 0.8279
No log 1.7647 90 0.6759 0.8095 0.6759 0.8222
No log 1.8039 92 0.7668 0.7625 0.7668 0.8756
No log 1.8431 94 1.4470 0.4366 1.4470 1.2029
No log 1.8824 96 1.7286 0.3310 1.7286 1.3148
No log 1.9216 98 1.2891 0.5854 1.2891 1.1354
No log 1.9608 100 0.9555 0.6927 0.9555 0.9775
No log 2.0 102 0.8572 0.7708 0.8572 0.9258
No log 2.0392 104 0.8652 0.7609 0.8652 0.9302
No log 2.0784 106 1.0501 0.6941 1.0501 1.0248
No log 2.1176 108 1.1758 0.6225 1.1758 1.0843
No log 2.1569 110 0.7972 0.7333 0.7972 0.8929
No log 2.1961 112 0.6635 0.7320 0.6635 0.8145
No log 2.2353 114 1.2685 0.6380 1.2685 1.1263
No log 2.2745 116 1.3155 0.6463 1.3155 1.1470
No log 2.3137 118 0.8970 0.6968 0.8970 0.9471
No log 2.3529 120 0.7347 0.7134 0.7347 0.8571
No log 2.3922 122 0.5620 0.7848 0.5620 0.7497
No log 2.4314 124 0.6168 0.7389 0.6168 0.7853
No log 2.4706 126 0.6517 0.7273 0.6517 0.8073
No log 2.5098 128 0.6334 0.7407 0.6334 0.7959
No log 2.5490 130 0.7873 0.7910 0.7873 0.8873
No log 2.5882 132 0.8367 0.7841 0.8367 0.9147
No log 2.6275 134 0.7866 0.7329 0.7866 0.8869
No log 2.6667 136 0.8695 0.6710 0.8695 0.9325
No log 2.7059 138 0.7745 0.7170 0.7745 0.8801
No log 2.7451 140 0.7407 0.7711 0.7407 0.8606
No log 2.7843 142 0.8139 0.7735 0.8139 0.9021
No log 2.8235 144 0.8690 0.7363 0.8690 0.9322
No log 2.8627 146 0.7154 0.8 0.7154 0.8458
No log 2.9020 148 0.6507 0.7320 0.6507 0.8066
No log 2.9412 150 0.6547 0.7821 0.6547 0.8092
No log 2.9804 152 0.8581 0.6788 0.8581 0.9263
No log 3.0196 154 0.9703 0.6509 0.9703 0.9850
No log 3.0588 156 0.8248 0.6879 0.8248 0.9082
No log 3.0980 158 0.7753 0.7059 0.7753 0.8805
No log 3.1373 160 0.6876 0.7273 0.6876 0.8292
No log 3.1765 162 0.6648 0.7273 0.6648 0.8154
No log 3.2157 164 0.7205 0.75 0.7205 0.8488
No log 3.2549 166 1.0402 0.6582 1.0402 1.0199
No log 3.2941 168 1.0334 0.6667 1.0334 1.0166
No log 3.3333 170 0.7756 0.7215 0.7756 0.8807
No log 3.3725 172 0.6212 0.7898 0.6212 0.7882
No log 3.4118 174 0.6329 0.7619 0.6329 0.7955
No log 3.4510 176 0.6685 0.7974 0.6685 0.8176
No log 3.4902 178 0.7626 0.7468 0.7626 0.8733
No log 3.5294 180 1.0242 0.6234 1.0242 1.0120
No log 3.5686 182 1.1713 0.6452 1.1713 1.0823
No log 3.6078 184 1.0337 0.6358 1.0337 1.0167
No log 3.6471 186 0.8535 0.6338 0.8535 0.9239
No log 3.6863 188 0.7460 0.7397 0.7460 0.8637
No log 3.7255 190 0.7279 0.7848 0.7279 0.8532
No log 3.7647 192 0.7426 0.7904 0.7426 0.8617
No log 3.8039 194 0.8508 0.7582 0.8508 0.9224
No log 3.8431 196 0.8814 0.7527 0.8814 0.9388
No log 3.8824 198 0.7976 0.7541 0.7976 0.8931
No log 3.9216 200 0.6997 0.8 0.6997 0.8365
No log 3.9608 202 0.7065 0.7582 0.7065 0.8405
No log 4.0 204 0.7861 0.7389 0.7861 0.8866
No log 4.0392 206 1.0435 0.6582 1.0435 1.0215
No log 4.0784 208 1.2892 0.6265 1.2892 1.1354
No log 4.1176 210 1.3528 0.6265 1.3528 1.1631
No log 4.1569 212 1.0732 0.6627 1.0732 1.0360
No log 4.1961 214 0.9543 0.6905 0.9543 0.9769
No log 4.2353 216 0.7979 0.7394 0.7979 0.8933
No log 4.2745 218 0.7473 0.7701 0.7473 0.8645
No log 4.3137 220 0.7354 0.7735 0.7354 0.8575
No log 4.3529 222 0.6334 0.8208 0.6334 0.7959
No log 4.3922 224 0.6840 0.7436 0.6840 0.8271
No log 4.4314 226 0.7895 0.6797 0.7895 0.8886
No log 4.4706 228 0.8011 0.6842 0.8011 0.8950
No log 4.5098 230 0.7813 0.6667 0.7813 0.8839
No log 4.5490 232 0.6891 0.7843 0.6891 0.8301
No log 4.5882 234 0.6624 0.7898 0.6624 0.8139
No log 4.6275 236 0.6879 0.7532 0.6879 0.8294
No log 4.6667 238 0.6946 0.7568 0.6946 0.8334
No log 4.7059 240 0.6868 0.7568 0.6868 0.8288
No log 4.7451 242 0.6580 0.7586 0.6580 0.8112
No log 4.7843 244 0.6430 0.7568 0.6430 0.8019
No log 4.8235 246 0.6332 0.7568 0.6332 0.7958
No log 4.8627 248 0.6621 0.7712 0.6621 0.8137
No log 4.9020 250 0.8173 0.6835 0.8173 0.9040
No log 4.9412 252 1.1167 0.6552 1.1167 1.0567
No log 4.9804 254 1.1406 0.6667 1.1406 1.0680
No log 5.0196 256 0.9486 0.6125 0.9486 0.9740
No log 5.0588 258 0.7107 0.7568 0.7107 0.8430
No log 5.0980 260 0.6962 0.7361 0.6962 0.8344
No log 5.1373 262 0.6913 0.7347 0.6913 0.8315
No log 5.1765 264 0.7394 0.7285 0.7394 0.8599
No log 5.2157 266 0.7604 0.7362 0.7604 0.8720
No log 5.2549 268 0.6976 0.7285 0.6976 0.8352
No log 5.2941 270 0.6409 0.7483 0.6409 0.8005
No log 5.3333 272 0.6486 0.7448 0.6486 0.8053
No log 5.3725 274 0.6629 0.7568 0.6629 0.8142
No log 5.4118 276 0.7152 0.7248 0.7152 0.8457
No log 5.4510 278 1.0317 0.6228 1.0317 1.0157
No log 5.4902 280 1.2208 0.6243 1.2208 1.1049
No log 5.5294 282 1.1119 0.6316 1.1119 1.0545
No log 5.5686 284 0.8484 0.7209 0.8484 0.9211
No log 5.6078 286 0.7283 0.7261 0.7283 0.8534
No log 5.6471 288 0.7468 0.7284 0.7468 0.8642
No log 5.6863 290 0.8732 0.6782 0.8732 0.9344
No log 5.7255 292 0.8754 0.6782 0.8754 0.9356
No log 5.7647 294 0.7091 0.7381 0.7091 0.8421
No log 5.8039 296 0.6418 0.7651 0.6418 0.8011
No log 5.8431 298 0.6659 0.7586 0.6659 0.8160
No log 5.8824 300 0.6681 0.7310 0.6681 0.8174
No log 5.9216 302 0.7413 0.7248 0.7413 0.8610
No log 5.9608 304 0.9513 0.64 0.9513 0.9753
No log 6.0 306 0.9870 0.6494 0.9870 0.9935
No log 6.0392 308 0.7921 0.7190 0.7921 0.8900
No log 6.0784 310 0.6964 0.7922 0.6964 0.8345
No log 6.1176 312 0.7145 0.7712 0.7145 0.8453
No log 6.1569 314 0.6733 0.7771 0.6733 0.8205
No log 6.1961 316 0.6514 0.7564 0.6514 0.8071
No log 6.2353 318 0.6521 0.7564 0.6521 0.8075
No log 6.2745 320 0.6539 0.7595 0.6539 0.8086
No log 6.3137 322 0.6378 0.7516 0.6378 0.7986
No log 6.3529 324 0.6284 0.7436 0.6284 0.7927
No log 6.3922 326 0.6401 0.75 0.6401 0.8000
No log 6.4314 328 0.7138 0.7709 0.7138 0.8449
No log 6.4706 330 0.7302 0.7802 0.7302 0.8545
No log 6.5098 332 0.6442 0.7412 0.6442 0.8026
No log 6.5490 334 0.6544 0.7778 0.6544 0.8090
No log 6.5882 336 0.6798 0.7778 0.6798 0.8245
No log 6.6275 338 0.6721 0.7778 0.6721 0.8198
No log 6.6667 340 0.6989 0.7578 0.6989 0.8360
No log 6.7059 342 0.7840 0.7317 0.7840 0.8854
No log 6.7451 344 0.8839 0.6918 0.8839 0.9402
No log 6.7843 346 0.8690 0.6962 0.8690 0.9322
No log 6.8235 348 0.7903 0.7034 0.7903 0.8890
No log 6.8627 350 0.7430 0.7651 0.7430 0.8620
No log 6.9020 352 0.7097 0.7651 0.7097 0.8424
No log 6.9412 354 0.7007 0.7712 0.7007 0.8371
No log 6.9804 356 0.6908 0.7578 0.6908 0.8311
No log 7.0196 358 0.6902 0.7578 0.6902 0.8308
No log 7.0588 360 0.6917 0.7651 0.6917 0.8317
No log 7.0980 362 0.7438 0.7550 0.7438 0.8624
No log 7.1373 364 0.7470 0.7467 0.7470 0.8643
No log 7.1765 366 0.7644 0.7260 0.7644 0.8743
No log 7.2157 368 0.7681 0.7260 0.7681 0.8764
No log 7.2549 370 0.7730 0.7324 0.7730 0.8792
No log 7.2941 372 0.7877 0.7376 0.7877 0.8875
No log 7.3333 374 0.7767 0.7376 0.7767 0.8813
No log 7.3725 376 0.7320 0.7483 0.7320 0.8556
No log 7.4118 378 0.7189 0.7651 0.7189 0.8479
No log 7.4510 380 0.7290 0.7237 0.7290 0.8538
No log 7.4902 382 0.7262 0.7190 0.7262 0.8521
No log 7.5294 384 0.6914 0.7651 0.6914 0.8315
No log 7.5686 386 0.6862 0.7534 0.6862 0.8284
No log 7.6078 388 0.7035 0.7484 0.7035 0.8387
No log 7.6471 390 0.7280 0.7143 0.7280 0.8532
No log 7.6863 392 0.7034 0.7532 0.7034 0.8387
No log 7.7255 394 0.6827 0.7733 0.6827 0.8262
No log 7.7647 396 0.6679 0.7733 0.6679 0.8173
No log 7.8039 398 0.6769 0.7421 0.6769 0.8227
No log 7.8431 400 0.7913 0.7486 0.7913 0.8896
No log 7.8824 402 0.8414 0.7213 0.8414 0.9173
No log 7.9216 404 0.7581 0.7356 0.7581 0.8707
No log 7.9608 406 0.6363 0.7436 0.6363 0.7977
No log 8.0 408 0.6360 0.7586 0.6360 0.7975
No log 8.0392 410 0.7023 0.7413 0.7023 0.8380
No log 8.0784 412 0.7124 0.7324 0.7124 0.8440
No log 8.1176 414 0.6951 0.7483 0.6951 0.8337
No log 8.1569 416 0.7256 0.7034 0.7256 0.8518
No log 8.1961 418 0.7474 0.7260 0.7474 0.8645
No log 8.2353 420 0.7308 0.7123 0.7308 0.8549
No log 8.2745 422 0.7454 0.7432 0.7454 0.8634
No log 8.3137 424 0.7298 0.7347 0.7298 0.8543
No log 8.3529 426 0.6996 0.7568 0.6996 0.8364
No log 8.3922 428 0.6957 0.7651 0.6957 0.8341
No log 8.4314 430 0.7092 0.7347 0.7092 0.8421
No log 8.4706 432 0.7842 0.7297 0.7842 0.8855
No log 8.5098 434 0.9491 0.6380 0.9491 0.9742
No log 8.5490 436 1.0296 0.6353 1.0296 1.0147
No log 8.5882 438 0.9122 0.6353 0.9122 0.9551
No log 8.6275 440 0.7032 0.7383 0.7032 0.8386
No log 8.6667 442 0.6600 0.7568 0.6600 0.8124
No log 8.7059 444 0.6875 0.7448 0.6875 0.8292
No log 8.7451 446 0.7169 0.7361 0.7169 0.8467
No log 8.7843 448 0.7306 0.7397 0.7306 0.8548
No log 8.8235 450 0.7502 0.6853 0.7502 0.8661
No log 8.8627 452 0.7923 0.7211 0.7923 0.8901
No log 8.9020 454 0.7794 0.7297 0.7794 0.8828
No log 8.9412 456 0.7303 0.7042 0.7303 0.8546
No log 8.9804 458 0.7403 0.7273 0.7403 0.8604
No log 9.0196 460 0.7332 0.7361 0.7332 0.8563
No log 9.0588 462 0.7184 0.75 0.7184 0.8476
No log 9.0980 464 0.7511 0.7383 0.7511 0.8666
No log 9.1373 466 0.7497 0.7134 0.7497 0.8658
No log 9.1765 468 0.7377 0.7561 0.7377 0.8589
No log 9.2157 470 0.7193 0.7792 0.7193 0.8481
No log 9.2549 472 0.7409 0.7651 0.7409 0.8608
No log 9.2941 474 0.7633 0.7568 0.7633 0.8737
No log 9.3333 476 0.7811 0.7361 0.7811 0.8838
No log 9.3725 478 0.7855 0.7273 0.7855 0.8863
No log 9.4118 480 0.7442 0.7183 0.7442 0.8627
No log 9.4510 482 0.7074 0.7273 0.7074 0.8410
No log 9.4902 484 0.6953 0.7273 0.6953 0.8339
No log 9.5294 486 0.6781 0.7273 0.6781 0.8235
No log 9.5686 488 0.6782 0.7273 0.6782 0.8235
No log 9.6078 490 0.6667 0.7397 0.6667 0.8165
No log 9.6471 492 0.6805 0.7432 0.6805 0.8249
No log 9.6863 494 0.6736 0.7451 0.6736 0.8208
No log 9.7255 496 0.6533 0.7662 0.6533 0.8083
No log 9.7647 498 0.6098 0.7692 0.6098 0.7809
0.3533 9.8039 500 0.6119 0.7632 0.6119 0.7822
0.3533 9.8431 502 0.6358 0.7517 0.6358 0.7974
0.3533 9.8824 504 0.6863 0.7432 0.6863 0.8284
0.3533 9.9216 506 0.7412 0.7123 0.7412 0.8609
0.3533 9.9608 508 0.7411 0.7083 0.7411 0.8609
0.3533 10.0 510 0.7249 0.7310 0.7249 0.8514

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
182
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k11_task1_organization

Finetuned
(4222)
this model