ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k9_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8754
  • Qwk: 0.6423
  • Mse: 0.8754
  • Rmse: 0.9356

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0476 2 6.9750 0.0118 6.9750 2.6410
No log 0.0952 4 5.0410 0.0366 5.0410 2.2452
No log 0.1429 6 3.4593 0.0556 3.4593 1.8599
No log 0.1905 8 3.3329 0.0452 3.3329 1.8256
No log 0.2381 10 2.2933 0.1702 2.2933 1.5143
No log 0.2857 12 1.9206 0.2114 1.9206 1.3859
No log 0.3333 14 1.8039 0.1835 1.8039 1.3431
No log 0.3810 16 1.7404 0.2321 1.7404 1.3193
No log 0.4286 18 2.0744 0.2443 2.0744 1.4403
No log 0.4762 20 2.3500 0.1449 2.3500 1.5330
No log 0.5238 22 2.5463 0.0544 2.5463 1.5957
No log 0.5714 24 2.3239 0.1286 2.3239 1.5244
No log 0.6190 26 2.0931 0.2353 2.0931 1.4468
No log 0.6667 28 2.0415 0.2353 2.0415 1.4288
No log 0.7143 30 2.7819 0.2373 2.7819 1.6679
No log 0.7619 32 2.9973 0.2256 2.9973 1.7313
No log 0.8095 34 2.8801 0.2591 2.8801 1.6971
No log 0.8571 36 2.8759 0.2944 2.8759 1.6959
No log 0.9048 38 2.5903 0.3385 2.5903 1.6094
No log 0.9524 40 2.2977 0.3175 2.2977 1.5158
No log 1.0 42 2.0808 0.3933 2.0808 1.4425
No log 1.0476 44 1.7309 0.4277 1.7309 1.3156
No log 1.0952 46 1.5181 0.3776 1.5181 1.2321
No log 1.1429 48 1.5250 0.4430 1.5250 1.2349
No log 1.1905 50 1.5530 0.4324 1.5530 1.2462
No log 1.2381 52 1.8007 0.4431 1.8007 1.3419
No log 1.2857 54 1.7727 0.4795 1.7727 1.3314
No log 1.3333 56 1.8086 0.4535 1.8086 1.3449
No log 1.3810 58 1.5357 0.4810 1.5357 1.2392
No log 1.4286 60 1.5242 0.5153 1.5242 1.2346
No log 1.4762 62 1.3767 0.5432 1.3767 1.1733
No log 1.5238 64 1.1230 0.6301 1.1230 1.0597
No log 1.5714 66 1.1349 0.6358 1.1349 1.0653
No log 1.6190 68 1.3504 0.5818 1.3504 1.1621
No log 1.6667 70 1.9950 0.4041 1.9950 1.4125
No log 1.7143 72 2.2144 0.3958 2.2144 1.4881
No log 1.7619 74 1.9762 0.4271 1.9762 1.4058
No log 1.8095 76 1.5448 0.4884 1.5448 1.2429
No log 1.8571 78 1.3044 0.5789 1.3044 1.1421
No log 1.9048 80 1.3322 0.5256 1.3322 1.1542
No log 1.9524 82 1.7734 0.4246 1.7734 1.3317
No log 2.0 84 1.7694 0.4270 1.7694 1.3302
No log 2.0476 86 1.2776 0.5526 1.2776 1.1303
No log 2.0952 88 1.2590 0.5306 1.2590 1.1221
No log 2.1429 90 1.4952 0.4667 1.4952 1.2228
No log 2.1905 92 1.4313 0.5170 1.4313 1.1964
No log 2.2381 94 1.5900 0.4314 1.5900 1.2609
No log 2.2857 96 2.1639 0.4041 2.1639 1.4710
No log 2.3333 98 2.6770 0.3529 2.6770 1.6362
No log 2.3810 100 2.3685 0.3789 2.3685 1.5390
No log 2.4286 102 1.8433 0.4025 1.8433 1.3577
No log 2.4762 104 1.5279 0.4397 1.5279 1.2361
No log 2.5238 106 1.5146 0.3972 1.5146 1.2307
No log 2.5714 108 1.5971 0.4459 1.5971 1.2638
No log 2.6190 110 1.6674 0.4533 1.6674 1.2913
No log 2.6667 112 2.0044 0.4022 2.0044 1.4158
No log 2.7143 114 1.6884 0.5029 1.6884 1.2994
No log 2.7619 116 1.0292 0.6447 1.0292 1.0145
No log 2.8095 118 0.9072 0.6241 0.9072 0.9525
No log 2.8571 120 0.8979 0.6479 0.8979 0.9476
No log 2.9048 122 0.8832 0.6933 0.8832 0.9398
No log 2.9524 124 1.1789 0.6286 1.1789 1.0858
No log 3.0 126 1.4484 0.5574 1.4484 1.2035
No log 3.0476 128 1.2739 0.6067 1.2739 1.1287
No log 3.0952 130 0.9452 0.7011 0.9452 0.9722
No log 3.1429 132 0.7612 0.7179 0.7612 0.8725
No log 3.1905 134 0.7085 0.7347 0.7085 0.8417
No log 3.2381 136 0.7231 0.7260 0.7231 0.8504
No log 3.2857 138 1.0408 0.6497 1.0408 1.0202
No log 3.3333 140 1.6274 0.4471 1.6274 1.2757
No log 3.3810 142 1.7251 0.4693 1.7251 1.3134
No log 3.4286 144 1.2707 0.5955 1.2707 1.1273
No log 3.4762 146 1.0455 0.6918 1.0455 1.0225
No log 3.5238 148 1.0175 0.7089 1.0175 1.0087
No log 3.5714 150 1.0982 0.6708 1.0982 1.0479
No log 3.6190 152 1.2094 0.5965 1.2094 1.0997
No log 3.6667 154 1.3326 0.5380 1.3326 1.1544
No log 3.7143 156 1.4045 0.4971 1.4045 1.1851
No log 3.7619 158 1.0579 0.6420 1.0579 1.0286
No log 3.8095 160 0.9215 0.7308 0.9215 0.9599
No log 3.8571 162 0.8751 0.7329 0.8751 0.9355
No log 3.9048 164 0.8445 0.7329 0.8445 0.9190
No log 3.9524 166 0.9109 0.6667 0.9109 0.9544
No log 4.0 168 0.9537 0.6792 0.9537 0.9766
No log 4.0476 170 0.9077 0.7117 0.9077 0.9528
No log 4.0952 172 0.8368 0.7284 0.8368 0.9147
No log 4.1429 174 0.7632 0.7320 0.7632 0.8736
No log 4.1905 176 0.7673 0.75 0.7673 0.8760
No log 4.2381 178 0.8080 0.7397 0.8080 0.8989
No log 4.2857 180 0.8267 0.7273 0.8267 0.9092
No log 4.3333 182 0.8555 0.6846 0.8555 0.9249
No log 4.3810 184 0.9269 0.6533 0.9269 0.9628
No log 4.4286 186 0.9512 0.6533 0.9512 0.9753
No log 4.4762 188 0.9556 0.6443 0.9556 0.9775
No log 4.5238 190 1.0180 0.6443 1.0180 1.0090
No log 4.5714 192 1.1400 0.5963 1.1400 1.0677
No log 4.6190 194 1.1124 0.6588 1.1124 1.0547
No log 4.6667 196 1.1483 0.6517 1.1483 1.0716
No log 4.7143 198 1.0600 0.6474 1.0600 1.0296
No log 4.7619 200 1.1124 0.6413 1.1124 1.0547
No log 4.8095 202 0.9726 0.6550 0.9726 0.9862
No log 4.8571 204 0.9037 0.7108 0.9037 0.9506
No log 4.9048 206 0.8670 0.7407 0.8670 0.9311
No log 4.9524 208 0.8012 0.7643 0.8012 0.8951
No log 5.0 210 0.7974 0.7333 0.7974 0.8930
No log 5.0476 212 0.8415 0.7308 0.8415 0.9173
No log 5.0952 214 0.9941 0.6752 0.9941 0.9971
No log 5.1429 216 1.0513 0.6667 1.0513 1.0253
No log 5.1905 218 0.8830 0.7075 0.8830 0.9397
No log 5.2381 220 0.8015 0.7297 0.8015 0.8953
No log 5.2857 222 0.7924 0.7383 0.7924 0.8902
No log 5.3333 224 0.7791 0.7285 0.7791 0.8827
No log 5.3810 226 0.7685 0.7285 0.7685 0.8767
No log 5.4286 228 0.8111 0.6962 0.8111 0.9006
No log 5.4762 230 0.8947 0.6875 0.8947 0.9459
No log 5.5238 232 1.1067 0.6477 1.1067 1.0520
No log 5.5714 234 1.0729 0.6477 1.0729 1.0358
No log 5.6190 236 0.8277 0.7125 0.8277 0.9098
No log 5.6667 238 0.7447 0.7329 0.7447 0.8629
No log 5.7143 240 0.7977 0.7205 0.7977 0.8931
No log 5.7619 242 0.9239 0.6936 0.9239 0.9612
No log 5.8095 244 1.0691 0.6780 1.0691 1.0340
No log 5.8571 246 1.2078 0.6243 1.2078 1.0990
No log 5.9048 248 1.1843 0.6597 1.1843 1.0883
No log 5.9524 250 1.1199 0.6667 1.1199 1.0583
No log 6.0 252 1.0323 0.6705 1.0323 1.0160
No log 6.0476 254 0.8591 0.7186 0.8591 0.9269
No log 6.0952 256 0.7533 0.7485 0.7533 0.8680
No log 6.1429 258 0.7475 0.7561 0.7475 0.8646
No log 6.1905 260 0.7284 0.7105 0.7284 0.8534
No log 6.2381 262 0.7226 0.6980 0.7226 0.8500
No log 6.2857 264 0.7837 0.7439 0.7837 0.8853
No log 6.3333 266 1.0233 0.6864 1.0233 1.0116
No log 6.3810 268 0.9712 0.6707 0.9712 0.9855
No log 6.4286 270 0.7567 0.7261 0.7567 0.8699
No log 6.4762 272 0.7144 0.6849 0.7144 0.8452
No log 6.5238 274 0.7848 0.7229 0.7848 0.8859
No log 6.5714 276 1.0017 0.6854 1.0017 1.0009
No log 6.6190 278 1.3939 0.6203 1.3939 1.1807
No log 6.6667 280 1.6613 0.5685 1.6613 1.2889
No log 6.7143 282 1.3803 0.5957 1.3803 1.1749
No log 6.7619 284 1.1198 0.6404 1.1198 1.0582
No log 6.8095 286 0.8219 0.7381 0.8219 0.9066
No log 6.8571 288 0.7249 0.7248 0.7249 0.8514
No log 6.9048 290 0.7353 0.7483 0.7353 0.8575
No log 6.9524 292 0.7332 0.7226 0.7332 0.8563
No log 7.0 294 0.7654 0.7407 0.7654 0.8749
No log 7.0476 296 0.8554 0.7066 0.8554 0.9249
No log 7.0952 298 0.9037 0.7119 0.9037 0.9506
No log 7.1429 300 0.9421 0.7111 0.9421 0.9706
No log 7.1905 302 0.9408 0.6932 0.9408 0.9699
No log 7.2381 304 0.9353 0.6584 0.9353 0.9671
No log 7.2857 306 0.8542 0.6803 0.8542 0.9242
No log 7.3333 308 0.7544 0.7397 0.7544 0.8685
No log 7.3810 310 0.7369 0.7297 0.7369 0.8584
No log 7.4286 312 0.7549 0.7114 0.7549 0.8688
No log 7.4762 314 0.7287 0.7692 0.7287 0.8536
No log 7.5238 316 0.7729 0.7630 0.7729 0.8791
No log 7.5714 318 0.7994 0.7416 0.7994 0.8941
No log 7.6190 320 0.8016 0.7558 0.8016 0.8953
No log 7.6667 322 0.8376 0.7241 0.8376 0.9152
No log 7.7143 324 0.8168 0.7337 0.8168 0.9038
No log 7.7619 326 0.7755 0.7765 0.7755 0.8806
No log 7.8095 328 0.7642 0.7765 0.7642 0.8742
No log 7.8571 330 0.7622 0.7765 0.7622 0.8731
No log 7.9048 332 0.7425 0.7665 0.7425 0.8617
No log 7.9524 334 0.7307 0.7665 0.7307 0.8548
No log 8.0 336 0.7250 0.7590 0.7250 0.8515
No log 8.0476 338 0.7243 0.7456 0.7243 0.8511
No log 8.0952 340 0.7700 0.7368 0.7700 0.8775
No log 8.1429 342 0.7736 0.7425 0.7736 0.8795
No log 8.1905 344 0.7574 0.7826 0.7574 0.8703
No log 8.2381 346 0.7605 0.7871 0.7605 0.8720
No log 8.2857 348 0.7684 0.7467 0.7684 0.8766
No log 8.3333 350 0.7708 0.7451 0.7708 0.8780
No log 8.3810 352 0.8069 0.7097 0.8069 0.8983
No log 8.4286 354 0.7580 0.7297 0.7580 0.8707
No log 8.4762 356 0.6978 0.7310 0.6978 0.8354
No log 8.5238 358 0.6994 0.7042 0.6994 0.8363
No log 8.5714 360 0.6718 0.7123 0.6718 0.8196
No log 8.6190 362 0.6456 0.7432 0.6456 0.8035
No log 8.6667 364 0.6370 0.8148 0.6370 0.7981
No log 8.7143 366 0.7031 0.7701 0.7031 0.8385
No log 8.7619 368 0.6557 0.7886 0.6557 0.8097
No log 8.8095 370 0.6241 0.8111 0.6241 0.7900
No log 8.8571 372 0.6786 0.7879 0.6786 0.8238
No log 8.9048 374 0.7469 0.6883 0.7469 0.8643
No log 8.9524 376 0.7444 0.7308 0.7444 0.8628
No log 9.0 378 0.7408 0.7799 0.7408 0.8607
No log 9.0476 380 0.7754 0.7975 0.7754 0.8806
No log 9.0952 382 0.7852 0.7722 0.7852 0.8861
No log 9.1429 384 0.7714 0.7799 0.7714 0.8783
No log 9.1905 386 0.7984 0.7229 0.7984 0.8935
No log 9.2381 388 0.8803 0.7305 0.8803 0.9382
No log 9.2857 390 0.9579 0.7283 0.9579 0.9787
No log 9.3333 392 0.9789 0.7209 0.9789 0.9894
No log 9.3810 394 0.8631 0.7229 0.8631 0.9291
No log 9.4286 396 0.7800 0.75 0.7800 0.8832
No log 9.4762 398 0.7638 0.7586 0.7638 0.8740
No log 9.5238 400 0.7499 0.7785 0.7499 0.8660
No log 9.5714 402 0.7440 0.7712 0.7440 0.8626
No log 9.6190 404 0.7487 0.7771 0.7487 0.8653
No log 9.6667 406 0.7453 0.7974 0.7453 0.8633
No log 9.7143 408 0.7307 0.7703 0.7307 0.8548
No log 9.7619 410 0.7284 0.7361 0.7284 0.8534
No log 9.8095 412 0.7343 0.7361 0.7343 0.8569
No log 9.8571 414 0.7706 0.7361 0.7706 0.8779
No log 9.9048 416 0.7804 0.7552 0.7804 0.8834
No log 9.9524 418 0.7981 0.7162 0.7981 0.8934
No log 10.0 420 0.8535 0.7134 0.8535 0.9239
No log 10.0476 422 0.9423 0.6946 0.9423 0.9707
No log 10.0952 424 0.9125 0.7263 0.9125 0.9552
No log 10.1429 426 0.7716 0.7229 0.7716 0.8784
No log 10.1905 428 0.7406 0.7547 0.7406 0.8606
No log 10.2381 430 0.7399 0.7595 0.7399 0.8602
No log 10.2857 432 0.7396 0.7662 0.7396 0.8600
No log 10.3333 434 0.7369 0.7625 0.7369 0.8585
No log 10.3810 436 0.8506 0.7337 0.8506 0.9223
No log 10.4286 438 0.9715 0.6821 0.9715 0.9856
No log 10.4762 440 0.9281 0.6628 0.9281 0.9634
No log 10.5238 442 0.8230 0.7337 0.8230 0.9072
No log 10.5714 444 0.7509 0.7712 0.7509 0.8666
No log 10.6190 446 0.7523 0.7662 0.7523 0.8674
No log 10.6667 448 0.7523 0.7692 0.7523 0.8674
No log 10.7143 450 0.8076 0.7399 0.8076 0.8987
No log 10.7619 452 0.8587 0.7159 0.8587 0.9267
No log 10.8095 454 0.8339 0.7168 0.8339 0.9132
No log 10.8571 456 0.7697 0.7654 0.7697 0.8773
No log 10.9048 458 0.7441 0.75 0.7441 0.8626
No log 10.9524 460 0.7346 0.75 0.7346 0.8571
No log 11.0 462 0.7104 0.7483 0.7104 0.8428
No log 11.0476 464 0.7037 0.7619 0.7037 0.8389
No log 11.0952 466 0.7059 0.7552 0.7059 0.8402
No log 11.1429 468 0.7247 0.7313 0.7247 0.8513
No log 11.1905 470 0.7310 0.7194 0.7310 0.8550
No log 11.2381 472 0.7299 0.7310 0.7299 0.8543
No log 11.2857 474 0.7443 0.7114 0.7443 0.8627
No log 11.3333 476 0.7910 0.7407 0.7910 0.8894
No log 11.3810 478 0.8077 0.7586 0.8077 0.8987
No log 11.4286 480 0.8838 0.7191 0.8838 0.9401
No log 11.4762 482 0.8112 0.7444 0.8112 0.9007
No log 11.5238 484 0.7106 0.7529 0.7106 0.8430
No log 11.5714 486 0.6932 0.7425 0.6932 0.8326
No log 11.6190 488 0.7690 0.7294 0.7690 0.8769
No log 11.6667 490 0.8087 0.7294 0.8087 0.8993
No log 11.7143 492 0.7502 0.7778 0.7502 0.8661
No log 11.7619 494 0.7470 0.7771 0.7470 0.8643
No log 11.8095 496 0.7682 0.7347 0.7682 0.8765
No log 11.8571 498 0.8144 0.7436 0.8144 0.9024
0.3764 11.9048 500 0.9148 0.7030 0.9148 0.9565
0.3764 11.9524 502 0.9487 0.7101 0.9487 0.9740
0.3764 12.0 504 1.0571 0.6667 1.0571 1.0282
0.3764 12.0476 506 1.0078 0.6705 1.0078 1.0039
0.3764 12.0952 508 0.8572 0.7262 0.8572 0.9258
0.3764 12.1429 510 0.7821 0.7468 0.7821 0.8843
0.3764 12.1905 512 0.7756 0.7484 0.7756 0.8807
0.3764 12.2381 514 0.7988 0.7515 0.7988 0.8938
0.3764 12.2857 516 0.8503 0.7439 0.8503 0.9221
0.3764 12.3333 518 0.8357 0.7362 0.8357 0.9142
0.3764 12.3810 520 0.8407 0.7105 0.8407 0.9169
0.3764 12.4286 522 0.8073 0.7333 0.8073 0.8985
0.3764 12.4762 524 0.7782 0.7785 0.7782 0.8821
0.3764 12.5238 526 0.7599 0.7550 0.7599 0.8717
0.3764 12.5714 528 0.7614 0.7632 0.7614 0.8726
0.3764 12.6190 530 0.7525 0.7613 0.7525 0.8675
0.3764 12.6667 532 0.7721 0.7515 0.7721 0.8787
0.3764 12.7143 534 0.8077 0.7442 0.8077 0.8987
0.3764 12.7619 536 0.8937 0.7059 0.8937 0.9453
0.3764 12.8095 538 0.9330 0.6905 0.9330 0.9659
0.3764 12.8571 540 0.8964 0.6835 0.8964 0.9468
0.3764 12.9048 542 0.8671 0.7067 0.8671 0.9312
0.3764 12.9524 544 0.8228 0.7260 0.8228 0.9071
0.3764 13.0 546 0.7800 0.7260 0.7800 0.8832
0.3764 13.0476 548 0.7708 0.7260 0.7708 0.8780
0.3764 13.0952 550 0.7798 0.7027 0.7798 0.8830
0.3764 13.1429 552 0.8392 0.6957 0.8392 0.9161
0.3764 13.1905 554 0.8924 0.6906 0.8924 0.9446
0.3764 13.2381 556 0.8984 0.6765 0.8984 0.9478
0.3764 13.2857 558 0.8754 0.6423 0.8754 0.9356

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
182
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k9_task1_organization

Finetuned
(4222)
this model