ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k12_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7138
  • Qwk: 0.3723
  • Mse: 0.7138
  • Rmse: 0.8449

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0308 2 4.4492 -0.0170 4.4492 2.1093
No log 0.0615 4 2.5417 -0.0097 2.5417 1.5943
No log 0.0923 6 1.5398 -0.0147 1.5398 1.2409
No log 0.1231 8 1.3416 -0.0397 1.3416 1.1583
No log 0.1538 10 2.1101 -0.0288 2.1101 1.4526
No log 0.1846 12 1.4742 0.0054 1.4742 1.2142
No log 0.2154 14 0.9282 -0.0312 0.9282 0.9634
No log 0.2462 16 0.9208 0.0232 0.9208 0.9596
No log 0.2769 18 0.9214 -0.0050 0.9214 0.9599
No log 0.3077 20 1.0402 -0.1147 1.0402 1.0199
No log 0.3385 22 1.0725 -0.0900 1.0725 1.0356
No log 0.3692 24 0.9298 -0.0379 0.9298 0.9643
No log 0.4 26 0.8806 0.0147 0.8806 0.9384
No log 0.4308 28 0.8619 0.0224 0.8619 0.9284
No log 0.4615 30 0.8842 0.1437 0.8842 0.9403
No log 0.4923 32 0.9606 -0.0112 0.9606 0.9801
No log 0.5231 34 0.8907 0.1867 0.8907 0.9438
No log 0.5538 36 0.8529 0.1878 0.8529 0.9235
No log 0.5846 38 1.0217 0.0504 1.0217 1.0108
No log 0.6154 40 0.9835 0.0501 0.9835 0.9917
No log 0.6462 42 0.8426 0.2138 0.8426 0.9179
No log 0.6769 44 0.7937 0.1210 0.7937 0.8909
No log 0.7077 46 0.7818 0.1281 0.7818 0.8842
No log 0.7385 48 0.7616 0.1951 0.7616 0.8727
No log 0.7692 50 0.7668 0.2855 0.7668 0.8757
No log 0.8 52 0.7905 0.2504 0.7905 0.8891
No log 0.8308 54 0.8262 0.2569 0.8262 0.9090
No log 0.8615 56 0.9010 0.0968 0.9010 0.9492
No log 0.8923 58 0.9901 0.0726 0.9901 0.9950
No log 0.9231 60 1.0944 0.1102 1.0944 1.0461
No log 0.9538 62 1.0632 0.1102 1.0632 1.0311
No log 0.9846 64 0.8863 0.1610 0.8863 0.9415
No log 1.0154 66 0.7517 0.2452 0.7517 0.8670
No log 1.0462 68 0.7489 0.1737 0.7489 0.8654
No log 1.0769 70 0.7639 0.2197 0.7639 0.8740
No log 1.1077 72 0.7817 0.2362 0.7817 0.8841
No log 1.1385 74 0.7808 0.2379 0.7808 0.8836
No log 1.1692 76 0.7973 0.2834 0.7973 0.8929
No log 1.2 78 0.7650 0.3097 0.7650 0.8747
No log 1.2308 80 0.7399 0.2555 0.7399 0.8602
No log 1.2615 82 0.7411 0.2183 0.7411 0.8609
No log 1.2923 84 0.7161 0.3205 0.7161 0.8462
No log 1.3231 86 0.8733 0.2171 0.8733 0.9345
No log 1.3538 88 0.9170 0.1916 0.9170 0.9576
No log 1.3846 90 0.8450 0.2174 0.8450 0.9192
No log 1.4154 92 0.7721 0.3132 0.7721 0.8787
No log 1.4462 94 0.7655 0.3145 0.7655 0.8749
No log 1.4769 96 0.7695 0.3329 0.7695 0.8772
No log 1.5077 98 0.8827 0.2222 0.8827 0.9395
No log 1.5385 100 0.9617 0.1493 0.9617 0.9806
No log 1.5692 102 0.8110 0.2658 0.8110 0.9006
No log 1.6 104 0.8673 0.2351 0.8673 0.9313
No log 1.6308 106 0.8299 0.2481 0.8299 0.9110
No log 1.6615 108 0.8499 0.2428 0.8499 0.9219
No log 1.6923 110 0.8596 0.2139 0.8596 0.9271
No log 1.7231 112 0.8864 0.2992 0.8864 0.9415
No log 1.7538 114 0.9005 0.2955 0.9005 0.9489
No log 1.7846 116 0.9220 0.3017 0.9220 0.9602
No log 1.8154 118 0.9573 0.3531 0.9573 0.9784
No log 1.8462 120 1.0082 0.2808 1.0082 1.0041
No log 1.8769 122 0.9159 0.4150 0.9159 0.9570
No log 1.9077 124 1.0320 0.3311 1.0320 1.0159
No log 1.9385 126 1.0524 0.3428 1.0524 1.0258
No log 1.9692 128 0.8739 0.3858 0.8739 0.9348
No log 2.0 130 0.8631 0.3149 0.8631 0.9291
No log 2.0308 132 0.8593 0.3226 0.8593 0.9270
No log 2.0615 134 0.7776 0.3220 0.7776 0.8818
No log 2.0923 136 0.8319 0.2914 0.8319 0.9121
No log 2.1231 138 0.8920 0.2550 0.8920 0.9445
No log 2.1538 140 0.8480 0.2604 0.8480 0.9209
No log 2.1846 142 0.7721 0.2345 0.7721 0.8787
No log 2.2154 144 0.7566 0.2269 0.7566 0.8698
No log 2.2462 146 0.7743 0.2372 0.7743 0.8799
No log 2.2769 148 0.8110 0.2919 0.8110 0.9005
No log 2.3077 150 0.8803 0.3088 0.8803 0.9382
No log 2.3385 152 0.8059 0.2794 0.8059 0.8977
No log 2.3692 154 0.8414 0.3002 0.8414 0.9173
No log 2.4 156 0.8287 0.2955 0.8287 0.9103
No log 2.4308 158 0.8137 0.3737 0.8137 0.9020
No log 2.4615 160 0.9294 0.3304 0.9294 0.9641
No log 2.4923 162 0.8926 0.3332 0.8926 0.9448
No log 2.5231 164 0.7995 0.2945 0.7995 0.8942
No log 2.5538 166 1.0119 0.2951 1.0119 1.0059
No log 2.5846 168 1.0350 0.2743 1.0350 1.0173
No log 2.6154 170 0.8208 0.3172 0.8208 0.9060
No log 2.6462 172 0.8006 0.2688 0.8006 0.8948
No log 2.6769 174 0.8880 0.3270 0.8880 0.9423
No log 2.7077 176 0.8688 0.3814 0.8688 0.9321
No log 2.7385 178 0.8770 0.3662 0.8770 0.9365
No log 2.7692 180 0.8616 0.3006 0.8616 0.9282
No log 2.8 182 0.8427 0.3329 0.8427 0.9180
No log 2.8308 184 0.8058 0.3027 0.8058 0.8977
No log 2.8615 186 0.8552 0.2672 0.8552 0.9248
No log 2.8923 188 0.7992 0.2782 0.7992 0.8940
No log 2.9231 190 0.7621 0.2732 0.7621 0.8730
No log 2.9538 192 0.7610 0.2860 0.7610 0.8723
No log 2.9846 194 0.8378 0.2542 0.8378 0.9153
No log 3.0154 196 1.0574 0.2935 1.0574 1.0283
No log 3.0462 198 0.9936 0.3267 0.9936 0.9968
No log 3.0769 200 0.8066 0.3200 0.8066 0.8981
No log 3.1077 202 0.8471 0.3509 0.8471 0.9204
No log 3.1385 204 0.8942 0.3631 0.8942 0.9456
No log 3.1692 206 0.8062 0.3068 0.8062 0.8979
No log 3.2 208 0.7858 0.3550 0.7858 0.8864
No log 3.2308 210 0.7880 0.3325 0.7880 0.8877
No log 3.2615 212 0.8011 0.2980 0.8011 0.8951
No log 3.2923 214 1.0217 0.3574 1.0217 1.0108
No log 3.3231 216 1.2153 0.2762 1.2153 1.1024
No log 3.3538 218 1.1219 0.3104 1.1219 1.0592
No log 3.3846 220 0.8593 0.3446 0.8593 0.9270
No log 3.4154 222 0.7234 0.2831 0.7234 0.8505
No log 3.4462 224 0.7719 0.3471 0.7719 0.8786
No log 3.4769 226 0.7544 0.3555 0.7544 0.8686
No log 3.5077 228 0.7142 0.3005 0.7142 0.8451
No log 3.5385 230 0.7607 0.3270 0.7607 0.8722
No log 3.5692 232 0.7658 0.3380 0.7658 0.8751
No log 3.6 234 0.7627 0.2774 0.7627 0.8733
No log 3.6308 236 0.7844 0.3723 0.7844 0.8856
No log 3.6615 238 0.8113 0.4131 0.8113 0.9007
No log 3.6923 240 0.8034 0.3920 0.8034 0.8963
No log 3.7231 242 0.7685 0.3485 0.7685 0.8767
No log 3.7538 244 0.7752 0.3430 0.7752 0.8805
No log 3.7846 246 0.7610 0.2913 0.7610 0.8724
No log 3.8154 248 0.7144 0.3168 0.7144 0.8452
No log 3.8462 250 0.7055 0.3515 0.7055 0.8400
No log 3.8769 252 0.7386 0.3402 0.7386 0.8594
No log 3.9077 254 0.7635 0.2876 0.7635 0.8738
No log 3.9385 256 0.7430 0.3176 0.7430 0.8620
No log 3.9692 258 0.7799 0.3124 0.7799 0.8831
No log 4.0 260 0.8048 0.3124 0.8048 0.8971
No log 4.0308 262 0.8266 0.3507 0.8266 0.9092
No log 4.0615 264 0.8265 0.3815 0.8265 0.9091
No log 4.0923 266 0.8529 0.2912 0.8529 0.9235
No log 4.1231 268 0.8195 0.3048 0.8195 0.9053
No log 4.1538 270 0.8344 0.3335 0.8344 0.9134
No log 4.1846 272 0.9096 0.3055 0.9096 0.9537
No log 4.2154 274 0.8656 0.3030 0.8656 0.9304
No log 4.2462 276 0.7767 0.2849 0.7767 0.8813
No log 4.2769 278 0.7565 0.3277 0.7565 0.8698
No log 4.3077 280 0.7343 0.3325 0.7343 0.8569
No log 4.3385 282 0.7255 0.2880 0.7255 0.8517
No log 4.3692 284 0.7417 0.2671 0.7417 0.8612
No log 4.4 286 0.8515 0.3382 0.8515 0.9227
No log 4.4308 288 1.0241 0.3505 1.0241 1.0120
No log 4.4615 290 0.9668 0.3298 0.9668 0.9833
No log 4.4923 292 0.7994 0.4322 0.7994 0.8941
No log 4.5231 294 0.7584 0.4181 0.7584 0.8709
No log 4.5538 296 0.7632 0.4533 0.7632 0.8736
No log 4.5846 298 0.7627 0.4298 0.7627 0.8733
No log 4.6154 300 0.7698 0.4157 0.7698 0.8774
No log 4.6462 302 0.7800 0.4527 0.7800 0.8832
No log 4.6769 304 0.7352 0.4049 0.7352 0.8574
No log 4.7077 306 0.7539 0.4107 0.7539 0.8683
No log 4.7385 308 0.7478 0.4107 0.7478 0.8647
No log 4.7692 310 0.7340 0.4298 0.7340 0.8567
No log 4.8 312 0.7460 0.4349 0.7460 0.8637
No log 4.8308 314 0.7954 0.3365 0.7954 0.8918
No log 4.8615 316 0.7954 0.3499 0.7954 0.8918
No log 4.8923 318 0.7563 0.3637 0.7563 0.8697
No log 4.9231 320 0.7663 0.4210 0.7663 0.8754
No log 4.9538 322 0.7900 0.3913 0.7900 0.8888
No log 4.9846 324 0.7512 0.4186 0.7512 0.8667
No log 5.0154 326 0.7331 0.3416 0.7331 0.8562
No log 5.0462 328 0.7458 0.3418 0.7458 0.8636
No log 5.0769 330 0.7245 0.3636 0.7245 0.8511
No log 5.1077 332 0.7608 0.3525 0.7608 0.8722
No log 5.1385 334 0.7877 0.4080 0.7877 0.8875
No log 5.1692 336 0.7516 0.4063 0.7516 0.8670
No log 5.2 338 0.7268 0.3796 0.7268 0.8525
No log 5.2308 340 0.6943 0.4012 0.6943 0.8333
No log 5.2615 342 0.6846 0.4045 0.6846 0.8274
No log 5.2923 344 0.6883 0.4106 0.6883 0.8296
No log 5.3231 346 0.6968 0.4288 0.6968 0.8348
No log 5.3538 348 0.7007 0.4257 0.7007 0.8371
No log 5.3846 350 0.6992 0.4348 0.6992 0.8362
No log 5.4154 352 0.6982 0.4469 0.6982 0.8356
No log 5.4462 354 0.6953 0.4039 0.6953 0.8338
No log 5.4769 356 0.7285 0.4115 0.7285 0.8535
No log 5.5077 358 0.7201 0.3674 0.7201 0.8486
No log 5.5385 360 0.7018 0.4402 0.7018 0.8378
No log 5.5692 362 0.7181 0.3896 0.7181 0.8474
No log 5.6 364 0.7347 0.4239 0.7347 0.8572
No log 5.6308 366 0.7336 0.4099 0.7336 0.8565
No log 5.6615 368 0.7137 0.4147 0.7137 0.8448
No log 5.6923 370 0.7039 0.4037 0.7039 0.8390
No log 5.7231 372 0.6932 0.4228 0.6932 0.8326
No log 5.7538 374 0.6934 0.4286 0.6934 0.8327
No log 5.7846 376 0.6929 0.4177 0.6929 0.8324
No log 5.8154 378 0.7096 0.4341 0.7096 0.8424
No log 5.8462 380 0.7335 0.4891 0.7335 0.8564
No log 5.8769 382 0.7407 0.4721 0.7407 0.8607
No log 5.9077 384 0.7515 0.4471 0.7515 0.8669
No log 5.9385 386 0.7779 0.4241 0.7779 0.8820
No log 5.9692 388 0.8381 0.3585 0.8381 0.9155
No log 6.0 390 0.8637 0.3034 0.8637 0.9293
No log 6.0308 392 0.8555 0.2562 0.8555 0.9249
No log 6.0615 394 0.7635 0.2419 0.7635 0.8738
No log 6.0923 396 0.7278 0.2917 0.7278 0.8531
No log 6.1231 398 0.7373 0.2669 0.7373 0.8587
No log 6.1538 400 0.7380 0.3091 0.7380 0.8591
No log 6.1846 402 0.7613 0.3643 0.7613 0.8725
No log 6.2154 404 0.7964 0.4011 0.7964 0.8924
No log 6.2462 406 0.8555 0.4170 0.8555 0.9249
No log 6.2769 408 0.8665 0.3806 0.8665 0.9308
No log 6.3077 410 0.7770 0.4154 0.7770 0.8815
No log 6.3385 412 0.7339 0.3729 0.7339 0.8567
No log 6.3692 414 0.7048 0.3651 0.7048 0.8395
No log 6.4 416 0.7060 0.2982 0.7060 0.8403
No log 6.4308 418 0.7268 0.3819 0.7268 0.8525
No log 6.4615 420 0.7134 0.3289 0.7134 0.8446
No log 6.4923 422 0.6989 0.3609 0.6989 0.8360
No log 6.5231 424 0.7439 0.3752 0.7439 0.8625
No log 6.5538 426 0.7817 0.3595 0.7817 0.8841
No log 6.5846 428 0.7305 0.4031 0.7305 0.8547
No log 6.6154 430 0.7173 0.4422 0.7173 0.8469
No log 6.6462 432 0.8717 0.3845 0.8717 0.9337
No log 6.6769 434 1.0609 0.3255 1.0609 1.0300
No log 6.7077 436 1.0133 0.3151 1.0133 1.0066
No log 6.7385 438 0.8349 0.3982 0.8349 0.9137
No log 6.7692 440 0.7387 0.3746 0.7387 0.8595
No log 6.8 442 0.7532 0.4161 0.7532 0.8678
No log 6.8308 444 0.7430 0.3769 0.7430 0.8620
No log 6.8615 446 0.7808 0.3349 0.7808 0.8837
No log 6.8923 448 0.8365 0.3401 0.8365 0.9146
No log 6.9231 450 0.8116 0.3368 0.8116 0.9009
No log 6.9538 452 0.7391 0.3481 0.7391 0.8597
No log 6.9846 454 0.7250 0.3633 0.7250 0.8515
No log 7.0154 456 0.7536 0.3082 0.7536 0.8681
No log 7.0462 458 0.8137 0.2689 0.8137 0.9020
No log 7.0769 460 0.8287 0.2901 0.8287 0.9104
No log 7.1077 462 0.7849 0.3356 0.7849 0.8859
No log 7.1385 464 0.8095 0.4082 0.8095 0.8997
No log 7.1692 466 0.8118 0.4545 0.8118 0.9010
No log 7.2 468 0.8081 0.4493 0.8081 0.8989
No log 7.2308 470 0.7688 0.4392 0.7688 0.8768
No log 7.2615 472 0.7482 0.4633 0.7482 0.8650
No log 7.2923 474 0.7168 0.4593 0.7168 0.8467
No log 7.3231 476 0.7170 0.4552 0.7170 0.8467
No log 7.3538 478 0.7526 0.4120 0.7526 0.8676
No log 7.3846 480 0.8073 0.3358 0.8073 0.8985
No log 7.4154 482 0.8712 0.3482 0.8712 0.9334
No log 7.4462 484 0.8412 0.3564 0.8412 0.9172
No log 7.4769 486 0.7768 0.3937 0.7768 0.8813
No log 7.5077 488 0.7510 0.4203 0.7510 0.8666
No log 7.5385 490 0.7532 0.4115 0.7532 0.8679
No log 7.5692 492 0.7559 0.3608 0.7559 0.8694
No log 7.6 494 0.7083 0.4286 0.7083 0.8416
No log 7.6308 496 0.6871 0.4303 0.6871 0.8289
No log 7.6615 498 0.6901 0.4293 0.6901 0.8307
0.4233 7.6923 500 0.7212 0.4144 0.7212 0.8492
0.4233 7.7231 502 0.7655 0.4075 0.7655 0.8749
0.4233 7.7538 504 0.7765 0.3992 0.7765 0.8812
0.4233 7.7846 506 0.7244 0.4261 0.7244 0.8511
0.4233 7.8154 508 0.6941 0.4563 0.6941 0.8331
0.4233 7.8462 510 0.7029 0.4762 0.7029 0.8384
0.4233 7.8769 512 0.6955 0.4558 0.6955 0.8340
0.4233 7.9077 514 0.7221 0.3986 0.7221 0.8498
0.4233 7.9385 516 0.7905 0.3941 0.7905 0.8891
0.4233 7.9692 518 0.7855 0.3755 0.7855 0.8863
0.4233 8.0 520 0.7412 0.4472 0.7412 0.8610
0.4233 8.0308 522 0.7488 0.4477 0.7488 0.8653
0.4233 8.0615 524 0.7469 0.4211 0.7469 0.8642
0.4233 8.0923 526 0.7237 0.4065 0.7237 0.8507
0.4233 8.1231 528 0.7363 0.4210 0.7363 0.8581
0.4233 8.1538 530 0.8110 0.3392 0.8110 0.9006
0.4233 8.1846 532 0.8111 0.3126 0.8111 0.9006
0.4233 8.2154 534 0.7559 0.3633 0.7559 0.8694
0.4233 8.2462 536 0.7137 0.4823 0.7137 0.8448
0.4233 8.2769 538 0.7159 0.4107 0.7159 0.8461
0.4233 8.3077 540 0.7317 0.4295 0.7317 0.8554
0.4233 8.3385 542 0.7413 0.4673 0.7413 0.8610
0.4233 8.3692 544 0.8046 0.3542 0.8046 0.8970
0.4233 8.4 546 0.8613 0.3351 0.8613 0.9281
0.4233 8.4308 548 0.8322 0.3283 0.8322 0.9123
0.4233 8.4615 550 0.7438 0.3914 0.7437 0.8624
0.4233 8.4923 552 0.7138 0.3723 0.7138 0.8449

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
5
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k12_task2_organization

Finetuned
(4222)
this model