ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k16_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7162
  • Qwk: 0.3504
  • Mse: 0.7162
  • Rmse: 0.8463

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0235 2 4.3267 -0.0197 4.3267 2.0801
No log 0.0471 4 2.1727 0.0784 2.1727 1.4740
No log 0.0706 6 1.1553 -0.0217 1.1553 1.0748
No log 0.0941 8 0.8811 0.0735 0.8811 0.9387
No log 0.1176 10 0.8855 0.1163 0.8855 0.9410
No log 0.1412 12 0.9492 0.0093 0.9492 0.9743
No log 0.1647 14 1.2053 0.0312 1.2053 1.0978
No log 0.1882 16 1.2658 0.0262 1.2658 1.1251
No log 0.2118 18 1.0324 0.0257 1.0324 1.0160
No log 0.2353 20 0.8690 0.2324 0.8690 0.9322
No log 0.2588 22 0.8716 0.1107 0.8716 0.9336
No log 0.2824 24 0.8645 0.1657 0.8645 0.9298
No log 0.3059 26 0.8783 0.0937 0.8783 0.9372
No log 0.3294 28 0.8689 0.0660 0.8689 0.9321
No log 0.3529 30 0.8908 0.0781 0.8908 0.9438
No log 0.3765 32 0.8826 0.2056 0.8826 0.9395
No log 0.4 34 0.8887 0.1735 0.8887 0.9427
No log 0.4235 36 0.9533 0.0855 0.9533 0.9764
No log 0.4471 38 1.0007 0.0444 1.0007 1.0004
No log 0.4706 40 1.1457 -0.0493 1.1457 1.0704
No log 0.4941 42 1.1005 0.0119 1.1005 1.0491
No log 0.5176 44 0.8991 0.0675 0.8991 0.9482
No log 0.5412 46 0.8516 0.1101 0.8516 0.9228
No log 0.5647 48 0.8485 0.1347 0.8485 0.9212
No log 0.5882 50 0.8509 0.1651 0.8509 0.9225
No log 0.6118 52 0.8707 0.1331 0.8707 0.9331
No log 0.6353 54 0.9146 0.0912 0.9146 0.9563
No log 0.6588 56 0.9072 0.1503 0.9072 0.9525
No log 0.6824 58 0.9546 0.1256 0.9546 0.9771
No log 0.7059 60 0.9464 0.1550 0.9464 0.9728
No log 0.7294 62 0.8747 0.1475 0.8747 0.9352
No log 0.7529 64 0.8655 0.1633 0.8655 0.9303
No log 0.7765 66 1.0379 0.0603 1.0379 1.0188
No log 0.8 68 1.1191 0.1310 1.1191 1.0579
No log 0.8235 70 0.9724 0.1583 0.9724 0.9861
No log 0.8471 72 0.8277 0.0797 0.8277 0.9098
No log 0.8706 74 0.8054 0.1571 0.8054 0.8974
No log 0.8941 76 0.9201 0.1205 0.9201 0.9592
No log 0.9176 78 1.1725 0.0630 1.1725 1.0828
No log 0.9412 80 1.2267 0.0305 1.2267 1.1075
No log 0.9647 82 1.0185 0.1016 1.0185 1.0092
No log 0.9882 84 0.8412 0.1931 0.8412 0.9172
No log 1.0118 86 0.7821 0.2452 0.7821 0.8843
No log 1.0353 88 0.8116 0.2486 0.8116 0.9009
No log 1.0588 90 0.8462 0.2071 0.8462 0.9199
No log 1.0824 92 0.8336 0.2283 0.8336 0.9130
No log 1.1059 94 0.8300 0.2862 0.8300 0.9110
No log 1.1294 96 0.8580 0.2519 0.8580 0.9263
No log 1.1529 98 0.8391 0.2469 0.8391 0.9160
No log 1.1765 100 0.8435 0.2136 0.8435 0.9184
No log 1.2 102 0.8185 0.2756 0.8185 0.9047
No log 1.2235 104 0.8319 0.2149 0.8319 0.9121
No log 1.2471 106 0.7989 0.2988 0.7989 0.8938
No log 1.2706 108 0.8103 0.1861 0.8103 0.9002
No log 1.2941 110 0.8527 0.2429 0.8527 0.9234
No log 1.3176 112 0.7947 0.2764 0.7947 0.8915
No log 1.3412 114 0.7858 0.3001 0.7858 0.8865
No log 1.3647 116 0.8388 0.2455 0.8388 0.9159
No log 1.3882 118 0.7915 0.3186 0.7915 0.8897
No log 1.4118 120 0.7882 0.2634 0.7882 0.8878
No log 1.4353 122 0.7852 0.2845 0.7852 0.8861
No log 1.4588 124 0.8095 0.3441 0.8095 0.8997
No log 1.4824 126 0.8321 0.3134 0.8321 0.9122
No log 1.5059 128 1.0706 0.3598 1.0706 1.0347
No log 1.5294 130 1.0935 0.3475 1.0935 1.0457
No log 1.5529 132 0.8293 0.4100 0.8293 0.9107
No log 1.5765 134 0.7863 0.3380 0.7863 0.8867
No log 1.6 136 0.7664 0.3663 0.7664 0.8754
No log 1.6235 138 0.7828 0.3906 0.7828 0.8848
No log 1.6471 140 1.0595 0.3607 1.0595 1.0293
No log 1.6706 142 1.1994 0.2764 1.1994 1.0952
No log 1.6941 144 1.0470 0.3601 1.0470 1.0232
No log 1.7176 146 0.8330 0.3762 0.8330 0.9127
No log 1.7412 148 0.7921 0.3708 0.7921 0.8900
No log 1.7647 150 0.8153 0.3620 0.8153 0.9029
No log 1.7882 152 0.9238 0.3578 0.9238 0.9611
No log 1.8118 154 1.1117 0.3207 1.1117 1.0544
No log 1.8353 156 1.0058 0.3259 1.0058 1.0029
No log 1.8588 158 0.9387 0.3177 0.9387 0.9689
No log 1.8824 160 1.0153 0.3259 1.0153 1.0076
No log 1.9059 162 1.1930 0.2753 1.1930 1.0922
No log 1.9294 164 1.1106 0.2983 1.1106 1.0539
No log 1.9529 166 0.8096 0.3451 0.8096 0.8998
No log 1.9765 168 0.7098 0.3283 0.7098 0.8425
No log 2.0 170 0.7143 0.3206 0.7143 0.8452
No log 2.0235 172 0.6658 0.3658 0.6658 0.8160
No log 2.0471 174 0.7283 0.3836 0.7283 0.8534
No log 2.0706 176 0.8396 0.3922 0.8396 0.9163
No log 2.0941 178 0.8699 0.4201 0.8699 0.9327
No log 2.1176 180 0.7710 0.3931 0.7710 0.8780
No log 2.1412 182 0.7292 0.3296 0.7292 0.8539
No log 2.1647 184 0.7315 0.3196 0.7315 0.8552
No log 2.1882 186 0.7272 0.3415 0.7272 0.8528
No log 2.2118 188 0.7372 0.3081 0.7372 0.8586
No log 2.2353 190 0.7299 0.3402 0.7299 0.8544
No log 2.2588 192 0.7240 0.3337 0.7240 0.8509
No log 2.2824 194 0.7337 0.3166 0.7337 0.8565
No log 2.3059 196 0.7699 0.3406 0.7699 0.8775
No log 2.3294 198 0.7489 0.2972 0.7489 0.8654
No log 2.3529 200 0.7549 0.2972 0.7549 0.8688
No log 2.3765 202 0.7372 0.3224 0.7372 0.8586
No log 2.4 204 0.7486 0.2880 0.7486 0.8652
No log 2.4235 206 0.7344 0.3241 0.7344 0.8569
No log 2.4471 208 0.7455 0.2704 0.7455 0.8634
No log 2.4706 210 0.7735 0.2929 0.7735 0.8795
No log 2.4941 212 0.7480 0.2968 0.7480 0.8649
No log 2.5176 214 0.7556 0.2493 0.7556 0.8693
No log 2.5412 216 0.7053 0.2972 0.7053 0.8398
No log 2.5647 218 0.6909 0.3838 0.6909 0.8312
No log 2.5882 220 0.6881 0.3596 0.6881 0.8295
No log 2.6118 222 0.7212 0.3125 0.7212 0.8492
No log 2.6353 224 0.8279 0.3459 0.8279 0.9099
No log 2.6588 226 0.7961 0.3775 0.7961 0.8923
No log 2.6824 228 0.7015 0.4217 0.7015 0.8376
No log 2.7059 230 0.6955 0.3811 0.6955 0.8340
No log 2.7294 232 0.7002 0.4140 0.7002 0.8368
No log 2.7529 234 0.7614 0.4108 0.7614 0.8726
No log 2.7765 236 0.8003 0.3735 0.8003 0.8946
No log 2.8 238 0.9743 0.3479 0.9743 0.9871
No log 2.8235 240 1.0075 0.3278 1.0075 1.0038
No log 2.8471 242 0.9563 0.3466 0.9563 0.9779
No log 2.8706 244 0.8384 0.3728 0.8384 0.9156
No log 2.8941 246 0.7442 0.3702 0.7442 0.8627
No log 2.9176 248 0.7219 0.3495 0.7219 0.8496
No log 2.9412 250 0.7281 0.4028 0.7281 0.8533
No log 2.9647 252 0.7229 0.3999 0.7229 0.8502
No log 2.9882 254 0.7362 0.4049 0.7362 0.8580
No log 3.0118 256 0.7998 0.3923 0.7998 0.8943
No log 3.0353 258 0.8926 0.4031 0.8926 0.9448
No log 3.0588 260 0.7809 0.3678 0.7809 0.8837
No log 3.0824 262 0.6564 0.3557 0.6564 0.8102
No log 3.1059 264 0.6824 0.3706 0.6824 0.8261
No log 3.1294 266 0.6748 0.3462 0.6748 0.8215
No log 3.1529 268 0.6796 0.2034 0.6796 0.8243
No log 3.1765 270 0.7133 0.3059 0.7133 0.8445
No log 3.2 272 0.6994 0.2938 0.6994 0.8363
No log 3.2235 274 0.7115 0.3510 0.7115 0.8435
No log 3.2471 276 0.7278 0.3828 0.7278 0.8531
No log 3.2706 278 0.7956 0.3548 0.7956 0.8919
No log 3.2941 280 0.8102 0.3699 0.8102 0.9001
No log 3.3176 282 0.7268 0.3947 0.7268 0.8525
No log 3.3412 284 0.6948 0.3647 0.6948 0.8335
No log 3.3647 286 0.7423 0.3472 0.7423 0.8615
No log 3.3882 288 0.8223 0.3577 0.8223 0.9068
No log 3.4118 290 0.8209 0.3678 0.8209 0.9060
No log 3.4353 292 0.7224 0.3870 0.7224 0.8499
No log 3.4588 294 0.7442 0.3691 0.7442 0.8627
No log 3.4824 296 0.6850 0.3828 0.6850 0.8276
No log 3.5059 298 0.6704 0.3258 0.6704 0.8188
No log 3.5294 300 0.6868 0.3412 0.6868 0.8287
No log 3.5529 302 0.7002 0.3290 0.7002 0.8368
No log 3.5765 304 0.7095 0.2980 0.7095 0.8423
No log 3.6 306 0.6844 0.2415 0.6844 0.8273
No log 3.6235 308 0.6805 0.2844 0.6805 0.8249
No log 3.6471 310 0.7057 0.2419 0.7057 0.8400
No log 3.6706 312 0.7193 0.3801 0.7193 0.8481
No log 3.6941 314 0.7354 0.3394 0.7354 0.8576
No log 3.7176 316 0.7831 0.3413 0.7831 0.8850
No log 3.7412 318 0.7511 0.3555 0.7511 0.8667
No log 3.7647 320 0.7626 0.3045 0.7626 0.8733
No log 3.7882 322 0.7319 0.3208 0.7319 0.8555
No log 3.8118 324 0.7322 0.2998 0.7322 0.8557
No log 3.8353 326 0.7259 0.3376 0.7259 0.8520
No log 3.8588 328 0.7186 0.3344 0.7186 0.8477
No log 3.8824 330 0.7199 0.3687 0.7199 0.8485
No log 3.9059 332 0.7141 0.4327 0.7141 0.8450
No log 3.9294 334 0.7504 0.4222 0.7504 0.8663
No log 3.9529 336 0.8497 0.3950 0.8497 0.9218
No log 3.9765 338 0.7761 0.4267 0.7761 0.8809
No log 4.0 340 0.6425 0.4506 0.6425 0.8015
No log 4.0235 342 0.6247 0.3471 0.6247 0.7904
No log 4.0471 344 0.6409 0.4507 0.6409 0.8005
No log 4.0706 346 0.6434 0.4605 0.6434 0.8022
No log 4.0941 348 0.6344 0.4797 0.6344 0.7965
No log 4.1176 350 0.6556 0.4327 0.6556 0.8097
No log 4.1412 352 0.6759 0.4282 0.6759 0.8221
No log 4.1647 354 0.7585 0.4306 0.7585 0.8709
No log 4.1882 356 0.7859 0.4250 0.7859 0.8865
No log 4.2118 358 0.7152 0.3696 0.7152 0.8457
No log 4.2353 360 0.6880 0.3747 0.6880 0.8295
No log 4.2588 362 0.6872 0.3772 0.6872 0.8289
No log 4.2824 364 0.7331 0.3446 0.7331 0.8562
No log 4.3059 366 0.8897 0.3556 0.8897 0.9433
No log 4.3294 368 0.8723 0.3782 0.8723 0.9340
No log 4.3529 370 0.7325 0.3288 0.7325 0.8558
No log 4.3765 372 0.6929 0.3682 0.6929 0.8324
No log 4.4 374 0.6833 0.3511 0.6833 0.8266
No log 4.4235 376 0.6843 0.3329 0.6843 0.8272
No log 4.4471 378 0.7003 0.2905 0.7003 0.8368
No log 4.4706 380 0.6676 0.3169 0.6676 0.8171
No log 4.4941 382 0.6639 0.3145 0.6639 0.8148
No log 4.5176 384 0.6676 0.2884 0.6676 0.8171
No log 4.5412 386 0.6721 0.3536 0.6721 0.8198
No log 4.5647 388 0.6735 0.3143 0.6735 0.8207
No log 4.5882 390 0.6982 0.3582 0.6982 0.8356
No log 4.6118 392 0.7081 0.3548 0.7081 0.8415
No log 4.6353 394 0.6854 0.3738 0.6854 0.8279
No log 4.6588 396 0.6526 0.2315 0.6526 0.8079
No log 4.6824 398 0.6457 0.3584 0.6457 0.8036
No log 4.7059 400 0.6655 0.3781 0.6655 0.8158
No log 4.7294 402 0.7421 0.4013 0.7421 0.8615
No log 4.7529 404 0.8328 0.3837 0.8328 0.9126
No log 4.7765 406 0.7733 0.3866 0.7733 0.8794
No log 4.8 408 0.6377 0.4453 0.6377 0.7986
No log 4.8235 410 0.6741 0.4131 0.6741 0.8211
No log 4.8471 412 0.6877 0.4108 0.6877 0.8293
No log 4.8706 414 0.6186 0.4415 0.6186 0.7865
No log 4.8941 416 0.6638 0.4431 0.6638 0.8147
No log 4.9176 418 0.8004 0.3867 0.8004 0.8946
No log 4.9412 420 0.7816 0.3867 0.7816 0.8841
No log 4.9647 422 0.6581 0.4419 0.6581 0.8112
No log 4.9882 424 0.6499 0.4773 0.6499 0.8061
No log 5.0118 426 0.6516 0.4889 0.6516 0.8072
No log 5.0353 428 0.6105 0.5011 0.6105 0.7814
No log 5.0588 430 0.6791 0.3958 0.6791 0.8241
No log 5.0824 432 0.7388 0.4155 0.7388 0.8595
No log 5.1059 434 0.6850 0.4047 0.6850 0.8277
No log 5.1294 436 0.6272 0.5288 0.6272 0.7920
No log 5.1529 438 0.6795 0.4311 0.6795 0.8243
No log 5.1765 440 0.7225 0.4311 0.7225 0.8500
No log 5.2 442 0.6760 0.4384 0.6760 0.8222
No log 5.2235 444 0.6889 0.4255 0.6889 0.8300
No log 5.2471 446 0.7971 0.4096 0.7971 0.8928
No log 5.2706 448 0.8075 0.3778 0.8075 0.8986
No log 5.2941 450 0.7685 0.3957 0.7685 0.8766
No log 5.3176 452 0.6860 0.4112 0.6860 0.8283
No log 5.3412 454 0.6643 0.3144 0.6643 0.8150
No log 5.3647 456 0.6705 0.3626 0.6705 0.8189
No log 5.3882 458 0.6709 0.3587 0.6709 0.8191
No log 5.4118 460 0.6658 0.3626 0.6658 0.8160
No log 5.4353 462 0.6625 0.3777 0.6625 0.8140
No log 5.4588 464 0.6724 0.3794 0.6724 0.8200
No log 5.4824 466 0.6631 0.4038 0.6631 0.8143
No log 5.5059 468 0.7171 0.3748 0.7171 0.8468
No log 5.5294 470 0.7688 0.3989 0.7688 0.8768
No log 5.5529 472 0.7162 0.3497 0.7162 0.8463
No log 5.5765 474 0.6510 0.3941 0.6510 0.8068
No log 5.6 476 0.6505 0.3559 0.6505 0.8066
No log 5.6235 478 0.6417 0.4181 0.6417 0.8011
No log 5.6471 480 0.6674 0.3059 0.6674 0.8169
No log 5.6706 482 0.7341 0.3105 0.7341 0.8568
No log 5.6941 484 0.8032 0.3953 0.8032 0.8962
No log 5.7176 486 0.7633 0.3874 0.7633 0.8737
No log 5.7412 488 0.6817 0.3957 0.6817 0.8257
No log 5.7647 490 0.6677 0.3938 0.6677 0.8171
No log 5.7882 492 0.6708 0.3931 0.6708 0.8190
No log 5.8118 494 0.6803 0.3662 0.6803 0.8248
No log 5.8353 496 0.6761 0.3646 0.6761 0.8223
No log 5.8588 498 0.6694 0.3747 0.6694 0.8182
0.4205 5.8824 500 0.6683 0.3528 0.6683 0.8175
0.4205 5.9059 502 0.6784 0.3413 0.6784 0.8237
0.4205 5.9294 504 0.6716 0.3809 0.6716 0.8195
0.4205 5.9529 506 0.6696 0.3917 0.6696 0.8183
0.4205 5.9765 508 0.6890 0.3653 0.6890 0.8301
0.4205 6.0 510 0.7259 0.4064 0.7259 0.8520
0.4205 6.0235 512 0.7077 0.4222 0.7077 0.8412
0.4205 6.0471 514 0.7150 0.3425 0.7150 0.8456
0.4205 6.0706 516 0.7411 0.3216 0.7411 0.8609
0.4205 6.0941 518 0.7114 0.3070 0.7114 0.8435
0.4205 6.1176 520 0.7044 0.3146 0.7044 0.8393
0.4205 6.1412 522 0.7162 0.3504 0.7162 0.8463

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
6
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k16_task2_organization

Finetuned
(4222)
this model