ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k4_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0131
  • Qwk: 0.5317
  • Mse: 1.0131
  • Rmse: 1.0065

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0952 2 5.4156 0.0062 5.4156 2.3272
No log 0.1905 4 3.4705 0.0543 3.4705 1.8629
No log 0.2857 6 2.5840 -0.0487 2.5840 1.6075
No log 0.3810 8 2.6466 -0.0533 2.6466 1.6268
No log 0.4762 10 1.8239 0.0215 1.8239 1.3505
No log 0.5714 12 1.2849 0.1815 1.2849 1.1335
No log 0.6667 14 1.2158 0.2149 1.2158 1.1026
No log 0.7619 16 1.2399 0.1195 1.2399 1.1135
No log 0.8571 18 1.2603 0.0738 1.2603 1.1226
No log 0.9524 20 1.2262 0.0738 1.2262 1.1073
No log 1.0476 22 1.1880 0.2013 1.1880 1.0900
No log 1.1429 24 1.2489 0.1817 1.2489 1.1175
No log 1.2381 26 1.1984 0.2922 1.1984 1.0947
No log 1.3333 28 1.1303 0.2784 1.1303 1.0631
No log 1.4286 30 1.1032 0.2386 1.1032 1.0503
No log 1.5238 32 1.0407 0.2036 1.0407 1.0202
No log 1.6190 34 1.0442 0.3912 1.0442 1.0218
No log 1.7143 36 1.1998 0.3783 1.1998 1.0953
No log 1.8095 38 1.2129 0.4062 1.2129 1.1013
No log 1.9048 40 1.1396 0.3890 1.1396 1.0675
No log 2.0 42 1.1464 0.3450 1.1464 1.0707
No log 2.0952 44 1.0749 0.2849 1.0749 1.0368
No log 2.1905 46 1.0347 0.2489 1.0347 1.0172
No log 2.2857 48 1.0169 0.2599 1.0169 1.0084
No log 2.3810 50 0.9889 0.3510 0.9889 0.9944
No log 2.4762 52 0.9979 0.4542 0.9979 0.9989
No log 2.5714 54 1.0182 0.4132 1.0182 1.0090
No log 2.6667 56 1.0284 0.3719 1.0284 1.0141
No log 2.7619 58 1.0311 0.2887 1.0311 1.0154
No log 2.8571 60 1.0592 0.2722 1.0592 1.0292
No log 2.9524 62 1.0033 0.3441 1.0033 1.0016
No log 3.0476 64 0.9529 0.3850 0.9529 0.9762
No log 3.1429 66 0.8925 0.3777 0.8925 0.9447
No log 3.2381 68 0.8827 0.4376 0.8827 0.9395
No log 3.3333 70 0.8978 0.4868 0.8978 0.9475
No log 3.4286 72 0.9338 0.4112 0.9338 0.9664
No log 3.5238 74 0.9318 0.4586 0.9318 0.9653
No log 3.6190 76 0.9227 0.4946 0.9227 0.9606
No log 3.7143 78 1.0972 0.4321 1.0972 1.0475
No log 3.8095 80 1.2822 0.4505 1.2822 1.1323
No log 3.9048 82 1.2541 0.2949 1.2541 1.1199
No log 4.0 84 1.2814 0.2040 1.2814 1.1320
No log 4.0952 86 1.1855 0.2548 1.1855 1.0888
No log 4.1905 88 1.0208 0.3472 1.0208 1.0103
No log 4.2857 90 0.8944 0.5566 0.8944 0.9457
No log 4.3810 92 0.7971 0.5315 0.7971 0.8928
No log 4.4762 94 0.8409 0.5178 0.8409 0.9170
No log 4.5714 96 0.7933 0.5412 0.7933 0.8907
No log 4.6667 98 0.8816 0.5426 0.8816 0.9389
No log 4.7619 100 0.9285 0.5490 0.9285 0.9636
No log 4.8571 102 1.0469 0.5499 1.0469 1.0232
No log 4.9524 104 1.0859 0.5418 1.0859 1.0421
No log 5.0476 106 1.0068 0.5523 1.0068 1.0034
No log 5.1429 108 0.8837 0.5556 0.8837 0.9400
No log 5.2381 110 0.9049 0.5730 0.9049 0.9513
No log 5.3333 112 0.9055 0.5480 0.9055 0.9516
No log 5.4286 114 0.9189 0.5478 0.9189 0.9586
No log 5.5238 116 0.8876 0.5635 0.8876 0.9421
No log 5.6190 118 0.7994 0.6168 0.7994 0.8941
No log 5.7143 120 0.7708 0.6151 0.7708 0.8780
No log 5.8095 122 0.7741 0.6308 0.7741 0.8799
No log 5.9048 124 0.8957 0.5731 0.8957 0.9464
No log 6.0 126 1.1345 0.4834 1.1345 1.0651
No log 6.0952 128 1.4205 0.4244 1.4205 1.1918
No log 6.1905 130 1.3489 0.4679 1.3489 1.1614
No log 6.2857 132 1.0459 0.5548 1.0459 1.0227
No log 6.3810 134 0.8139 0.6386 0.8139 0.9022
No log 6.4762 136 0.6821 0.6398 0.6821 0.8259
No log 6.5714 138 0.6774 0.6331 0.6774 0.8230
No log 6.6667 140 0.7593 0.6494 0.7593 0.8714
No log 6.7619 142 1.0570 0.5696 1.0570 1.0281
No log 6.8571 144 1.2540 0.5083 1.2540 1.1198
No log 6.9524 146 1.3757 0.4658 1.3757 1.1729
No log 7.0476 148 1.3163 0.4807 1.3163 1.1473
No log 7.1429 150 1.3546 0.4480 1.3546 1.1639
No log 7.2381 152 1.1468 0.5418 1.1468 1.0709
No log 7.3333 154 0.9882 0.5350 0.9882 0.9941
No log 7.4286 156 0.8672 0.6007 0.8672 0.9313
No log 7.5238 158 0.7702 0.6360 0.7702 0.8776
No log 7.6190 160 0.7239 0.6428 0.7239 0.8508
No log 7.7143 162 0.7693 0.6386 0.7693 0.8771
No log 7.8095 164 0.8309 0.6177 0.8309 0.9115
No log 7.9048 166 0.7899 0.6285 0.7899 0.8888
No log 8.0 168 0.6976 0.6410 0.6976 0.8352
No log 8.0952 170 0.7080 0.6596 0.7080 0.8415
No log 8.1905 172 0.7778 0.6611 0.7778 0.8819
No log 8.2857 174 0.9323 0.6268 0.9323 0.9655
No log 8.3810 176 1.1300 0.5860 1.1300 1.0630
No log 8.4762 178 1.1954 0.5656 1.1954 1.0934
No log 8.5714 180 1.3058 0.5325 1.3058 1.1427
No log 8.6667 182 1.1942 0.5246 1.1942 1.0928
No log 8.7619 184 0.9140 0.5706 0.9140 0.9560
No log 8.8571 186 0.7600 0.6372 0.7600 0.8718
No log 8.9524 188 0.7576 0.6212 0.7576 0.8704
No log 9.0476 190 0.8098 0.6116 0.8098 0.8999
No log 9.1429 192 0.9652 0.5793 0.9652 0.9825
No log 9.2381 194 1.0886 0.5525 1.0886 1.0434
No log 9.3333 196 1.0575 0.5547 1.0575 1.0283
No log 9.4286 198 0.9027 0.5905 0.9027 0.9501
No log 9.5238 200 0.8538 0.6140 0.8538 0.9240
No log 9.6190 202 0.8955 0.5731 0.8955 0.9463
No log 9.7143 204 0.9863 0.5558 0.9863 0.9931
No log 9.8095 206 1.1375 0.4932 1.1375 1.0665
No log 9.9048 208 1.0649 0.5477 1.0649 1.0319
No log 10.0 210 0.8998 0.5693 0.8998 0.9486
No log 10.0952 212 0.7184 0.6742 0.7184 0.8476
No log 10.1905 214 0.6922 0.6585 0.6922 0.8320
No log 10.2857 216 0.7683 0.6499 0.7683 0.8765
No log 10.3810 218 0.9847 0.5707 0.9847 0.9923
No log 10.4762 220 1.1006 0.5846 1.1006 1.0491
No log 10.5714 222 0.9487 0.5988 0.9487 0.9740
No log 10.6667 224 0.6850 0.7127 0.6850 0.8276
No log 10.7619 226 0.6421 0.6560 0.6421 0.8013
No log 10.8571 228 0.6504 0.6630 0.6504 0.8065
No log 10.9524 230 0.6925 0.6898 0.6925 0.8322
No log 11.0476 232 0.9208 0.6189 0.9208 0.9596
No log 11.1429 234 1.2887 0.5229 1.2887 1.1352
No log 11.2381 236 1.2996 0.4656 1.2996 1.1400
No log 11.3333 238 1.0477 0.5421 1.0477 1.0236
No log 11.4286 240 0.8448 0.6165 0.8448 0.9191
No log 11.5238 242 0.7345 0.6462 0.7345 0.8570
No log 11.6190 244 0.7724 0.6354 0.7724 0.8789
No log 11.7143 246 0.8573 0.6326 0.8573 0.9259
No log 11.8095 248 1.0036 0.5955 1.0036 1.0018
No log 11.9048 250 1.1904 0.5487 1.1904 1.0910
No log 12.0 252 1.3632 0.4736 1.3632 1.1676
No log 12.0952 254 1.2972 0.4662 1.2972 1.1390
No log 12.1905 256 1.0559 0.5348 1.0559 1.0276
No log 12.2857 258 0.9589 0.5508 0.9589 0.9792
No log 12.3810 260 0.9197 0.5793 0.9197 0.9590
No log 12.4762 262 0.8737 0.6149 0.8737 0.9347
No log 12.5714 264 0.7199 0.6617 0.7199 0.8485
No log 12.6667 266 0.6770 0.6853 0.6770 0.8228
No log 12.7619 268 0.7141 0.6180 0.7141 0.8450
No log 12.8571 270 0.7530 0.5992 0.7530 0.8677
No log 12.9524 272 0.7466 0.6192 0.7466 0.8641
No log 13.0476 274 0.7165 0.6503 0.7165 0.8465
No log 13.1429 276 0.7622 0.6232 0.7622 0.8730
No log 13.2381 278 0.8062 0.6351 0.8062 0.8979
No log 13.3333 280 0.9338 0.5893 0.9338 0.9663
No log 13.4286 282 0.9869 0.5377 0.9869 0.9934
No log 13.5238 284 1.0329 0.5517 1.0329 1.0163
No log 13.6190 286 0.9798 0.5656 0.9798 0.9899
No log 13.7143 288 0.9594 0.5796 0.9594 0.9795
No log 13.8095 290 0.9142 0.5845 0.9142 0.9561
No log 13.9048 292 0.8940 0.5845 0.8940 0.9455
No log 14.0 294 0.8582 0.5937 0.8582 0.9264
No log 14.0952 296 0.8486 0.5898 0.8486 0.9212
No log 14.1905 298 0.9437 0.5401 0.9437 0.9714
No log 14.2857 300 0.9717 0.5401 0.9717 0.9858
No log 14.3810 302 0.8276 0.6139 0.8276 0.9097
No log 14.4762 304 0.8246 0.6284 0.8246 0.9081
No log 14.5714 306 0.9310 0.5863 0.9310 0.9649
No log 14.6667 308 0.9891 0.5529 0.9891 0.9945
No log 14.7619 310 1.1257 0.5459 1.1257 1.0610
No log 14.8571 312 1.0503 0.5524 1.0503 1.0248
No log 14.9524 314 1.0397 0.5435 1.0397 1.0196
No log 15.0476 316 0.8966 0.6079 0.8966 0.9469
No log 15.1429 318 0.8488 0.6170 0.8488 0.9213
No log 15.2381 320 0.7513 0.6341 0.7513 0.8668
No log 15.3333 322 0.7495 0.6358 0.7495 0.8657
No log 15.4286 324 0.7475 0.6412 0.7475 0.8646
No log 15.5238 326 0.7984 0.6191 0.7984 0.8935
No log 15.6190 328 0.8208 0.6003 0.8208 0.9060
No log 15.7143 330 0.8106 0.6049 0.8106 0.9003
No log 15.8095 332 0.7814 0.6240 0.7814 0.8840
No log 15.9048 334 0.7919 0.6189 0.7919 0.8899
No log 16.0 336 0.8389 0.5949 0.8389 0.9159
No log 16.0952 338 0.9095 0.5882 0.9095 0.9537
No log 16.1905 340 0.9531 0.5976 0.9531 0.9763
No log 16.2857 342 0.9056 0.5931 0.9056 0.9516
No log 16.3810 344 0.9386 0.5882 0.9386 0.9688
No log 16.4762 346 0.9174 0.5882 0.9174 0.9578
No log 16.5714 348 0.8836 0.5974 0.8836 0.9400
No log 16.6667 350 0.8578 0.6070 0.8578 0.9262
No log 16.7619 352 0.8419 0.6241 0.8419 0.9175
No log 16.8571 354 0.8125 0.6373 0.8125 0.9014
No log 16.9524 356 0.9219 0.6264 0.9219 0.9602
No log 17.0476 358 1.0580 0.5920 1.0580 1.0286
No log 17.1429 360 0.9773 0.5706 0.9773 0.9886
No log 17.2381 362 0.8399 0.6116 0.8399 0.9165
No log 17.3333 364 0.8011 0.6132 0.8011 0.8950
No log 17.4286 366 0.8110 0.6241 0.8110 0.9005
No log 17.5238 368 0.8572 0.6312 0.8572 0.9258
No log 17.6190 370 1.0283 0.5675 1.0283 1.0140
No log 17.7143 372 1.0895 0.5511 1.0895 1.0438
No log 17.8095 374 1.0299 0.5647 1.0299 1.0148
No log 17.9048 376 0.8537 0.5974 0.8537 0.9239
No log 18.0 378 0.7630 0.6212 0.7630 0.8735
No log 18.0952 380 0.7840 0.6329 0.7840 0.8854
No log 18.1905 382 0.8942 0.5745 0.8942 0.9456
No log 18.2857 384 1.0183 0.5909 1.0183 1.0091
No log 18.3810 386 0.9066 0.6207 0.9066 0.9522
No log 18.4762 388 0.7225 0.6584 0.7225 0.8500
No log 18.5714 390 0.6409 0.6664 0.6409 0.8006
No log 18.6667 392 0.6604 0.6708 0.6604 0.8126
No log 18.7619 394 0.7811 0.6450 0.7811 0.8838
No log 18.8571 396 0.8670 0.5886 0.8670 0.9311
No log 18.9524 398 0.8451 0.6126 0.8451 0.9193
No log 19.0476 400 0.8282 0.6069 0.8282 0.9101
No log 19.1429 402 0.8497 0.6000 0.8497 0.9218
No log 19.2381 404 0.8401 0.6089 0.8401 0.9166
No log 19.3333 406 0.7743 0.6135 0.7743 0.8799
No log 19.4286 408 0.7051 0.6443 0.7051 0.8397
No log 19.5238 410 0.7295 0.6365 0.7295 0.8541
No log 19.6190 412 0.8341 0.6177 0.8341 0.9133
No log 19.7143 414 0.9043 0.5793 0.9043 0.9509
No log 19.8095 416 0.8519 0.5975 0.8519 0.9230
No log 19.9048 418 0.8372 0.6142 0.8372 0.9150
No log 20.0 420 0.9090 0.5553 0.9090 0.9534
No log 20.0952 422 0.8650 0.6105 0.8650 0.9301
No log 20.1905 424 0.7713 0.6466 0.7713 0.8783
No log 20.2857 426 0.7529 0.6495 0.7529 0.8677
No log 20.3810 428 0.8065 0.6356 0.8065 0.8981
No log 20.4762 430 0.9242 0.5719 0.9242 0.9613
No log 20.5714 432 0.9318 0.5393 0.9318 0.9653
No log 20.6667 434 0.9005 0.5718 0.9005 0.9490
No log 20.7619 436 0.8400 0.5852 0.8400 0.9165
No log 20.8571 438 0.8427 0.5852 0.8427 0.9180
No log 20.9524 440 0.9197 0.5353 0.9197 0.9590
No log 21.0476 442 1.1111 0.5446 1.1111 1.0541
No log 21.1429 444 1.2474 0.5248 1.2474 1.1169
No log 21.2381 446 1.1956 0.5344 1.1956 1.0934
No log 21.3333 448 1.1067 0.5094 1.1067 1.0520
No log 21.4286 450 1.0762 0.5301 1.0762 1.0374
No log 21.5238 452 1.0462 0.5401 1.0462 1.0229
No log 21.6190 454 0.9994 0.5517 0.9994 0.9997
No log 21.7143 456 1.0202 0.5732 1.0202 1.0100
No log 21.8095 458 0.9642 0.5874 0.9642 0.9819
No log 21.9048 460 0.8628 0.6247 0.8628 0.9289
No log 22.0 462 0.8474 0.6196 0.8474 0.9205
No log 22.0952 464 0.7983 0.6189 0.7983 0.8935
No log 22.1905 466 0.7430 0.6571 0.7430 0.8620
No log 22.2857 468 0.7840 0.6189 0.7840 0.8855
No log 22.3810 470 0.8239 0.6516 0.8239 0.9077
No log 22.4762 472 0.8475 0.6410 0.8475 0.9206
No log 22.5714 474 0.7828 0.6487 0.7828 0.8847
No log 22.6667 476 0.7779 0.6388 0.7779 0.8820
No log 22.7619 478 0.8282 0.6221 0.8282 0.9100
No log 22.8571 480 0.9723 0.5954 0.9723 0.9861
No log 22.9524 482 0.9838 0.6136 0.9838 0.9919
No log 23.0476 484 0.8748 0.5896 0.8748 0.9353
No log 23.1429 486 0.7563 0.6281 0.7563 0.8696
No log 23.2381 488 0.7319 0.6108 0.7319 0.8555
No log 23.3333 490 0.7518 0.6337 0.7518 0.8671
No log 23.4286 492 0.7780 0.6081 0.7780 0.8821
No log 23.5238 494 0.7868 0.6065 0.7868 0.8870
No log 23.6190 496 0.7727 0.6216 0.7727 0.8790
No log 23.7143 498 0.8485 0.5985 0.8485 0.9212
0.4191 23.8095 500 0.9267 0.5905 0.9267 0.9627
0.4191 23.9048 502 1.0122 0.5535 1.0122 1.0061
0.4191 24.0 504 0.9286 0.5804 0.9286 0.9636
0.4191 24.0952 506 0.8056 0.6444 0.8056 0.8976
0.4191 24.1905 508 0.6810 0.7075 0.6810 0.8253
0.4191 24.2857 510 0.6541 0.6949 0.6541 0.8088
0.4191 24.3810 512 0.6939 0.7021 0.6939 0.8330
0.4191 24.4762 514 0.7142 0.7000 0.7142 0.8451
0.4191 24.5714 516 0.7117 0.6964 0.7117 0.8436
0.4191 24.6667 518 0.6763 0.6831 0.6763 0.8224
0.4191 24.7619 520 0.7344 0.6706 0.7344 0.8570
0.4191 24.8571 522 0.7446 0.6706 0.7446 0.8629
0.4191 24.9524 524 0.8346 0.6568 0.8346 0.9136
0.4191 25.0476 526 0.9372 0.6196 0.9372 0.9681
0.4191 25.1429 528 0.9127 0.6156 0.9127 0.9553
0.4191 25.2381 530 0.7935 0.6332 0.7935 0.8908
0.4191 25.3333 532 0.7478 0.6367 0.7478 0.8648
0.4191 25.4286 534 0.7816 0.6097 0.7816 0.8841
0.4191 25.5238 536 0.9362 0.5745 0.9362 0.9676
0.4191 25.6190 538 1.0181 0.5578 1.0181 1.0090
0.4191 25.7143 540 1.0293 0.5418 1.0293 1.0145
0.4191 25.8095 542 1.0463 0.5561 1.0463 1.0229
0.4191 25.9048 544 1.0131 0.5317 1.0131 1.0065

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
7
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k4_task1_organization

Finetuned
(4222)
this model