ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k3_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6083
  • Qwk: 0.7613
  • Mse: 0.6083
  • Rmse: 0.7800

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0870 2 6.7965 0.0242 6.7965 2.6070
No log 0.1739 4 5.2676 0.0366 5.2676 2.2951
No log 0.2609 6 3.0917 0.0595 3.0917 1.7583
No log 0.3478 8 2.5803 0.0132 2.5803 1.6063
No log 0.4348 10 3.5005 0.0787 3.5005 1.8710
No log 0.5217 12 2.6461 0.0141 2.6461 1.6267
No log 0.6087 14 1.7385 0.2478 1.7385 1.3185
No log 0.6957 16 1.6891 0.0583 1.6891 1.2996
No log 0.7826 18 1.6908 0.1296 1.6908 1.3003
No log 0.8696 20 1.6275 0.1308 1.6275 1.2757
No log 0.9565 22 1.4872 0.1538 1.4872 1.2195
No log 1.0435 24 1.5368 0.3158 1.5368 1.2397
No log 1.1304 26 1.3818 0.3158 1.3818 1.1755
No log 1.2174 28 1.2881 0.3932 1.2881 1.1349
No log 1.3043 30 1.1907 0.3621 1.1907 1.0912
No log 1.3913 32 1.1694 0.3652 1.1694 1.0814
No log 1.4783 34 1.1661 0.3214 1.1661 1.0798
No log 1.5652 36 1.0867 0.4138 1.0867 1.0425
No log 1.6522 38 1.0225 0.4959 1.0225 1.0112
No log 1.7391 40 0.9464 0.6970 0.9464 0.9728
No log 1.8261 42 0.8481 0.7482 0.8481 0.9209
No log 1.9130 44 0.7664 0.7260 0.7664 0.8754
No log 2.0 46 0.7207 0.7651 0.7207 0.8489
No log 2.0870 48 0.9562 0.5926 0.9562 0.9779
No log 2.1739 50 1.2811 0.5224 1.2811 1.1319
No log 2.2609 52 1.1457 0.5481 1.1457 1.0704
No log 2.3478 54 1.0695 0.5630 1.0695 1.0342
No log 2.4348 56 1.2533 0.5180 1.2533 1.1195
No log 2.5217 58 1.4337 0.4714 1.4337 1.1974
No log 2.6087 60 1.1399 0.5693 1.1399 1.0677
No log 2.6957 62 0.8858 0.5926 0.8858 0.9412
No log 2.7826 64 0.9129 0.6567 0.9129 0.9555
No log 2.8696 66 1.2352 0.5 1.2352 1.1114
No log 2.9565 68 1.1034 0.5606 1.1034 1.0504
No log 3.0435 70 0.9522 0.5891 0.9522 0.9758
No log 3.1304 72 0.9261 0.6142 0.9261 0.9624
No log 3.2174 74 0.9060 0.6515 0.9060 0.9519
No log 3.3043 76 0.8631 0.6667 0.8631 0.9290
No log 3.3913 78 0.8798 0.6522 0.8798 0.9380
No log 3.4783 80 0.8796 0.6324 0.8796 0.9379
No log 3.5652 82 0.9746 0.6222 0.9746 0.9872
No log 3.6522 84 1.0278 0.5899 1.0278 1.0138
No log 3.7391 86 0.9362 0.6479 0.9362 0.9676
No log 3.8261 88 0.9420 0.6383 0.9420 0.9706
No log 3.9130 90 1.0282 0.6383 1.0282 1.0140
No log 4.0 92 1.3175 0.5 1.3175 1.1478
No log 4.0870 94 1.2619 0.5324 1.2619 1.1234
No log 4.1739 96 0.9363 0.6716 0.9363 0.9676
No log 4.2609 98 0.7244 0.7194 0.7244 0.8511
No log 4.3478 100 0.7534 0.7586 0.7534 0.8680
No log 4.4348 102 0.7001 0.7286 0.7001 0.8367
No log 4.5217 104 0.7011 0.7324 0.7011 0.8373
No log 4.6087 106 0.7827 0.6897 0.7827 0.8847
No log 4.6957 108 0.7922 0.7034 0.7922 0.8900
No log 4.7826 110 0.8235 0.6377 0.8235 0.9075
No log 4.8696 112 0.8034 0.6331 0.8034 0.8963
No log 4.9565 114 0.7506 0.6897 0.7506 0.8664
No log 5.0435 116 0.8669 0.6525 0.8669 0.9311
No log 5.1304 118 0.9777 0.6241 0.9777 0.9888
No log 5.2174 120 0.8365 0.6525 0.8365 0.9146
No log 5.3043 122 0.7675 0.7034 0.7675 0.8761
No log 5.3913 124 0.8405 0.7222 0.8405 0.9168
No log 5.4783 126 0.8176 0.7020 0.8176 0.9042
No log 5.5652 128 0.7173 0.7027 0.7173 0.8469
No log 5.6522 130 0.7593 0.6897 0.7593 0.8714
No log 5.7391 132 0.7562 0.6806 0.7562 0.8696
No log 5.8261 134 0.7676 0.6475 0.7676 0.8761
No log 5.9130 136 0.7785 0.6812 0.7785 0.8823
No log 6.0 138 0.8224 0.6667 0.8224 0.9069
No log 6.0870 140 0.7774 0.6963 0.7774 0.8817
No log 6.1739 142 0.7749 0.6906 0.7749 0.8803
No log 6.2609 144 0.7668 0.6906 0.7668 0.8757
No log 6.3478 146 0.7373 0.7273 0.7373 0.8587
No log 6.4348 148 0.7192 0.7310 0.7192 0.8481
No log 6.5217 150 0.7090 0.7333 0.7090 0.8420
No log 6.6087 152 0.7434 0.7248 0.7434 0.8622
No log 6.6957 154 0.9131 0.6957 0.9131 0.9555
No log 6.7826 156 1.0124 0.6497 1.0124 1.0062
No log 6.8696 158 1.0566 0.6364 1.0566 1.0279
No log 6.9565 160 0.9694 0.6316 0.9694 0.9846
No log 7.0435 162 0.8067 0.7042 0.8067 0.8982
No log 7.1304 164 0.7349 0.6944 0.7349 0.8573
No log 7.2174 166 0.6986 0.7310 0.6986 0.8358
No log 7.3043 168 0.6657 0.7582 0.6657 0.8159
No log 7.3913 170 0.6363 0.7516 0.6363 0.7977
No log 7.4783 172 0.6082 0.7625 0.6082 0.7799
No log 7.5652 174 0.7006 0.7248 0.7006 0.8370
No log 7.6522 176 0.7965 0.6906 0.7965 0.8925
No log 7.7391 178 0.8408 0.6715 0.8408 0.9169
No log 7.8261 180 0.8113 0.6815 0.8113 0.9007
No log 7.9130 182 0.7450 0.7101 0.7450 0.8631
No log 8.0 184 0.7569 0.6765 0.7569 0.8700
No log 8.0870 186 0.8218 0.6364 0.8218 0.9065
No log 8.1739 188 0.8568 0.7353 0.8568 0.9256
No log 8.2609 190 0.8843 0.7007 0.8843 0.9404
No log 8.3478 192 0.8448 0.7092 0.8448 0.9191
No log 8.4348 194 0.8084 0.7133 0.8084 0.8991
No log 8.5217 196 0.8005 0.6901 0.8005 0.8947
No log 8.6087 198 0.8825 0.6099 0.8825 0.9394
No log 8.6957 200 0.9992 0.5899 0.9992 0.9996
No log 8.7826 202 0.9517 0.6471 0.9517 0.9756
No log 8.8696 204 0.7884 0.6619 0.7884 0.8879
No log 8.9565 206 0.6833 0.7483 0.6833 0.8266
No log 9.0435 208 0.6564 0.7534 0.6564 0.8102
No log 9.1304 210 0.6524 0.7260 0.6524 0.8077
No log 9.2174 212 0.7133 0.7114 0.7133 0.8446
No log 9.3043 214 0.7107 0.7097 0.7107 0.8430
No log 9.3913 216 0.6151 0.7248 0.6151 0.7843
No log 9.4783 218 0.5884 0.7632 0.5884 0.7671
No log 9.5652 220 0.6203 0.7453 0.6203 0.7876
No log 9.6522 222 0.6408 0.7586 0.6408 0.8005
No log 9.7391 224 0.5502 0.8136 0.5502 0.7417
No log 9.8261 226 0.5383 0.8095 0.5383 0.7337
No log 9.9130 228 0.5737 0.7738 0.5737 0.7575
No log 10.0 230 0.7389 0.7317 0.7389 0.8596
No log 10.0870 232 0.7915 0.6933 0.7915 0.8897
No log 10.1739 234 0.7053 0.7042 0.7053 0.8398
No log 10.2609 236 0.6509 0.7639 0.6509 0.8068
No log 10.3478 238 0.6944 0.7448 0.6944 0.8333
No log 10.4348 240 0.6575 0.7867 0.6575 0.8109
No log 10.5217 242 0.6312 0.8077 0.6312 0.7945
No log 10.6087 244 0.5962 0.7413 0.5962 0.7721
No log 10.6957 246 0.7272 0.7117 0.7272 0.8528
No log 10.7826 248 0.9181 0.7159 0.9181 0.9582
No log 10.8696 250 0.9041 0.7273 0.9041 0.9509
No log 10.9565 252 0.7246 0.7261 0.7246 0.8512
No log 11.0435 254 0.6210 0.7123 0.6210 0.7880
No log 11.1304 256 0.6441 0.7975 0.6441 0.8026
No log 11.2174 258 0.6400 0.7722 0.6400 0.8000
No log 11.3043 260 0.6608 0.7133 0.6608 0.8129
No log 11.3913 262 0.6943 0.7050 0.6943 0.8333
No log 11.4783 264 0.7331 0.7194 0.7331 0.8562
No log 11.5652 266 0.8334 0.6619 0.8334 0.9129
No log 11.6522 268 0.8575 0.6277 0.8575 0.9260
No log 11.7391 270 0.7644 0.6765 0.7644 0.8743
No log 11.8261 272 0.6426 0.7552 0.6426 0.8016
No log 11.9130 274 0.6004 0.7792 0.6004 0.7749
No log 12.0 276 0.5728 0.7871 0.5728 0.7569
No log 12.0870 278 0.5652 0.7831 0.5652 0.7518
No log 12.1739 280 0.6229 0.7727 0.6229 0.7893
No log 12.2609 282 0.6352 0.7574 0.6352 0.7970
No log 12.3478 284 0.6230 0.7712 0.6230 0.7893
No log 12.4348 286 0.6832 0.7143 0.6832 0.8266
No log 12.5217 288 0.7361 0.7429 0.7361 0.8580
No log 12.6087 290 0.7906 0.7092 0.7906 0.8892
No log 12.6957 292 0.7854 0.6901 0.7854 0.8863
No log 12.7826 294 0.7310 0.7172 0.7310 0.8550
No log 12.8696 296 0.6754 0.7432 0.6754 0.8219
No log 12.9565 298 0.6937 0.7347 0.6937 0.8329
No log 13.0435 300 0.7777 0.7172 0.7777 0.8819
No log 13.1304 302 0.8698 0.6806 0.8698 0.9326
No log 13.2174 304 0.8839 0.6806 0.8839 0.9402
No log 13.3043 306 0.8234 0.6901 0.8234 0.9074
No log 13.3913 308 0.7608 0.6950 0.7608 0.8722
No log 13.4783 310 0.6967 0.7222 0.6967 0.8347
No log 13.5652 312 0.6828 0.7083 0.6828 0.8263
No log 13.6522 314 0.7175 0.7342 0.7175 0.8470
No log 13.7391 316 0.7009 0.7545 0.7009 0.8372
No log 13.8261 318 0.5856 0.7619 0.5856 0.7653
No log 13.9130 320 0.5438 0.7516 0.5438 0.7374
No log 14.0 322 0.5434 0.8075 0.5434 0.7371
No log 14.0870 324 0.5506 0.7949 0.5506 0.7420
No log 14.1739 326 0.5678 0.7722 0.5678 0.7535
No log 14.2609 328 0.6371 0.6986 0.6371 0.7982
No log 14.3478 330 0.7045 0.7034 0.7045 0.8394
No log 14.4348 332 0.6817 0.6901 0.6817 0.8256
No log 14.5217 334 0.6294 0.7222 0.6294 0.7933
No log 14.6087 336 0.6336 0.7361 0.6336 0.7960
No log 14.6957 338 0.6504 0.7222 0.6504 0.8065
No log 14.7826 340 0.7387 0.7034 0.7387 0.8595
No log 14.8696 342 0.7726 0.7485 0.7726 0.8790
No log 14.9565 344 0.6830 0.7389 0.6830 0.8264
No log 15.0435 346 0.5874 0.7582 0.5874 0.7664
No log 15.1304 348 0.5782 0.7895 0.5782 0.7604
No log 15.2174 350 0.5740 0.7895 0.5740 0.7576
No log 15.3043 352 0.5592 0.7821 0.5592 0.7478
No log 15.3913 354 0.5783 0.7643 0.5783 0.7605
No log 15.4783 356 0.6894 0.7152 0.6894 0.8303
No log 15.5652 358 0.8083 0.7152 0.8083 0.8991
No log 15.6522 360 0.7990 0.7034 0.7990 0.8938
No log 15.7391 362 0.7269 0.7391 0.7269 0.8526
No log 15.8261 364 0.6937 0.7286 0.6937 0.8329
No log 15.9130 366 0.6747 0.7413 0.6747 0.8214
No log 16.0 368 0.7028 0.7391 0.7028 0.8384
No log 16.0870 370 0.7329 0.7042 0.7329 0.8561
No log 16.1739 372 0.7263 0.7042 0.7263 0.8522
No log 16.2609 374 0.6680 0.7234 0.6680 0.8173
No log 16.3478 376 0.6554 0.7324 0.6554 0.8096
No log 16.4348 378 0.6463 0.7092 0.6463 0.8039
No log 16.5217 380 0.6519 0.6993 0.6519 0.8074
No log 16.6087 382 0.7800 0.7355 0.7800 0.8832
No log 16.6957 384 0.8796 0.7135 0.8796 0.9379
No log 16.7826 386 0.8313 0.7412 0.8313 0.9117
No log 16.8696 388 0.7209 0.7403 0.7209 0.8491
No log 16.9565 390 0.6469 0.6993 0.6469 0.8043
No log 17.0435 392 0.6308 0.7429 0.6308 0.7942
No log 17.1304 394 0.6557 0.7391 0.6557 0.8098
No log 17.2174 396 0.6678 0.7206 0.6678 0.8172
No log 17.3043 398 0.6687 0.7299 0.6687 0.8177
No log 17.3913 400 0.6655 0.7299 0.6655 0.8158
No log 17.4783 402 0.6923 0.7286 0.6923 0.8321
No log 17.5652 404 0.6951 0.7092 0.6951 0.8337
No log 17.6522 406 0.6981 0.7299 0.6981 0.8355
No log 17.7391 408 0.6696 0.7259 0.6696 0.8183
No log 17.8261 410 0.6481 0.7338 0.6481 0.8051
No log 17.9130 412 0.6289 0.7465 0.6289 0.7930
No log 18.0 414 0.6130 0.7448 0.6130 0.7829
No log 18.0870 416 0.6391 0.7172 0.6391 0.7995
No log 18.1739 418 0.6525 0.7285 0.6525 0.8078
No log 18.2609 420 0.6984 0.6993 0.6984 0.8357
No log 18.3478 422 0.7015 0.6993 0.7015 0.8376
No log 18.4348 424 0.6479 0.7083 0.6479 0.8049
No log 18.5217 426 0.6102 0.7568 0.6102 0.7812
No log 18.6087 428 0.6221 0.7619 0.6221 0.7887
No log 18.6957 430 0.6196 0.7682 0.6196 0.7871
No log 18.7826 432 0.6425 0.7465 0.6425 0.8016
No log 18.8696 434 0.7384 0.7564 0.7384 0.8593
No log 18.9565 436 0.7683 0.7468 0.7683 0.8765
No log 19.0435 438 0.7516 0.7355 0.7516 0.8670
No log 19.1304 440 0.6607 0.76 0.6607 0.8129
No log 19.2174 442 0.6014 0.7552 0.6014 0.7755
No log 19.3043 444 0.5984 0.7552 0.5984 0.7736
No log 19.3913 446 0.5899 0.7273 0.5899 0.7681
No log 19.4783 448 0.6044 0.7632 0.6044 0.7774
No log 19.5652 450 0.6503 0.7333 0.6503 0.8064
No log 19.6522 452 0.6377 0.7333 0.6377 0.7985
No log 19.7391 454 0.6340 0.7451 0.6340 0.7962
No log 19.8261 456 0.6477 0.7451 0.6477 0.8048
No log 19.9130 458 0.6841 0.7448 0.6841 0.8271
No log 20.0 460 0.6696 0.7376 0.6696 0.8183
No log 20.0870 462 0.6422 0.7586 0.6422 0.8014
No log 20.1739 464 0.6321 0.7483 0.6321 0.7951
No log 20.2609 466 0.6653 0.7586 0.6653 0.8157
No log 20.3478 468 0.7282 0.7162 0.7282 0.8533
No log 20.4348 470 0.8254 0.6939 0.8254 0.9085
No log 20.5217 472 0.8982 0.675 0.8982 0.9478
No log 20.6087 474 0.9288 0.6667 0.9288 0.9638
No log 20.6957 476 0.8928 0.6757 0.8928 0.9449
No log 20.7826 478 0.7766 0.6993 0.7766 0.8812
No log 20.8696 480 0.7318 0.6861 0.7318 0.8554
No log 20.9565 482 0.7215 0.7299 0.7215 0.8494
No log 21.0435 484 0.7576 0.7050 0.7576 0.8704
No log 21.1304 486 0.8378 0.6901 0.8378 0.9153
No log 21.2174 488 0.8329 0.6713 0.8329 0.9126
No log 21.3043 490 0.7573 0.6806 0.7573 0.8702
No log 21.3913 492 0.6638 0.7534 0.6638 0.8147
No log 21.4783 494 0.6208 0.7712 0.6208 0.7879
No log 21.5652 496 0.5694 0.7771 0.5694 0.7546
No log 21.6522 498 0.5471 0.7643 0.5471 0.7397
0.3686 21.7391 500 0.5431 0.7643 0.5431 0.7369
0.3686 21.8261 502 0.5557 0.7875 0.5557 0.7455
0.3686 21.9130 504 0.6282 0.7907 0.6282 0.7926
0.3686 22.0 506 0.6934 0.7746 0.6934 0.8327
0.3686 22.0870 508 0.7206 0.7283 0.7206 0.8489
0.3686 22.1739 510 0.6874 0.775 0.6874 0.8291
0.3686 22.2609 512 0.6474 0.7517 0.6474 0.8046
0.3686 22.3478 514 0.6205 0.7660 0.6205 0.7877
0.3686 22.4348 516 0.6134 0.7660 0.6134 0.7832
0.3686 22.5217 518 0.6145 0.7606 0.6145 0.7839
0.3686 22.6087 520 0.6451 0.7552 0.6451 0.8032
0.3686 22.6957 522 0.7010 0.7564 0.7010 0.8372
0.3686 22.7826 524 0.6915 0.7673 0.6915 0.8316
0.3686 22.8696 526 0.6316 0.775 0.6316 0.7947
0.3686 22.9565 528 0.5589 0.7925 0.5589 0.7476
0.3686 23.0435 530 0.5466 0.8101 0.5466 0.7393
0.3686 23.1304 532 0.5849 0.7922 0.5849 0.7648
0.3686 23.2174 534 0.6205 0.7397 0.6205 0.7877
0.3686 23.3043 536 0.6262 0.7586 0.6262 0.7913
0.3686 23.3913 538 0.6058 0.7778 0.6058 0.7783
0.3686 23.4783 540 0.5877 0.7862 0.5877 0.7666
0.3686 23.5652 542 0.5880 0.7838 0.5880 0.7668
0.3686 23.6522 544 0.6254 0.7534 0.6254 0.7908
0.3686 23.7391 546 0.6427 0.7417 0.6427 0.8017
0.3686 23.8261 548 0.6254 0.75 0.6254 0.7909
0.3686 23.9130 550 0.6155 0.7550 0.6155 0.7845
0.3686 24.0 552 0.6083 0.7613 0.6083 0.7800

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
6
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k3_task1_organization

Finetuned
(4222)
this model