ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k1_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0580
  • Qwk: 0.6131
  • Mse: 1.0580
  • Rmse: 1.0286

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.25 2 6.0434 0.0261 6.0434 2.4583
No log 0.5 4 3.9893 0.1429 3.9893 1.9973
No log 0.75 6 2.6914 0.0500 2.6914 1.6406
No log 1.0 8 1.9588 0.0984 1.9588 1.3996
No log 1.25 10 1.7472 0.1111 1.7472 1.3218
No log 1.5 12 1.6663 0.0917 1.6663 1.2908
No log 1.75 14 1.6471 0.1416 1.6471 1.2834
No log 2.0 16 1.5418 0.1333 1.5418 1.2417
No log 2.25 18 1.4940 0.1714 1.4940 1.2223
No log 2.5 20 1.4679 0.1346 1.4679 1.2116
No log 2.75 22 1.3960 0.1905 1.3960 1.1815
No log 3.0 24 1.3620 0.2963 1.3620 1.1671
No log 3.25 26 1.3401 0.4655 1.3401 1.1576
No log 3.5 28 1.1229 0.5854 1.1229 1.0597
No log 3.75 30 1.1577 0.56 1.1577 1.0760
No log 4.0 32 1.3164 0.4762 1.3164 1.1473
No log 4.25 34 1.0964 0.5781 1.0964 1.0471
No log 4.5 36 0.9775 0.6316 0.9775 0.9887
No log 4.75 38 1.1159 0.5846 1.1159 1.0564
No log 5.0 40 1.1260 0.5581 1.1260 1.0611
No log 5.25 42 1.1567 0.5271 1.1567 1.0755
No log 5.5 44 1.3267 0.4580 1.3267 1.1518
No log 5.75 46 1.1694 0.5512 1.1694 1.0814
No log 6.0 48 0.8457 0.6571 0.8457 0.9196
No log 6.25 50 0.8401 0.6475 0.8401 0.9166
No log 6.5 52 0.8524 0.6569 0.8524 0.9233
No log 6.75 54 0.9563 0.6619 0.9563 0.9779
No log 7.0 56 1.6587 0.3741 1.6587 1.2879
No log 7.25 58 1.7114 0.3546 1.7114 1.3082
No log 7.5 60 1.1264 0.5379 1.1264 1.0613
No log 7.75 62 0.7990 0.7273 0.7990 0.8938
No log 8.0 64 0.7380 0.6993 0.7380 0.8591
No log 8.25 66 0.8104 0.7234 0.8104 0.9002
No log 8.5 68 1.1166 0.5793 1.1166 1.0567
No log 8.75 70 1.2192 0.5286 1.2192 1.1042
No log 9.0 72 1.0648 0.5846 1.0648 1.0319
No log 9.25 74 1.1797 0.5271 1.1797 1.0862
No log 9.5 76 1.1183 0.5271 1.1183 1.0575
No log 9.75 78 1.0274 0.6462 1.0274 1.0136
No log 10.0 80 0.9583 0.6562 0.9583 0.9789
No log 10.25 82 0.8749 0.6512 0.8749 0.9354
No log 10.5 84 0.8730 0.6716 0.8730 0.9344
No log 10.75 86 1.0355 0.5694 1.0355 1.0176
No log 11.0 88 1.3525 0.5698 1.3525 1.1630
No log 11.25 90 1.2799 0.5195 1.2799 1.1313
No log 11.5 92 1.2196 0.5034 1.2196 1.1043
No log 11.75 94 1.0260 0.6119 1.0260 1.0129
No log 12.0 96 1.0012 0.5556 1.0012 1.0006
No log 12.25 98 1.0046 0.6308 1.0046 1.0023
No log 12.5 100 1.1427 0.4964 1.1427 1.0689
No log 12.75 102 1.2576 0.4755 1.2576 1.1214
No log 13.0 104 1.2733 0.4966 1.2733 1.1284
No log 13.25 106 1.0226 0.6618 1.0226 1.0112
No log 13.5 108 0.9201 0.7194 0.9201 0.9592
No log 13.75 110 0.9727 0.6259 0.9727 0.9862
No log 14.0 112 0.9704 0.6216 0.9704 0.9851
No log 14.25 114 0.8100 0.7034 0.8100 0.9000
No log 14.5 116 0.8216 0.6620 0.8216 0.9064
No log 14.75 118 1.0284 0.5926 1.0284 1.0141
No log 15.0 120 1.1467 0.5850 1.1467 1.0708
No log 15.25 122 1.0504 0.6111 1.0504 1.0249
No log 15.5 124 0.8948 0.6618 0.8948 0.9460
No log 15.75 126 0.9033 0.6508 0.9033 0.9504
No log 16.0 128 0.9267 0.6032 0.9267 0.9627
No log 16.25 130 1.0021 0.6107 1.0021 1.0011
No log 16.5 132 1.0232 0.5938 1.0232 1.0116
No log 16.75 134 1.0181 0.6429 1.0181 1.0090
No log 17.0 136 1.0595 0.6093 1.0595 1.0293
No log 17.25 138 1.0922 0.6104 1.0922 1.0451
No log 17.5 140 0.9201 0.6667 0.9201 0.9592
No log 17.75 142 0.8927 0.7361 0.8927 0.9448
No log 18.0 144 1.0840 0.6216 1.0840 1.0411
No log 18.25 146 1.0815 0.6111 1.0815 1.0400
No log 18.5 148 0.8133 0.7068 0.8133 0.9018
No log 18.75 150 0.7630 0.6767 0.7630 0.8735
No log 19.0 152 0.7418 0.7 0.7418 0.8613
No log 19.25 154 0.7183 0.7518 0.7183 0.8475
No log 19.5 156 0.9407 0.6351 0.9407 0.9699
No log 19.75 158 1.1308 0.6194 1.1308 1.0634
No log 20.0 160 1.0770 0.6 1.0770 1.0378
No log 20.25 162 1.2141 0.5828 1.2141 1.1019
No log 20.5 164 1.0783 0.6143 1.0783 1.0384
No log 20.75 166 0.9227 0.6466 0.9227 0.9605
No log 21.0 168 0.8481 0.6950 0.8481 0.9209
No log 21.25 170 0.9505 0.6710 0.9505 0.9750
No log 21.5 172 1.1145 0.625 1.1145 1.0557
No log 21.75 174 1.3307 0.5714 1.3307 1.1536
No log 22.0 176 1.4338 0.5093 1.4338 1.1974
No log 22.25 178 1.3464 0.5590 1.3464 1.1603
No log 22.5 180 1.0743 0.5816 1.0743 1.0365
No log 22.75 182 0.8602 0.6963 0.8602 0.9275
No log 23.0 184 0.8335 0.6822 0.8335 0.9130
No log 23.25 186 0.8319 0.6614 0.8319 0.9121
No log 23.5 188 0.8881 0.6667 0.8881 0.9424
No log 23.75 190 1.0436 0.5778 1.0436 1.0216
No log 24.0 192 1.1918 0.5816 1.1918 1.0917
No log 24.25 194 1.0548 0.6099 1.0548 1.0270
No log 24.5 196 0.8004 0.6944 0.8004 0.8946
No log 24.75 198 0.6809 0.75 0.6809 0.8252
No log 25.0 200 0.6927 0.7413 0.6927 0.8323
No log 25.25 202 0.7597 0.7376 0.7597 0.8716
No log 25.5 204 0.9863 0.5714 0.9863 0.9931
No log 25.75 206 1.5548 0.5363 1.5548 1.2469
No log 26.0 208 1.9109 0.4848 1.9109 1.3823
No log 26.25 210 1.8090 0.4516 1.8090 1.3450
No log 26.5 212 1.4038 0.4744 1.4038 1.1848
No log 26.75 214 1.0982 0.6015 1.0982 1.0479
No log 27.0 216 0.9441 0.64 0.9441 0.9716
No log 27.25 218 0.8690 0.6508 0.8690 0.9322
No log 27.5 220 0.8980 0.6522 0.8980 0.9476
No log 27.75 222 0.9651 0.6232 0.9651 0.9824
No log 28.0 224 1.0026 0.6029 1.0026 1.0013
No log 28.25 226 0.9311 0.6377 0.9311 0.9649
No log 28.5 228 0.7594 0.6901 0.7594 0.8714
No log 28.75 230 0.6972 0.7211 0.6972 0.8350
No log 29.0 232 0.8108 0.6667 0.8108 0.9005
No log 29.25 234 0.8886 0.6667 0.8886 0.9426
No log 29.5 236 0.8329 0.6846 0.8329 0.9126
No log 29.75 238 0.9348 0.6133 0.9348 0.9669
No log 30.0 240 1.0132 0.6065 1.0132 1.0066
No log 30.25 242 0.9766 0.6040 0.9766 0.9883
No log 30.5 244 0.8360 0.6714 0.8360 0.9143
No log 30.75 246 0.7462 0.7101 0.7462 0.8638
No log 31.0 248 0.7229 0.7059 0.7229 0.8502
No log 31.25 250 0.7320 0.7194 0.7320 0.8556
No log 31.5 252 0.8151 0.6853 0.8151 0.9028
No log 31.75 254 0.9163 0.6207 0.9163 0.9572
No log 32.0 256 1.0681 0.5811 1.0681 1.0335
No log 32.25 258 1.2252 0.5342 1.2252 1.1069
No log 32.5 260 1.0575 0.6197 1.0575 1.0283
No log 32.75 262 0.8359 0.7101 0.8359 0.9143
No log 33.0 264 0.7626 0.7023 0.7626 0.8732
No log 33.25 266 0.7741 0.7068 0.7741 0.8798
No log 33.5 268 0.8100 0.7068 0.8100 0.9000
No log 33.75 270 0.8567 0.6963 0.8567 0.9256
No log 34.0 272 0.9446 0.6269 0.9446 0.9719
No log 34.25 274 0.9539 0.6475 0.9539 0.9767
No log 34.5 276 0.8931 0.6714 0.8931 0.9450
No log 34.75 278 0.8309 0.7023 0.8309 0.9115
No log 35.0 280 0.8487 0.7023 0.8487 0.9212
No log 35.25 282 0.9441 0.6619 0.9441 0.9717
No log 35.5 284 1.0647 0.6471 1.0647 1.0319
No log 35.75 286 1.0482 0.6308 1.0482 1.0238
No log 36.0 288 1.0268 0.624 1.0268 1.0133
No log 36.25 290 1.0085 0.6129 1.0085 1.0042
No log 36.5 292 1.0605 0.6308 1.0605 1.0298
No log 36.75 294 1.1918 0.5674 1.1918 1.0917
No log 37.0 296 1.3213 0.5 1.3213 1.1495
No log 37.25 298 1.2890 0.5161 1.2890 1.1354
No log 37.5 300 1.1634 0.5676 1.1634 1.0786
No log 37.75 302 0.9560 0.6716 0.9560 0.9777
No log 38.0 304 0.8738 0.6815 0.8738 0.9348
No log 38.25 306 0.8495 0.6715 0.8495 0.9217
No log 38.5 308 0.8580 0.6809 0.8580 0.9263
No log 38.75 310 0.9014 0.6294 0.9014 0.9494
No log 39.0 312 0.8705 0.6809 0.8705 0.9330
No log 39.25 314 0.7634 0.6957 0.7634 0.8737
No log 39.5 316 0.7352 0.7153 0.7352 0.8574
No log 39.75 318 0.7603 0.6957 0.7603 0.8719
No log 40.0 320 0.8175 0.6950 0.8175 0.9041
No log 40.25 322 0.9782 0.6503 0.9782 0.9890
No log 40.5 324 1.0370 0.6514 1.0370 1.0183
No log 40.75 326 0.8475 0.6708 0.8475 0.9206
No log 41.0 328 0.7110 0.7075 0.7110 0.8432
No log 41.25 330 0.7380 0.6842 0.7380 0.8591
No log 41.5 332 0.8417 0.6957 0.8417 0.9175
No log 41.75 334 0.7915 0.6755 0.7915 0.8897
No log 42.0 336 0.8554 0.6383 0.8554 0.9249
No log 42.25 338 0.8519 0.6765 0.8519 0.9230
No log 42.5 340 0.8469 0.6812 0.8469 0.9203
No log 42.75 342 0.8243 0.6912 0.8243 0.9079
No log 43.0 344 0.8631 0.6912 0.8631 0.9290
No log 43.25 346 0.9410 0.6715 0.9410 0.9700
No log 43.5 348 1.0106 0.6277 1.0106 1.0053
No log 43.75 350 0.9939 0.6471 0.9939 0.9969
No log 44.0 352 0.9523 0.6618 0.9523 0.9759
No log 44.25 354 0.9637 0.6618 0.9637 0.9817
No log 44.5 356 1.0042 0.6471 1.0042 1.0021
No log 44.75 358 1.0852 0.6232 1.0852 1.0417
No log 45.0 360 1.2444 0.525 1.2444 1.1155
No log 45.25 362 1.3566 0.5424 1.3566 1.1647
No log 45.5 364 1.3405 0.5424 1.3405 1.1578
No log 45.75 366 1.2306 0.5342 1.2306 1.1093
No log 46.0 368 0.9911 0.6383 0.9911 0.9955
No log 46.25 370 0.8483 0.6716 0.8483 0.9210
No log 46.5 372 0.8155 0.6716 0.8155 0.9031
No log 46.75 374 0.8317 0.6716 0.8317 0.9120
No log 47.0 376 0.9380 0.6525 0.9380 0.9685
No log 47.25 378 1.0442 0.6667 1.0442 1.0219
No log 47.5 380 0.9907 0.6533 0.9907 0.9953
No log 47.75 382 0.8545 0.6957 0.8545 0.9244
No log 48.0 384 0.7984 0.7153 0.7984 0.8935
No log 48.25 386 0.8040 0.7164 0.8040 0.8967
No log 48.5 388 0.8079 0.7164 0.8079 0.8988
No log 48.75 390 0.7967 0.7164 0.7967 0.8926
No log 49.0 392 0.8132 0.7153 0.8132 0.9018
No log 49.25 394 0.8581 0.7153 0.8581 0.9263
No log 49.5 396 0.9294 0.6429 0.9294 0.9640
No log 49.75 398 0.9568 0.6277 0.9568 0.9781
No log 50.0 400 0.9593 0.6277 0.9593 0.9795
No log 50.25 402 0.9249 0.6667 0.9249 0.9617
No log 50.5 404 0.8882 0.7121 0.8882 0.9424
No log 50.75 406 0.8933 0.7194 0.8933 0.9451
No log 51.0 408 0.9321 0.6475 0.9321 0.9655
No log 51.25 410 0.9890 0.5942 0.9890 0.9945
No log 51.5 412 1.0307 0.5972 1.0307 1.0152
No log 51.75 414 0.9892 0.5926 0.9892 0.9946
No log 52.0 416 0.9504 0.6222 0.9504 0.9749
No log 52.25 418 0.9264 0.6567 0.9264 0.9625
No log 52.5 420 0.8731 0.6716 0.8731 0.9344
No log 52.75 422 0.8332 0.6963 0.8332 0.9128
No log 53.0 424 0.7815 0.7299 0.7815 0.8840
No log 53.25 426 0.7639 0.7391 0.7639 0.8740
No log 53.5 428 0.7904 0.7299 0.7904 0.8891
No log 53.75 430 0.8083 0.7083 0.8083 0.8990
No log 54.0 432 0.8071 0.7092 0.8071 0.8984
No log 54.25 434 0.8213 0.7101 0.8213 0.9062
No log 54.5 436 0.8386 0.7101 0.8386 0.9158
No log 54.75 438 0.8172 0.7391 0.8172 0.9040
No log 55.0 440 0.7935 0.7259 0.7935 0.8908
No log 55.25 442 0.8106 0.6963 0.8106 0.9004
No log 55.5 444 0.8529 0.6716 0.8529 0.9235
No log 55.75 446 0.9613 0.6423 0.9613 0.9804
No log 56.0 448 1.1032 0.5526 1.1032 1.0503
No log 56.25 450 1.1277 0.5676 1.1277 1.0619
No log 56.5 452 1.0341 0.6187 1.0341 1.0169
No log 56.75 454 0.9163 0.6370 0.9163 0.9572
No log 57.0 456 0.8363 0.6917 0.8363 0.9145
No log 57.25 458 0.8126 0.7299 0.8126 0.9014
No log 57.5 460 0.8046 0.7299 0.8046 0.8970
No log 57.75 462 0.8150 0.7286 0.8150 0.9028
No log 58.0 464 0.8306 0.6715 0.8306 0.9114
No log 58.25 466 0.8860 0.6528 0.8860 0.9413
No log 58.5 468 0.9279 0.6621 0.9279 0.9633
No log 58.75 470 0.9801 0.6294 0.9801 0.9900
No log 59.0 472 0.9580 0.6286 0.9580 0.9788
No log 59.25 474 0.9449 0.6471 0.9449 0.9721
No log 59.5 476 0.9457 0.6466 0.9457 0.9725
No log 59.75 478 0.9064 0.7015 0.9064 0.9520
No log 60.0 480 0.9013 0.7015 0.9013 0.9494
No log 60.25 482 0.9293 0.6131 0.9293 0.9640
No log 60.5 484 1.0466 0.5931 1.0466 1.0231
No log 60.75 486 1.2370 0.5509 1.2370 1.1122
No log 61.0 488 1.3091 0.5325 1.3091 1.1442
No log 61.25 490 1.2488 0.5325 1.2488 1.1175
No log 61.5 492 1.1442 0.5455 1.1442 1.0697
No log 61.75 494 1.0592 0.5714 1.0592 1.0292
No log 62.0 496 0.9891 0.6259 0.9891 0.9945
No log 62.25 498 1.0005 0.6301 1.0005 1.0002
0.3118 62.5 500 0.9967 0.6301 0.9967 0.9983
0.3118 62.75 502 1.0340 0.6301 1.0340 1.0169
0.3118 63.0 504 1.1171 0.5931 1.1171 1.0569
0.3118 63.25 506 1.1605 0.5931 1.1605 1.0773
0.3118 63.5 508 1.1307 0.5899 1.1307 1.0633
0.3118 63.75 510 1.0580 0.6131 1.0580 1.0286

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
5
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k1_task1_organization

Finetuned
(4222)
this model