ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k5_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0106
  • Qwk: 0.5994
  • Mse: 1.0106
  • Rmse: 1.0053

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0769 2 5.2079 -0.0368 5.2079 2.2821
No log 0.1538 4 3.3087 0.0101 3.3087 1.8190
No log 0.2308 6 2.2462 -0.0542 2.2462 1.4987
No log 0.3077 8 1.4124 0.1672 1.4124 1.1884
No log 0.3846 10 1.3341 0.1813 1.3341 1.1550
No log 0.4615 12 1.6603 -0.0969 1.6603 1.2885
No log 0.5385 14 1.4730 0.0195 1.4730 1.2137
No log 0.6154 16 1.3253 0.0779 1.3253 1.1512
No log 0.6923 18 1.2758 0.2597 1.2758 1.1295
No log 0.7692 20 1.3709 0.1560 1.3709 1.1708
No log 0.8462 22 1.2275 0.2495 1.2275 1.1079
No log 0.9231 24 1.1185 0.2199 1.1185 1.0576
No log 1.0 26 1.3060 0.1460 1.3060 1.1428
No log 1.0769 28 1.5053 0.0612 1.5053 1.2269
No log 1.1538 30 1.3617 0.1792 1.3617 1.1669
No log 1.2308 32 1.0225 0.2838 1.0225 1.0112
No log 1.3077 34 0.9321 0.4092 0.9321 0.9655
No log 1.3846 36 1.4948 0.4177 1.4948 1.2226
No log 1.4615 38 1.4888 0.3861 1.4888 1.2202
No log 1.5385 40 1.1731 0.4330 1.1731 1.0831
No log 1.6154 42 0.9162 0.4413 0.9162 0.9572
No log 1.6923 44 0.9521 0.5206 0.9521 0.9757
No log 1.7692 46 0.9486 0.5391 0.9486 0.9739
No log 1.8462 48 0.9258 0.5376 0.9258 0.9622
No log 1.9231 50 0.8843 0.5078 0.8843 0.9404
No log 2.0 52 0.8638 0.4552 0.8638 0.9294
No log 2.0769 54 0.8588 0.4081 0.8588 0.9267
No log 2.1538 56 0.8868 0.4900 0.8868 0.9417
No log 2.2308 58 0.9612 0.5211 0.9612 0.9804
No log 2.3077 60 1.0461 0.4652 1.0461 1.0228
No log 2.3846 62 0.9469 0.5838 0.9469 0.9731
No log 2.4615 64 0.7832 0.5811 0.7832 0.8850
No log 2.5385 66 0.7729 0.5769 0.7729 0.8791
No log 2.6154 68 0.8543 0.5802 0.8543 0.9243
No log 2.6923 70 0.8195 0.6375 0.8195 0.9053
No log 2.7692 72 0.7656 0.6258 0.7656 0.8750
No log 2.8462 74 0.8827 0.6264 0.8827 0.9395
No log 2.9231 76 1.0743 0.4522 1.0743 1.0365
No log 3.0 78 1.1539 0.3948 1.1539 1.0742
No log 3.0769 80 1.0917 0.4620 1.0917 1.0449
No log 3.1538 82 1.0254 0.5493 1.0254 1.0126
No log 3.2308 84 0.9058 0.5645 0.9058 0.9517
No log 3.3077 86 0.9397 0.5913 0.9397 0.9694
No log 3.3846 88 0.9118 0.5713 0.9118 0.9549
No log 3.4615 90 0.8595 0.5627 0.8595 0.9271
No log 3.5385 92 0.9100 0.6276 0.9100 0.9539
No log 3.6154 94 0.8277 0.6233 0.8277 0.9098
No log 3.6923 96 0.8144 0.6174 0.8144 0.9024
No log 3.7692 98 0.8372 0.5860 0.8372 0.9150
No log 3.8462 100 0.8239 0.5860 0.8239 0.9077
No log 3.9231 102 0.8027 0.6326 0.8027 0.8959
No log 4.0 104 0.9919 0.5261 0.9919 0.9959
No log 4.0769 106 1.1385 0.5021 1.1385 1.0670
No log 4.1538 108 0.9708 0.5540 0.9708 0.9853
No log 4.2308 110 0.8297 0.6213 0.8297 0.9109
No log 4.3077 112 0.8579 0.6319 0.8579 0.9262
No log 4.3846 114 1.0292 0.5342 1.0292 1.0145
No log 4.4615 116 1.1671 0.4818 1.1671 1.0803
No log 4.5385 118 1.1753 0.4965 1.1753 1.0841
No log 4.6154 120 1.0593 0.5658 1.0593 1.0292
No log 4.6923 122 0.8324 0.6329 0.8324 0.9123
No log 4.7692 124 0.6939 0.6671 0.6939 0.8330
No log 4.8462 126 0.7188 0.6600 0.7188 0.8478
No log 4.9231 128 0.8753 0.6098 0.8753 0.9356
No log 5.0 130 1.2487 0.5399 1.2487 1.1174
No log 5.0769 132 1.3379 0.5299 1.3379 1.1567
No log 5.1538 134 1.2219 0.5497 1.2219 1.1054
No log 5.2308 136 1.0782 0.6054 1.0782 1.0384
No log 5.3077 138 0.9443 0.6166 0.9443 0.9717
No log 5.3846 140 0.8961 0.6107 0.8961 0.9466
No log 5.4615 142 0.9481 0.5926 0.9481 0.9737
No log 5.5385 144 0.9075 0.5829 0.9075 0.9526
No log 5.6154 146 0.8981 0.5755 0.8981 0.9477
No log 5.6923 148 0.7976 0.6061 0.7976 0.8931
No log 5.7692 150 0.7561 0.6386 0.7561 0.8695
No log 5.8462 152 0.7601 0.6515 0.7601 0.8718
No log 5.9231 154 0.7528 0.6565 0.7528 0.8676
No log 6.0 156 0.8738 0.6293 0.8738 0.9348
No log 6.0769 158 0.9638 0.5992 0.9638 0.9818
No log 6.1538 160 1.0206 0.5801 1.0206 1.0103
No log 6.2308 162 0.9558 0.5835 0.9558 0.9777
No log 6.3077 164 0.8170 0.6676 0.8170 0.9039
No log 6.3846 166 0.7574 0.6811 0.7574 0.8703
No log 6.4615 168 0.8979 0.6519 0.8979 0.9476
No log 6.5385 170 1.1396 0.5749 1.1396 1.0675
No log 6.6154 172 1.2221 0.5505 1.2221 1.1055
No log 6.6923 174 1.2686 0.5286 1.2686 1.1263
No log 6.7692 176 1.1197 0.5503 1.1197 1.0582
No log 6.8462 178 1.1102 0.5262 1.1102 1.0537
No log 6.9231 180 1.1831 0.5020 1.1831 1.0877
No log 7.0 182 1.2721 0.4415 1.2721 1.1279
No log 7.0769 184 1.3656 0.4486 1.3656 1.1686
No log 7.1538 186 1.2548 0.5392 1.2548 1.1202
No log 7.2308 188 0.9321 0.5807 0.9321 0.9654
No log 7.3077 190 0.8390 0.5957 0.8390 0.9160
No log 7.3846 192 0.9148 0.6043 0.9148 0.9564
No log 7.4615 194 0.8444 0.5647 0.8444 0.9189
No log 7.5385 196 0.9132 0.5624 0.9132 0.9556
No log 7.6154 198 0.9750 0.5477 0.9750 0.9874
No log 7.6923 200 0.8557 0.5899 0.8557 0.9250
No log 7.7692 202 0.7044 0.6067 0.7044 0.8393
No log 7.8462 204 0.6877 0.6432 0.6877 0.8293
No log 7.9231 206 0.7658 0.6347 0.7658 0.8751
No log 8.0 208 1.1710 0.5476 1.1710 1.0821
No log 8.0769 210 1.3013 0.4905 1.3013 1.1408
No log 8.1538 212 1.0364 0.5264 1.0364 1.0180
No log 8.2308 214 0.9105 0.5753 0.9105 0.9542
No log 8.3077 216 0.9367 0.5621 0.9367 0.9678
No log 8.3846 218 1.1835 0.5107 1.1835 1.0879
No log 8.4615 220 1.5770 0.4588 1.5770 1.2558
No log 8.5385 222 1.6320 0.4164 1.6320 1.2775
No log 8.6154 224 1.4039 0.4726 1.4039 1.1849
No log 8.6923 226 0.9829 0.5824 0.9829 0.9914
No log 8.7692 228 0.6912 0.6856 0.6912 0.8314
No log 8.8462 230 0.6746 0.6973 0.6746 0.8213
No log 8.9231 232 0.6839 0.6844 0.6839 0.8270
No log 9.0 234 0.8366 0.6427 0.8366 0.9147
No log 9.0769 236 0.9447 0.5987 0.9447 0.9720
No log 9.1538 238 0.8696 0.6094 0.8696 0.9325
No log 9.2308 240 0.7764 0.6769 0.7764 0.8811
No log 9.3077 242 0.8091 0.6532 0.8091 0.8995
No log 9.3846 244 0.9351 0.5777 0.9351 0.9670
No log 9.4615 246 1.0501 0.5552 1.0501 1.0248
No log 9.5385 248 1.1270 0.5553 1.1270 1.0616
No log 9.6154 250 0.9820 0.6071 0.9820 0.9909
No log 9.6923 252 0.8268 0.6609 0.8268 0.9093
No log 9.7692 254 0.8194 0.6609 0.8194 0.9052
No log 9.8462 256 0.8983 0.6260 0.8983 0.9478
No log 9.9231 258 1.0159 0.5718 1.0159 1.0079
No log 10.0 260 1.2729 0.5374 1.2729 1.1282
No log 10.0769 262 1.2983 0.5194 1.2983 1.1394
No log 10.1538 264 1.1642 0.5229 1.1642 1.0790
No log 10.2308 266 0.9314 0.6167 0.9314 0.9651
No log 10.3077 268 0.7816 0.6434 0.7816 0.8841
No log 10.3846 270 0.7893 0.6445 0.7893 0.8884
No log 10.4615 272 0.8896 0.5925 0.8896 0.9432
No log 10.5385 274 1.0813 0.5486 1.0813 1.0398
No log 10.6154 276 1.1086 0.5310 1.1086 1.0529
No log 10.6923 278 0.9933 0.5671 0.9933 0.9967
No log 10.7692 280 0.8894 0.6032 0.8894 0.9431
No log 10.8462 282 0.8031 0.6372 0.8031 0.8962
No log 10.9231 284 0.8561 0.6122 0.8561 0.9253
No log 11.0 286 0.9757 0.6056 0.9757 0.9878
No log 11.0769 288 1.0982 0.5710 1.0982 1.0479
No log 11.1538 290 1.0009 0.6279 1.0009 1.0004
No log 11.2308 292 0.8355 0.6321 0.8355 0.9141
No log 11.3077 294 0.8099 0.6283 0.8099 0.8999
No log 11.3846 296 0.7079 0.6841 0.7079 0.8414
No log 11.4615 298 0.6561 0.6379 0.6561 0.8100
No log 11.5385 300 0.6673 0.6903 0.6673 0.8169
No log 11.6154 302 0.7684 0.6444 0.7684 0.8766
No log 11.6923 304 0.8240 0.6438 0.8240 0.9077
No log 11.7692 306 0.7425 0.6284 0.7425 0.8617
No log 11.8462 308 0.7307 0.6634 0.7307 0.8548
No log 11.9231 310 0.8283 0.6196 0.8283 0.9101
No log 12.0 312 1.0033 0.5540 1.0033 1.0017
No log 12.0769 314 1.1668 0.5610 1.1668 1.0802
No log 12.1538 316 1.1229 0.5597 1.1229 1.0597
No log 12.2308 318 1.0320 0.5616 1.0320 1.0159
No log 12.3077 320 1.0127 0.5616 1.0127 1.0063
No log 12.3846 322 1.0785 0.5686 1.0785 1.0385
No log 12.4615 324 1.0527 0.5566 1.0527 1.0260
No log 12.5385 326 0.9645 0.6135 0.9645 0.9821
No log 12.6154 328 0.8283 0.6731 0.8283 0.9101
No log 12.6923 330 0.8145 0.6778 0.8145 0.9025
No log 12.7692 332 0.8057 0.6785 0.8057 0.8976
No log 12.8462 334 0.8548 0.6741 0.8548 0.9245
No log 12.9231 336 1.0168 0.6014 1.0168 1.0083
No log 13.0 338 0.9836 0.6017 0.9836 0.9918
No log 13.0769 340 0.8659 0.6398 0.8659 0.9305
No log 13.1538 342 0.7764 0.6744 0.7764 0.8811
No log 13.2308 344 0.7934 0.6753 0.7934 0.8907
No log 13.3077 346 0.8904 0.6695 0.8904 0.9436
No log 13.3846 348 1.0953 0.6059 1.0953 1.0466
No log 13.4615 350 1.2352 0.5992 1.2352 1.1114
No log 13.5385 352 1.3259 0.5722 1.3259 1.1515
No log 13.6154 354 1.3200 0.5630 1.3200 1.1489
No log 13.6923 356 1.1450 0.5710 1.1450 1.0700
No log 13.7692 358 1.0459 0.5683 1.0459 1.0227
No log 13.8462 360 0.9284 0.5998 0.9284 0.9635
No log 13.9231 362 0.9641 0.5894 0.9641 0.9819
No log 14.0 364 1.0213 0.6230 1.0213 1.0106
No log 14.0769 366 1.0513 0.6190 1.0513 1.0253
No log 14.1538 368 0.9301 0.6033 0.9301 0.9644
No log 14.2308 370 0.8889 0.5991 0.8889 0.9428
No log 14.3077 372 0.9484 0.6016 0.9484 0.9738
No log 14.3846 374 0.9997 0.5991 0.9997 0.9998
No log 14.4615 376 0.9990 0.5917 0.9990 0.9995
No log 14.5385 378 1.1116 0.5710 1.1116 1.0543
No log 14.6154 380 1.1909 0.5690 1.1909 1.0913
No log 14.6923 382 1.1608 0.5733 1.1608 1.0774
No log 14.7692 384 1.0444 0.5953 1.0444 1.0220
No log 14.8462 386 0.8790 0.5991 0.8790 0.9375
No log 14.9231 388 0.9225 0.6110 0.9225 0.9605
No log 15.0 390 1.0123 0.6268 1.0123 1.0061
No log 15.0769 392 1.0460 0.6044 1.0460 1.0228
No log 15.1538 394 1.0661 0.5662 1.0661 1.0325
No log 15.2308 396 0.9256 0.5541 0.9256 0.9621
No log 15.3077 398 0.7952 0.6256 0.7952 0.8918
No log 15.3846 400 0.7695 0.6414 0.7695 0.8772
No log 15.4615 402 0.7856 0.6676 0.7856 0.8864
No log 15.5385 404 0.9582 0.6049 0.9582 0.9789
No log 15.6154 406 0.9912 0.6233 0.9912 0.9956
No log 15.6923 408 0.9337 0.6357 0.9337 0.9663
No log 15.7692 410 0.8510 0.6563 0.8510 0.9225
No log 15.8462 412 0.6988 0.6317 0.6988 0.8360
No log 15.9231 414 0.7334 0.6427 0.7334 0.8564
No log 16.0 416 0.8590 0.6513 0.8590 0.9268
No log 16.0769 418 1.1314 0.5897 1.1314 1.0637
No log 16.1538 420 1.2719 0.5567 1.2719 1.1278
No log 16.2308 422 1.2323 0.5602 1.2323 1.1101
No log 16.3077 424 0.9685 0.5633 0.9685 0.9841
No log 16.3846 426 0.8154 0.6256 0.8154 0.9030
No log 16.4615 428 0.8379 0.5871 0.8379 0.9154
No log 16.5385 430 0.9499 0.5756 0.9499 0.9746
No log 16.6154 432 1.0353 0.5683 1.0353 1.0175
No log 16.6923 434 1.0232 0.5779 1.0232 1.0115
No log 16.7692 436 1.0118 0.5647 1.0118 1.0059
No log 16.8462 438 1.0709 0.5486 1.0709 1.0348
No log 16.9231 440 1.2774 0.5604 1.2774 1.1302
No log 17.0 442 1.3161 0.5443 1.3161 1.1472
No log 17.0769 444 1.3593 0.5415 1.3593 1.1659
No log 17.1538 446 1.4020 0.5129 1.4020 1.1840
No log 17.2308 448 1.2767 0.4375 1.2767 1.1299
No log 17.3077 450 1.0581 0.5271 1.0581 1.0286
No log 17.3846 452 0.8676 0.5911 0.8676 0.9314
No log 17.4615 454 0.7833 0.6233 0.7833 0.8851
No log 17.5385 456 0.8300 0.6073 0.8300 0.9110
No log 17.6154 458 1.0336 0.5684 1.0336 1.0167
No log 17.6923 460 1.3978 0.5302 1.3978 1.1823
No log 17.7692 462 1.5561 0.4807 1.5561 1.2475
No log 17.8462 464 1.4251 0.5160 1.4251 1.1938
No log 17.9231 466 1.1300 0.5926 1.1300 1.0630
No log 18.0 468 0.8972 0.6333 0.8972 0.9472
No log 18.0769 470 0.8232 0.6475 0.8232 0.9073
No log 18.1538 472 0.8573 0.6319 0.8573 0.9259
No log 18.2308 474 0.9925 0.6142 0.9925 0.9962
No log 18.3077 476 1.0085 0.6252 1.0085 1.0043
No log 18.3846 478 0.9454 0.6302 0.9454 0.9723
No log 18.4615 480 0.9083 0.6380 0.9083 0.9531
No log 18.5385 482 0.7798 0.6684 0.7798 0.8831
No log 18.6154 484 0.7209 0.6486 0.7209 0.8491
No log 18.6923 486 0.7104 0.6486 0.7104 0.8428
No log 18.7692 488 0.7504 0.6694 0.7504 0.8662
No log 18.8462 490 0.8560 0.6456 0.8560 0.9252
No log 18.9231 492 0.9171 0.5890 0.9171 0.9577
No log 19.0 494 0.8669 0.6456 0.8669 0.9311
No log 19.0769 496 0.7599 0.6734 0.7599 0.8717
No log 19.1538 498 0.7085 0.6976 0.7085 0.8417
0.3934 19.2308 500 0.6944 0.6945 0.6944 0.8333
0.3934 19.3077 502 0.7581 0.6807 0.7581 0.8707
0.3934 19.3846 504 0.9469 0.6139 0.9469 0.9731
0.3934 19.4615 506 1.0363 0.6358 1.0363 1.0180
0.3934 19.5385 508 1.0110 0.5991 1.0110 1.0055
0.3934 19.6154 510 1.0106 0.5994 1.0106 1.0053

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
6
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k5_task1_organization

Finetuned
(4222)
this model