ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k4_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8455
  • Qwk: -0.0334
  • Mse: 0.8455
  • Rmse: 0.9195

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1538 2 3.9708 0.0103 3.9708 1.9927
No log 0.3077 4 2.1363 0.0672 2.1363 1.4616
No log 0.4615 6 1.8444 0.0304 1.8444 1.3581
No log 0.6154 8 1.1100 -0.0423 1.1100 1.0536
No log 0.7692 10 0.6934 0.0506 0.6934 0.8327
No log 0.9231 12 0.7235 0.1021 0.7235 0.8506
No log 1.0769 14 0.8884 0.0378 0.8884 0.9425
No log 1.2308 16 1.0089 0.0083 1.0089 1.0044
No log 1.3846 18 1.2718 -0.0500 1.2718 1.1278
No log 1.5385 20 1.3198 -0.0500 1.3198 1.1488
No log 1.6923 22 1.3811 -0.0500 1.3811 1.1752
No log 1.8462 24 0.9302 0.0006 0.9302 0.9645
No log 2.0 26 0.7676 -0.0188 0.7676 0.8761
No log 2.1538 28 0.9214 -0.1261 0.9214 0.9599
No log 2.3077 30 1.2921 -0.0510 1.2921 1.1367
No log 2.4615 32 1.6775 -0.0029 1.6775 1.2952
No log 2.6154 34 1.3327 0.0 1.3327 1.1544
No log 2.7692 36 1.0668 0.0298 1.0668 1.0329
No log 2.9231 38 0.9820 -0.0677 0.9820 0.9909
No log 3.0769 40 0.8354 -0.0823 0.8354 0.9140
No log 3.2308 42 0.8197 -0.0812 0.8197 0.9054
No log 3.3846 44 0.8626 0.0377 0.8626 0.9288
No log 3.5385 46 0.7737 0.0549 0.7737 0.8796
No log 3.6923 48 0.8400 0.0377 0.8400 0.9165
No log 3.8462 50 0.7805 0.0628 0.7805 0.8835
No log 4.0 52 0.7993 0.0225 0.7993 0.8941
No log 4.1538 54 0.7891 -0.0240 0.7891 0.8883
No log 4.3077 56 0.9479 -0.0459 0.9479 0.9736
No log 4.4615 58 0.9849 -0.0474 0.9849 0.9924
No log 4.6154 60 0.7652 0.0807 0.7652 0.8747
No log 4.7692 62 0.7079 0.0506 0.7079 0.8413
No log 4.9231 64 0.6942 -0.0035 0.6942 0.8332
No log 5.0769 66 0.7808 0.1387 0.7808 0.8836
No log 5.2308 68 3.2423 -0.0160 3.2423 1.8006
No log 5.3846 70 2.9777 -0.0043 2.9777 1.7256
No log 5.5385 72 1.4978 0.0443 1.4978 1.2238
No log 5.6923 74 0.8540 0.0071 0.8540 0.9241
No log 5.8462 76 0.7598 -0.0679 0.7598 0.8717
No log 6.0 78 0.7303 -0.1223 0.7303 0.8546
No log 6.1538 80 0.7148 -0.1223 0.7148 0.8454
No log 6.3077 82 0.7373 0.0670 0.7373 0.8587
No log 6.4615 84 0.9607 -0.0425 0.9607 0.9801
No log 6.6154 86 1.0105 -0.1263 1.0105 1.0052
No log 6.7692 88 0.7291 0.1627 0.7291 0.8539
No log 6.9231 90 0.6974 0.0909 0.6974 0.8351
No log 7.0769 92 0.7238 0.0807 0.7238 0.8508
No log 7.2308 94 0.7849 0.1627 0.7849 0.8859
No log 7.3846 96 0.7506 0.0857 0.7506 0.8664
No log 7.5385 98 0.7671 0.0355 0.7671 0.8758
No log 7.6923 100 0.8345 0.0807 0.8345 0.9135
No log 7.8462 102 0.9679 -0.0767 0.9679 0.9838
No log 8.0 104 0.8172 0.0807 0.8172 0.9040
No log 8.1538 106 0.7732 -0.0385 0.7732 0.8793
No log 8.3077 108 0.7033 -0.0035 0.7033 0.8387
No log 8.4615 110 0.8184 0.0099 0.8184 0.9047
No log 8.6154 112 1.1162 -0.0236 1.1162 1.0565
No log 8.7692 114 1.1310 -0.0236 1.1310 1.0635
No log 8.9231 116 0.8335 0.0071 0.8335 0.9130
No log 9.0769 118 0.7433 0.0374 0.7433 0.8621
No log 9.2308 120 0.7913 0.0318 0.7913 0.8896
No log 9.3846 122 0.9118 -0.1580 0.9118 0.9549
No log 9.5385 124 1.0052 -0.0347 1.0052 1.0026
No log 9.6923 126 1.0565 -0.0612 1.0565 1.0278
No log 9.8462 128 1.0882 -0.0960 1.0882 1.0431
No log 10.0 130 1.1531 -0.0870 1.1531 1.0738
No log 10.1538 132 1.0458 -0.1405 1.0458 1.0227
No log 10.3077 134 0.9713 -0.1086 0.9713 0.9855
No log 10.4615 136 1.0063 -0.1633 1.0063 1.0032
No log 10.6154 138 1.1517 -0.1253 1.1517 1.0732
No log 10.7692 140 1.0278 -0.1234 1.0278 1.0138
No log 10.9231 142 0.9454 0.0913 0.9454 0.9723
No log 11.0769 144 0.9435 -0.1449 0.9435 0.9713
No log 11.2308 146 1.1687 -0.0079 1.1687 1.0811
No log 11.3846 148 1.0037 -0.1152 1.0037 1.0019
No log 11.5385 150 0.8623 0.1267 0.8623 0.9286
No log 11.6923 152 0.8615 0.1038 0.8615 0.9282
No log 11.8462 154 0.8533 0.0482 0.8533 0.9238
No log 12.0 156 0.9178 -0.0322 0.9178 0.9580
No log 12.1538 158 0.9116 -0.0425 0.9116 0.9548
No log 12.3077 160 0.8118 0.0043 0.8118 0.9010
No log 12.4615 162 0.7661 0.0828 0.7661 0.8753
No log 12.6154 164 0.8270 0.0570 0.8270 0.9094
No log 12.7692 166 0.8552 -0.0132 0.8552 0.9248
No log 12.9231 168 1.0865 0.0348 1.0865 1.0424
No log 13.0769 170 0.9987 -0.0679 0.9987 0.9994
No log 13.2308 172 1.0021 0.0007 1.0021 1.0011
No log 13.3846 174 1.1000 0.0169 1.1000 1.0488
No log 13.5385 176 0.9326 0.0632 0.9326 0.9657
No log 13.6923 178 0.8863 -0.0156 0.8863 0.9415
No log 13.8462 180 1.0689 0.0250 1.0689 1.0339
No log 14.0 182 0.9343 0.0711 0.9343 0.9666
No log 14.1538 184 0.7972 -0.0506 0.7972 0.8928
No log 14.3077 186 0.8992 0.0699 0.8992 0.9483
No log 14.4615 188 0.8724 0.0268 0.8724 0.9340
No log 14.6154 190 0.7967 -0.0488 0.7967 0.8926
No log 14.7692 192 0.8525 0.0146 0.8525 0.9233
No log 14.9231 194 0.9153 -0.0408 0.9153 0.9567
No log 15.0769 196 0.8589 0.0956 0.8589 0.9267
No log 15.2308 198 0.7996 0.1734 0.7996 0.8942
No log 15.3846 200 0.8365 0.0944 0.8365 0.9146
No log 15.5385 202 0.8597 0.0955 0.8597 0.9272
No log 15.6923 204 0.8799 0.0584 0.8799 0.9380
No log 15.8462 206 0.8936 0.0062 0.8936 0.9453
No log 16.0 208 0.8332 0.1051 0.8332 0.9128
No log 16.1538 210 0.8458 0.1345 0.8458 0.9197
No log 16.3077 212 0.8614 0.0833 0.8614 0.9281
No log 16.4615 214 0.8788 -0.0506 0.8788 0.9374
No log 16.6154 216 0.9143 0.0218 0.9143 0.9562
No log 16.7692 218 0.9381 -0.0813 0.9381 0.9685
No log 16.9231 220 0.8838 0.0827 0.8838 0.9401
No log 17.0769 222 0.8135 0.0277 0.8135 0.9020
No log 17.2308 224 0.8804 -0.0008 0.8804 0.9383
No log 17.3846 226 0.8813 -0.0008 0.8813 0.9388
No log 17.5385 228 0.7996 0.0639 0.7996 0.8942
No log 17.6923 230 0.8026 0.0940 0.8026 0.8959
No log 17.8462 232 0.8444 0.0952 0.8444 0.9189
No log 18.0 234 0.8599 0.1942 0.8599 0.9273
No log 18.1538 236 1.0942 0.0257 1.0942 1.0460
No log 18.3077 238 1.1041 0.0257 1.1041 1.0508
No log 18.4615 240 0.8735 0.1561 0.8735 0.9346
No log 18.6154 242 0.8524 -0.0014 0.8524 0.9232
No log 18.7692 244 0.8756 0.0070 0.8756 0.9357
No log 18.9231 246 0.7813 0.1408 0.7813 0.8839
No log 19.0769 248 0.7695 0.1553 0.7695 0.8772
No log 19.2308 250 0.8860 0.0346 0.8860 0.9413
No log 19.3846 252 0.8875 -0.0031 0.8875 0.9421
No log 19.5385 254 0.7744 0.1189 0.7744 0.8800
No log 19.6923 256 0.8116 0.0249 0.8116 0.9009
No log 19.8462 258 0.8054 0.0633 0.8054 0.8974
No log 20.0 260 0.8094 0.0277 0.8094 0.8997
No log 20.1538 262 0.9481 -0.0345 0.9481 0.9737
No log 20.3077 264 0.9905 0.0015 0.9905 0.9952
No log 20.4615 266 0.9476 0.1132 0.9476 0.9735
No log 20.6154 268 0.9365 0.0818 0.9365 0.9677
No log 20.7692 270 0.9048 0.0306 0.9048 0.9512
No log 20.9231 272 0.8954 0.0509 0.8954 0.9463
No log 21.0769 274 0.8633 0.0884 0.8633 0.9291
No log 21.2308 276 0.8585 0.0541 0.8585 0.9266
No log 21.3846 278 0.8764 0.0245 0.8764 0.9361
No log 21.5385 280 0.9255 0.1591 0.9255 0.9620
No log 21.6923 282 0.9218 0.0838 0.9218 0.9601
No log 21.8462 284 0.9218 0.0520 0.9218 0.9601
No log 22.0 286 0.9348 0.0084 0.9348 0.9668
No log 22.1538 288 0.8945 0.0218 0.8945 0.9458
No log 22.3077 290 0.9073 0.0421 0.9073 0.9525
No log 22.4615 292 0.9300 0.0563 0.9300 0.9644
No log 22.6154 294 0.9456 0.0552 0.9456 0.9724
No log 22.7692 296 0.9121 0.0552 0.9121 0.9550
No log 22.9231 298 0.9166 0.0200 0.9166 0.9574
No log 23.0769 300 0.9221 0.0262 0.9221 0.9603
No log 23.2308 302 0.9311 -0.0047 0.9311 0.9649
No log 23.3846 304 0.9311 -0.0117 0.9311 0.9650
No log 23.5385 306 0.9093 -0.0117 0.9093 0.9536
No log 23.6923 308 0.8917 -0.0491 0.8917 0.9443
No log 23.8462 310 0.9024 0.0875 0.9024 0.9499
No log 24.0 312 0.8856 0.0870 0.8856 0.9410
No log 24.1538 314 0.8925 0.0208 0.8925 0.9447
No log 24.3077 316 0.9409 -0.0031 0.9409 0.9700
No log 24.4615 318 0.8871 0.0016 0.8871 0.9419
No log 24.6154 320 0.8135 0.0709 0.8135 0.9020
No log 24.7692 322 0.7935 0.0749 0.7935 0.8908
No log 24.9231 324 0.8053 0.0893 0.8053 0.8974
No log 25.0769 326 0.8610 0.1078 0.8610 0.9279
No log 25.2308 328 0.8503 0.1078 0.8503 0.9221
No log 25.3846 330 0.7835 0.0481 0.7835 0.8851
No log 25.5385 332 0.7851 0.1604 0.7851 0.8861
No log 25.6923 334 0.8046 0.1604 0.8046 0.8970
No log 25.8462 336 0.8407 0.0526 0.8407 0.9169
No log 26.0 338 0.8881 0.0964 0.8881 0.9424
No log 26.1538 340 0.8598 0.0559 0.8598 0.9273
No log 26.3077 342 0.8067 0.1232 0.8067 0.8982
No log 26.4615 344 0.8379 0.0095 0.8379 0.9154
No log 26.6154 346 0.8813 0.0442 0.8813 0.9388
No log 26.7692 348 0.8392 -0.1083 0.8392 0.9161
No log 26.9231 350 0.8249 0.0474 0.8249 0.9082
No log 27.0769 352 0.9123 0.0347 0.9123 0.9552
No log 27.2308 354 0.9027 0.0590 0.9027 0.9501
No log 27.3846 356 0.8557 0.0771 0.8557 0.9251
No log 27.5385 358 0.9191 0.0174 0.9191 0.9587
No log 27.6923 360 0.9434 0.0174 0.9434 0.9713
No log 27.8462 362 0.9102 0.1519 0.9102 0.9540
No log 28.0 364 1.0073 0.0301 1.0073 1.0036
No log 28.1538 366 1.0219 -0.0295 1.0219 1.0109
No log 28.3077 368 0.9180 0.0879 0.9180 0.9581
No log 28.4615 370 0.9593 -0.0679 0.9593 0.9794
No log 28.6154 372 1.0691 -0.0797 1.0691 1.0340
No log 28.7692 374 0.9994 -0.0642 0.9994 0.9997
No log 28.9231 376 0.9805 0.0768 0.9805 0.9902
No log 29.0769 378 0.9965 0.0560 0.9965 0.9983
No log 29.2308 380 0.9365 0.0875 0.9365 0.9678
No log 29.3846 382 0.8905 0.0236 0.8905 0.9437
No log 29.5385 384 0.9004 -0.0391 0.9004 0.9489
No log 29.6923 386 0.9100 -0.0442 0.9100 0.9539
No log 29.8462 388 0.8424 0.0152 0.8424 0.9178
No log 30.0 390 0.8125 0.0085 0.8125 0.9014
No log 30.1538 392 0.8505 0.0209 0.8505 0.9222
No log 30.3077 394 0.8844 0.0995 0.8844 0.9404
No log 30.4615 396 0.8852 0.0810 0.8852 0.9408
No log 30.6154 398 0.8902 -0.0095 0.8902 0.9435
No log 30.7692 400 0.8754 -0.0049 0.8754 0.9356
No log 30.9231 402 0.8701 0.0330 0.8701 0.9328
No log 31.0769 404 0.8627 0.0330 0.8627 0.9288
No log 31.2308 406 0.9344 -0.0728 0.9344 0.9666
No log 31.3846 408 0.9616 -0.0315 0.9616 0.9806
No log 31.5385 410 0.9539 -0.1004 0.9539 0.9767
No log 31.6923 412 0.9476 0.1166 0.9476 0.9734
No log 31.8462 414 1.0120 0.0365 1.0120 1.0060
No log 32.0 416 0.9692 0.0028 0.9692 0.9845
No log 32.1538 418 0.8665 0.1635 0.8665 0.9309
No log 32.3077 420 0.8930 -0.0192 0.8930 0.9450
No log 32.4615 422 1.0538 -0.0050 1.0538 1.0266
No log 32.6154 424 1.0691 -0.0073 1.0691 1.0340
No log 32.7692 426 0.9237 -0.0618 0.9237 0.9611
No log 32.9231 428 0.8454 0.0 0.8454 0.9194
No log 33.0769 430 0.8192 -0.0426 0.8192 0.9051
No log 33.2308 432 0.7975 -0.0541 0.7975 0.8930
No log 33.3846 434 0.8216 -0.0704 0.8216 0.9064
No log 33.5385 436 0.8511 -0.0221 0.8511 0.9225
No log 33.6923 438 0.8665 -0.0563 0.8665 0.9309
No log 33.8462 440 0.8623 -0.0970 0.8623 0.9286
No log 34.0 442 0.8440 0.0791 0.8440 0.9187
No log 34.1538 444 0.8174 0.0412 0.8174 0.9041
No log 34.3077 446 0.7731 0.0357 0.7731 0.8793
No log 34.4615 448 0.7684 0.0768 0.7684 0.8766
No log 34.6154 450 0.8121 0.1095 0.8121 0.9012
No log 34.7692 452 0.9308 0.0837 0.9308 0.9648
No log 34.9231 454 0.9511 -0.0030 0.9511 0.9753
No log 35.0769 456 0.8755 0.0236 0.8755 0.9357
No log 35.2308 458 0.8515 0.0570 0.8515 0.9228
No log 35.3846 460 0.8692 0.0239 0.8692 0.9323
No log 35.5385 462 0.8295 0.0606 0.8295 0.9107
No log 35.6923 464 0.8171 0.0283 0.8171 0.9039
No log 35.8462 466 0.8829 -0.0283 0.8829 0.9396
No log 36.0 468 0.9135 0.0065 0.9135 0.9558
No log 36.1538 470 0.8950 0.0161 0.8950 0.9460
No log 36.3077 472 0.9149 -0.0212 0.9149 0.9565
No log 36.4615 474 0.8805 0.0200 0.8805 0.9383
No log 36.6154 476 0.8436 0.1136 0.8436 0.9185
No log 36.7692 478 0.8505 0.0581 0.8505 0.9222
No log 36.9231 480 0.8392 0.0581 0.8392 0.9161
No log 37.0769 482 0.8135 0.0053 0.8135 0.9019
No log 37.2308 484 0.8044 0.1573 0.8044 0.8969
No log 37.3846 486 0.8101 0.1139 0.8101 0.9000
No log 37.5385 488 0.8092 0.0733 0.8092 0.8995
No log 37.6923 490 0.8171 0.0733 0.8171 0.9039
No log 37.8462 492 0.8221 0.1135 0.8221 0.9067
No log 38.0 494 0.8282 0.0697 0.8282 0.9101
No log 38.1538 496 0.8495 0.0697 0.8495 0.9217
No log 38.3077 498 0.8514 0.0697 0.8514 0.9227
0.3218 38.4615 500 0.8465 0.0733 0.8465 0.9201
0.3218 38.6154 502 0.8477 0.0697 0.8477 0.9207
0.3218 38.7692 504 0.8829 0.0129 0.8829 0.9396
0.3218 38.9231 506 0.8696 0.0697 0.8696 0.9325
0.3218 39.0769 508 0.8832 0.0861 0.8832 0.9398
0.3218 39.2308 510 0.9329 0.0673 0.9329 0.9659
0.3218 39.3846 512 0.8850 -0.0238 0.8850 0.9408
0.3218 39.5385 514 0.8494 0.1094 0.8494 0.9216
0.3218 39.6923 516 0.8709 -0.0373 0.8709 0.9332
0.3218 39.8462 518 0.8472 -0.0373 0.8472 0.9204
0.3218 40.0 520 0.7904 0.0680 0.7904 0.8891
0.3218 40.1538 522 0.7983 -0.0370 0.7983 0.8935
0.3218 40.3077 524 0.8455 -0.0334 0.8455 0.9195

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
9
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k4_task3_organization

Finetuned
(4222)
this model