ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k5_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9105
  • Qwk: 0.0545
  • Mse: 0.9105
  • Rmse: 0.9542

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.125 2 3.5160 0.0048 3.5160 1.8751
No log 0.25 4 1.8904 0.0943 1.8904 1.3749
No log 0.375 6 1.4957 -0.0460 1.4957 1.2230
No log 0.5 8 1.3978 0.0731 1.3978 1.1823
No log 0.625 10 1.4233 0.0677 1.4233 1.1930
No log 0.75 12 1.2199 0.0493 1.2199 1.1045
No log 0.875 14 0.7043 0.0460 0.7043 0.8392
No log 1.0 16 0.7052 -0.0035 0.7052 0.8398
No log 1.125 18 0.9155 0.1374 0.9155 0.9568
No log 1.25 20 0.9848 0.0469 0.9848 0.9924
No log 1.375 22 0.8832 0.1589 0.8832 0.9398
No log 1.5 24 0.8492 0.1107 0.8492 0.9215
No log 1.625 26 1.2594 0.0048 1.2594 1.1222
No log 1.75 28 1.4924 0.0048 1.4924 1.2216
No log 1.875 30 1.2713 0.0048 1.2713 1.1275
No log 2.0 32 0.9138 0.0353 0.9138 0.9559
No log 2.125 34 0.8757 -0.0595 0.8757 0.9358
No log 2.25 36 0.8938 -0.0923 0.8938 0.9454
No log 2.375 38 1.0338 0.0006 1.0338 1.0168
No log 2.5 40 0.8572 -0.0408 0.8572 0.9258
No log 2.625 42 0.8017 -0.0725 0.8017 0.8954
No log 2.75 44 0.8571 0.0129 0.8571 0.9258
No log 2.875 46 1.0544 0.0810 1.0544 1.0268
No log 3.0 48 1.9536 0.0659 1.9536 1.3977
No log 3.125 50 2.2395 0.0408 2.2395 1.4965
No log 3.25 52 1.5815 0.0095 1.5815 1.2576
No log 3.375 54 0.8227 0.0225 0.8227 0.9070
No log 3.5 56 0.7734 -0.0069 0.7734 0.8794
No log 3.625 58 0.7791 -0.0188 0.7791 0.8827
No log 3.75 60 1.0952 -0.0331 1.0952 1.0465
No log 3.875 62 1.4573 -0.0479 1.4573 1.2072
No log 4.0 64 1.2023 -0.0398 1.2023 1.0965
No log 4.125 66 0.7852 0.1148 0.7852 0.8861
No log 4.25 68 0.7154 0.0857 0.7154 0.8458
No log 4.375 70 0.7732 0.0282 0.7732 0.8793
No log 4.5 72 1.0713 -0.0253 1.0713 1.0350
No log 4.625 74 1.1923 0.0025 1.1923 1.0919
No log 4.75 76 1.0888 -0.0471 1.0888 1.0435
No log 4.875 78 1.1670 -0.0156 1.1670 1.0803
No log 5.0 80 1.4054 -0.0511 1.4054 1.1855
No log 5.125 82 1.2606 -0.0434 1.2606 1.1228
No log 5.25 84 1.0436 -0.0225 1.0436 1.0216
No log 5.375 86 0.9533 0.0771 0.9533 0.9764
No log 5.5 88 1.0780 -0.0047 1.0780 1.0383
No log 5.625 90 1.2459 -0.0807 1.2459 1.1162
No log 5.75 92 1.1851 -0.1077 1.1851 1.0886
No log 5.875 94 1.0600 0.0365 1.0600 1.0295
No log 6.0 96 1.1574 0.0248 1.1574 1.0758
No log 6.125 98 1.2397 -0.0231 1.2397 1.1134
No log 6.25 100 0.9557 0.0081 0.9557 0.9776
No log 6.375 102 0.8725 0.1133 0.8725 0.9341
No log 6.5 104 0.8579 0.1263 0.8579 0.9262
No log 6.625 106 0.9611 0.0920 0.9611 0.9804
No log 6.75 108 1.3349 -0.0870 1.3349 1.1554
No log 6.875 110 1.3079 -0.0870 1.3079 1.1436
No log 7.0 112 1.2634 -0.0876 1.2634 1.1240
No log 7.125 114 1.1076 -0.1186 1.1076 1.0524
No log 7.25 116 0.9912 0.0109 0.9912 0.9956
No log 7.375 118 0.9867 -0.0184 0.9867 0.9933
No log 7.5 120 1.0583 0.0364 1.0583 1.0288
No log 7.625 122 1.0413 0.1103 1.0413 1.0204
No log 7.75 124 1.0725 0.0566 1.0725 1.0356
No log 7.875 126 0.9960 -0.0087 0.9960 0.9980
No log 8.0 128 0.9838 0.0627 0.9838 0.9918
No log 8.125 130 0.8641 0.0966 0.8641 0.9296
No log 8.25 132 0.8182 0.0816 0.8182 0.9045
No log 8.375 134 0.8659 0.0161 0.8659 0.9305
No log 8.5 136 0.9260 0.0734 0.9260 0.9623
No log 8.625 138 0.8513 0.0087 0.8513 0.9227
No log 8.75 140 0.8293 0.0038 0.8293 0.9106
No log 8.875 142 0.8853 -0.0101 0.8853 0.9409
No log 9.0 144 0.7358 0.1617 0.7358 0.8578
No log 9.125 146 0.7308 0.1395 0.7308 0.8549
No log 9.25 148 0.8639 0.0041 0.8639 0.9295
No log 9.375 150 0.9763 -0.1538 0.9763 0.9881
No log 9.5 152 1.0997 -0.0508 1.0997 1.0487
No log 9.625 154 0.9018 0.1050 0.9018 0.9497
No log 9.75 156 0.8912 0.0559 0.8912 0.9440
No log 9.875 158 0.8816 0.0559 0.8816 0.9389
No log 10.0 160 1.0003 -0.0438 1.0003 1.0002
No log 10.125 162 1.0619 -0.0526 1.0619 1.0305
No log 10.25 164 0.9243 -0.0815 0.9243 0.9614
No log 10.375 166 0.8177 -0.0163 0.8177 0.9043
No log 10.5 168 0.8762 -0.0052 0.8762 0.9360
No log 10.625 170 1.0429 0.0111 1.0429 1.0212
No log 10.75 172 1.1767 -0.0101 1.1767 1.0848
No log 10.875 174 0.9038 -0.0030 0.9038 0.9507
No log 11.0 176 0.8201 0.1408 0.8201 0.9056
No log 11.125 178 0.8365 0.1714 0.8365 0.9146
No log 11.25 180 0.8762 0.0892 0.8762 0.9361
No log 11.375 182 0.9118 0.0821 0.9118 0.9549
No log 11.5 184 0.9671 -0.0409 0.9671 0.9834
No log 11.625 186 0.9448 -0.0395 0.9448 0.9720
No log 11.75 188 0.9407 -0.0151 0.9407 0.9699
No log 11.875 190 0.9873 0.0486 0.9873 0.9936
No log 12.0 192 0.9678 0.0486 0.9678 0.9838
No log 12.125 194 0.8637 0.0684 0.8637 0.9293
No log 12.25 196 0.8284 0.0504 0.8284 0.9101
No log 12.375 198 0.9179 0.0277 0.9179 0.9581
No log 12.5 200 0.8952 -0.0393 0.8952 0.9461
No log 12.625 202 0.8642 -0.0746 0.8642 0.9296
No log 12.75 204 0.9597 0.0556 0.9597 0.9796
No log 12.875 206 1.2330 0.0258 1.2330 1.1104
No log 13.0 208 1.3117 0.0731 1.3117 1.1453
No log 13.125 210 1.0034 -0.0513 1.0034 1.0017
No log 13.25 212 0.7772 0.1734 0.7772 0.8816
No log 13.375 214 0.7596 0.1490 0.7596 0.8716
No log 13.5 216 0.7278 0.0394 0.7278 0.8531
No log 13.625 218 0.9091 0.0515 0.9091 0.9535
No log 13.75 220 1.0566 0.0046 1.0566 1.0279
No log 13.875 222 0.9034 0.0207 0.9034 0.9505
No log 14.0 224 0.7578 0.0513 0.7578 0.8705
No log 14.125 226 0.7480 0.0513 0.7480 0.8649
No log 14.25 228 0.7490 -0.0091 0.7490 0.8655
No log 14.375 230 0.8456 0.1605 0.8456 0.9196
No log 14.5 232 0.8990 0.0233 0.8990 0.9482
No log 14.625 234 0.9793 0.0111 0.9793 0.9896
No log 14.75 236 0.9264 0.0111 0.9264 0.9625
No log 14.875 238 0.7697 0.0123 0.7697 0.8773
No log 15.0 240 0.8007 0.1037 0.8007 0.8948
No log 15.125 242 0.9504 0.0065 0.9504 0.9749
No log 15.25 244 0.9162 0.0365 0.9162 0.9572
No log 15.375 246 0.8801 0.0892 0.8801 0.9381
No log 15.5 248 1.1746 0.0041 1.1746 1.0838
No log 15.625 250 1.5673 0.0585 1.5673 1.2519
No log 15.75 252 1.3754 -0.0114 1.3754 1.1728
No log 15.875 254 0.9309 -0.0138 0.9309 0.9648
No log 16.0 256 0.8474 0.0670 0.8474 0.9206
No log 16.125 258 0.8662 -0.0322 0.8662 0.9307
No log 16.25 260 0.9815 -0.0571 0.9815 0.9907
No log 16.375 262 1.0719 -0.0606 1.0719 1.0353
No log 16.5 264 1.0189 -0.0583 1.0189 1.0094
No log 16.625 266 0.8689 0.0748 0.8689 0.9321
No log 16.75 268 0.7837 0.0338 0.7837 0.8853
No log 16.875 270 0.7813 0.0432 0.7813 0.8839
No log 17.0 272 0.8199 0.0269 0.8199 0.9055
No log 17.125 274 0.9825 -0.0157 0.9825 0.9912
No log 17.25 276 1.1180 -0.0269 1.1180 1.0574
No log 17.375 278 1.0378 0.0515 1.0378 1.0187
No log 17.5 280 0.8909 0.0762 0.8909 0.9439
No log 17.625 282 0.8147 -0.0262 0.8147 0.9026
No log 17.75 284 0.7531 0.0318 0.7531 0.8678
No log 17.875 286 0.7056 0.0479 0.7056 0.8400
No log 18.0 288 0.7054 0.0964 0.7054 0.8399
No log 18.125 290 0.7722 0.1342 0.7722 0.8787
No log 18.25 292 0.8025 0.1342 0.8025 0.8958
No log 18.375 294 0.7728 0.0236 0.7728 0.8791
No log 18.5 296 0.8072 0.0700 0.8072 0.8984
No log 18.625 298 0.8405 0.0660 0.8405 0.9168
No log 18.75 300 0.8871 0.1324 0.8871 0.9418
No log 18.875 302 1.0453 -0.0146 1.0453 1.0224
No log 19.0 304 1.1144 -0.0229 1.1144 1.0556
No log 19.125 306 1.1062 -0.0261 1.1062 1.0518
No log 19.25 308 0.8991 -0.0440 0.8991 0.9482
No log 19.375 310 0.7694 0.0338 0.7694 0.8771
No log 19.5 312 0.7458 0.0375 0.7458 0.8636
No log 19.625 314 0.8329 0.0684 0.8329 0.9126
No log 19.75 316 1.0758 -0.0345 1.0758 1.0372
No log 19.875 318 1.0203 0.0378 1.0203 1.0101
No log 20.0 320 0.8649 0.1360 0.8649 0.9300
No log 20.125 322 0.7682 0.1047 0.7682 0.8765
No log 20.25 324 0.7255 -0.0118 0.7255 0.8517
No log 20.375 326 0.7750 -0.0686 0.7750 0.8803
No log 20.5 328 0.9961 0.0486 0.9961 0.9981
No log 20.625 330 1.1285 0.0305 1.1285 1.0623
No log 20.75 332 0.9833 0.1636 0.9833 0.9916
No log 20.875 334 0.7737 -0.0371 0.7737 0.8796
No log 21.0 336 0.7044 -0.0032 0.7044 0.8393
No log 21.125 338 0.6958 0.0909 0.6958 0.8342
No log 21.25 340 0.7680 0.0909 0.7680 0.8763
No log 21.375 342 0.9068 0.0805 0.9068 0.9523
No log 21.5 344 0.9812 0.0353 0.9812 0.9905
No log 21.625 346 0.9300 0.0431 0.9300 0.9644
No log 21.75 348 0.8213 0.1065 0.8213 0.9062
No log 21.875 350 0.8171 0.1493 0.8171 0.9039
No log 22.0 352 0.8988 -0.0211 0.8988 0.9481
No log 22.125 354 0.9296 -0.0211 0.9296 0.9642
No log 22.25 356 0.8512 0.1354 0.8512 0.9226
No log 22.375 358 0.8227 0.1193 0.8227 0.9070
No log 22.5 360 0.7873 0.1646 0.7873 0.8873
No log 22.625 362 0.7944 0.1646 0.7944 0.8913
No log 22.75 364 0.8278 -0.0118 0.8278 0.9098
No log 22.875 366 0.9046 0.0067 0.9046 0.9511
No log 23.0 368 0.9059 0.0046 0.9059 0.9518
No log 23.125 370 0.8500 -0.0539 0.8500 0.9220
No log 23.25 372 0.8740 -0.0218 0.8740 0.9349
No log 23.375 374 0.8060 0.1147 0.8060 0.8978
No log 23.5 376 0.7626 0.1423 0.7626 0.8733
No log 23.625 378 0.8369 0.0642 0.8369 0.9148
No log 23.75 380 0.9892 0.0046 0.9892 0.9946
No log 23.875 382 1.0495 0.0026 1.0495 1.0245
No log 24.0 384 0.9970 0.0067 0.9970 0.9985
No log 24.125 386 0.8358 0.1553 0.8358 0.9142
No log 24.25 388 0.7711 0.0639 0.7711 0.8781
No log 24.375 390 0.7983 0.0909 0.7983 0.8935
No log 24.5 392 0.7975 0.1395 0.7975 0.8930
No log 24.625 394 0.7466 0.0318 0.7466 0.8641
No log 24.75 396 0.7427 0.0807 0.7427 0.8618
No log 24.875 398 0.7564 0.1691 0.7564 0.8697
No log 25.0 400 0.7788 0.1506 0.7788 0.8825
No log 25.125 402 0.7444 0.2105 0.7444 0.8628
No log 25.25 404 0.7394 0.2105 0.7394 0.8599
No log 25.375 406 0.7250 0.2180 0.7250 0.8515
No log 25.5 408 0.7459 0.2034 0.7459 0.8637
No log 25.625 410 0.7867 0.1291 0.7867 0.8870
No log 25.75 412 0.7486 0.1259 0.7486 0.8652
No log 25.875 414 0.7556 0.0768 0.7556 0.8693
No log 26.0 416 0.7515 0.0338 0.7515 0.8669
No log 26.125 418 0.7908 -0.0218 0.7908 0.8893
No log 26.25 420 0.8867 -0.0526 0.8867 0.9417
No log 26.375 422 1.0032 -0.0013 1.0032 1.0016
No log 26.5 424 0.9503 0.0006 0.9503 0.9748
No log 26.625 426 0.8342 0.0277 0.8342 0.9134
No log 26.75 428 0.7721 -0.0195 0.7721 0.8787
No log 26.875 430 0.7663 -0.0195 0.7663 0.8754
No log 27.0 432 0.8034 0.1196 0.8034 0.8963
No log 27.125 434 0.8955 -0.0301 0.8955 0.9463
No log 27.25 436 0.9592 -0.0398 0.9592 0.9794
No log 27.375 438 0.9387 -0.0031 0.9387 0.9689
No log 27.5 440 0.8197 0.1196 0.8197 0.9053
No log 27.625 442 0.7342 0.1202 0.7342 0.8569
No log 27.75 444 0.7251 0.0355 0.7251 0.8515
No log 27.875 446 0.7518 0.0214 0.7518 0.8671
No log 28.0 448 0.8597 0.0642 0.8597 0.9272
No log 28.125 450 1.0421 0.0404 1.0421 1.0208
No log 28.25 452 1.1349 -0.0117 1.1349 1.0653
No log 28.375 454 1.0388 0.0260 1.0388 1.0192
No log 28.5 456 0.8613 0.0486 0.8613 0.9281
No log 28.625 458 0.7198 0.1097 0.7198 0.8484
No log 28.75 460 0.7064 0.0432 0.7064 0.8405
No log 28.875 462 0.7098 0.0395 0.7098 0.8425
No log 29.0 464 0.7472 0.1506 0.7472 0.8644
No log 29.125 466 0.8965 0.0431 0.8965 0.9468
No log 29.25 468 1.0091 -0.0385 1.0091 1.0045
No log 29.375 470 0.9948 -0.0031 0.9948 0.9974
No log 29.5 472 0.8862 -0.0218 0.8862 0.9414
No log 29.625 474 0.7633 0.1449 0.7633 0.8736
No log 29.75 476 0.7227 0.0338 0.7227 0.8501
No log 29.875 478 0.7294 0.0338 0.7294 0.8540
No log 30.0 480 0.7733 0.2248 0.7733 0.8794
No log 30.125 482 0.8493 0.1107 0.8493 0.9216
No log 30.25 484 0.9412 0.0089 0.9412 0.9702
No log 30.375 486 1.0689 0.0156 1.0689 1.0339
No log 30.5 488 1.0219 0.0156 1.0219 1.0109
No log 30.625 490 0.8974 0.0067 0.8974 0.9473
No log 30.75 492 0.8076 0.1553 0.8076 0.8987
No log 30.875 494 0.7758 0.2277 0.7758 0.8808
No log 31.0 496 0.7903 0.2494 0.7903 0.8890
No log 31.125 498 0.7785 0.2566 0.7785 0.8823
0.271 31.25 500 0.8149 0.0609 0.8149 0.9027
0.271 31.375 502 0.9377 0.0046 0.9377 0.9683
0.271 31.5 504 1.0442 0.0790 1.0442 1.0219
0.271 31.625 506 1.0050 0.0238 1.0050 1.0025
0.271 31.75 508 0.8862 0.0458 0.8862 0.9414
0.271 31.875 510 0.8099 0.1453 0.8099 0.9000
0.271 32.0 512 0.7688 0.0871 0.7688 0.8768
0.271 32.125 514 0.7977 0.1955 0.7977 0.8932
0.271 32.25 516 0.8676 0.0909 0.8676 0.9314
0.271 32.375 518 0.8960 0.0134 0.8960 0.9466
0.271 32.5 520 0.8562 0.0909 0.8562 0.9253
0.271 32.625 522 0.7814 0.1150 0.7814 0.8840
0.271 32.75 524 0.7153 0.2502 0.7153 0.8458
0.271 32.875 526 0.6857 0.1199 0.6857 0.8281
0.271 33.0 528 0.6875 0.1199 0.6875 0.8291
0.271 33.125 530 0.7084 0.2502 0.7084 0.8416
0.271 33.25 532 0.7981 0.1064 0.7981 0.8934
0.271 33.375 534 0.9299 0.0404 0.9299 0.9643
0.271 33.5 536 0.9380 0.0710 0.9380 0.9685
0.271 33.625 538 0.8306 -0.0163 0.8306 0.9114
0.271 33.75 540 0.7759 0.1994 0.7759 0.8809
0.271 33.875 542 0.7442 0.1336 0.7442 0.8627
0.271 34.0 544 0.7758 0.2054 0.7758 0.8808
0.271 34.125 546 0.8620 0.0157 0.8620 0.9284
0.271 34.25 548 0.9440 0.0353 0.9440 0.9716
0.271 34.375 550 0.9665 0.0260 0.9665 0.9831
0.271 34.5 552 0.8583 0.0182 0.8583 0.9265
0.271 34.625 554 0.7356 0.2248 0.7356 0.8577
0.271 34.75 556 0.7162 0.1612 0.7162 0.8463
0.271 34.875 558 0.7299 0.1096 0.7299 0.8543
0.271 35.0 560 0.7781 0.1701 0.7781 0.8821
0.271 35.125 562 0.8800 0.0545 0.8800 0.9381
0.271 35.25 564 0.9924 -0.0049 0.9924 0.9962
0.271 35.375 566 1.0116 -0.0067 1.0116 1.0058
0.271 35.5 568 1.0017 -0.0049 1.0017 1.0009
0.271 35.625 570 0.9105 0.0545 0.9105 0.9542

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
9
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k5_task3_organization

Finetuned
(4222)
this model