ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run3_AugV5_k2_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.2322
  • Qwk: 0.5774
  • Mse: 1.2322
  • Rmse: 1.1101

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1667 2 5.1770 -0.0093 5.1770 2.2753
No log 0.3333 4 3.4441 0.0521 3.4441 1.8558
No log 0.5 6 1.9299 0.0992 1.9299 1.3892
No log 0.6667 8 1.4513 0.1304 1.4513 1.2047
No log 0.8333 10 1.2487 0.1488 1.2487 1.1174
No log 1.0 12 1.3748 0.1665 1.3748 1.1725
No log 1.1667 14 1.2829 0.1631 1.2829 1.1327
No log 1.3333 16 1.3418 0.2093 1.3418 1.1584
No log 1.5 18 1.3157 0.2234 1.3157 1.1470
No log 1.6667 20 1.1889 0.2680 1.1889 1.0904
No log 1.8333 22 1.1496 0.3116 1.1496 1.0722
No log 2.0 24 1.2826 0.3708 1.2826 1.1325
No log 2.1667 26 1.3018 0.3619 1.3018 1.1410
No log 2.3333 28 0.9937 0.4299 0.9937 0.9968
No log 2.5 30 1.0967 0.3272 1.0967 1.0473
No log 2.6667 32 1.1307 0.2221 1.1307 1.0634
No log 2.8333 34 0.9950 0.3319 0.9950 0.9975
No log 3.0 36 0.9568 0.4369 0.9568 0.9781
No log 3.1667 38 1.1208 0.3471 1.1208 1.0587
No log 3.3333 40 1.0823 0.3840 1.0823 1.0403
No log 3.5 42 0.9649 0.4104 0.9649 0.9823
No log 3.6667 44 1.0424 0.5100 1.0424 1.0210
No log 3.8333 46 1.1006 0.4872 1.1006 1.0491
No log 4.0 48 1.1827 0.3990 1.1827 1.0875
No log 4.1667 50 1.1390 0.4375 1.1390 1.0672
No log 4.3333 52 1.1044 0.4577 1.1044 1.0509
No log 4.5 54 1.0147 0.4815 1.0147 1.0073
No log 4.6667 56 0.9373 0.4903 0.9373 0.9682
No log 4.8333 58 0.9511 0.5210 0.9511 0.9752
No log 5.0 60 1.4790 0.3817 1.4790 1.2161
No log 5.1667 62 1.5638 0.3515 1.5638 1.2505
No log 5.3333 64 1.0427 0.4794 1.0427 1.0211
No log 5.5 66 0.9641 0.5191 0.9641 0.9819
No log 5.6667 68 0.9344 0.5667 0.9344 0.9667
No log 5.8333 70 0.8307 0.6297 0.8307 0.9114
No log 6.0 72 0.8769 0.5763 0.8769 0.9364
No log 6.1667 74 1.0872 0.5707 1.0872 1.0427
No log 6.3333 76 1.0272 0.5884 1.0272 1.0135
No log 6.5 78 0.8515 0.6310 0.8515 0.9228
No log 6.6667 80 0.8062 0.6538 0.8062 0.8979
No log 6.8333 82 0.8176 0.6317 0.8176 0.9042
No log 7.0 84 0.8765 0.6124 0.8765 0.9362
No log 7.1667 86 1.0964 0.5020 1.0964 1.0471
No log 7.3333 88 1.0907 0.4992 1.0907 1.0444
No log 7.5 90 0.7939 0.6375 0.7939 0.8910
No log 7.6667 92 0.7630 0.6432 0.7630 0.8735
No log 7.8333 94 0.7442 0.6406 0.7442 0.8627
No log 8.0 96 0.7392 0.6591 0.7392 0.8598
No log 8.1667 98 0.8027 0.6326 0.8027 0.8959
No log 8.3333 100 0.9051 0.5994 0.9051 0.9514
No log 8.5 102 0.7876 0.6410 0.7876 0.8875
No log 8.6667 104 0.7491 0.6626 0.7491 0.8655
No log 8.8333 106 0.7313 0.6671 0.7313 0.8552
No log 9.0 108 0.7256 0.6752 0.7256 0.8518
No log 9.1667 110 0.7157 0.6990 0.7157 0.8460
No log 9.3333 112 0.8318 0.6445 0.8318 0.9120
No log 9.5 114 0.8468 0.6508 0.8468 0.9202
No log 9.6667 116 0.6952 0.6555 0.6952 0.8338
No log 9.8333 118 0.7303 0.6239 0.7303 0.8546
No log 10.0 120 0.7271 0.6423 0.7271 0.8527
No log 10.1667 122 0.7930 0.6818 0.7930 0.8905
No log 10.3333 124 1.1457 0.5870 1.1457 1.0704
No log 10.5 126 1.3329 0.4918 1.3329 1.1545
No log 10.6667 128 1.1208 0.5617 1.1208 1.0587
No log 10.8333 130 0.7552 0.6825 0.7552 0.8690
No log 11.0 132 0.7163 0.6398 0.7163 0.8463
No log 11.1667 134 0.7482 0.6297 0.7482 0.8650
No log 11.3333 136 0.7014 0.6560 0.7014 0.8375
No log 11.5 138 0.9502 0.5818 0.9502 0.9748
No log 11.6667 140 1.1288 0.5029 1.1288 1.0624
No log 11.8333 142 0.9681 0.5475 0.9681 0.9839
No log 12.0 144 0.7567 0.5524 0.7567 0.8699
No log 12.1667 146 0.7337 0.6124 0.7337 0.8566
No log 12.3333 148 0.7147 0.6569 0.7147 0.8454
No log 12.5 150 0.8005 0.6656 0.8005 0.8947
No log 12.6667 152 0.7639 0.6448 0.7639 0.8740
No log 12.8333 154 0.7777 0.6636 0.7777 0.8819
No log 13.0 156 0.8592 0.6564 0.8592 0.9269
No log 13.1667 158 0.8973 0.6648 0.8973 0.9472
No log 13.3333 160 0.8134 0.6504 0.8134 0.9019
No log 13.5 162 0.7184 0.6201 0.7184 0.8476
No log 13.6667 164 0.6950 0.6426 0.6950 0.8337
No log 13.8333 166 0.7226 0.6359 0.7226 0.8501
No log 14.0 168 0.7968 0.6381 0.7968 0.8927
No log 14.1667 170 0.7501 0.6562 0.7501 0.8661
No log 14.3333 172 0.7534 0.6317 0.7534 0.8680
No log 14.5 174 0.8277 0.6002 0.8277 0.9098
No log 14.6667 176 0.8562 0.6029 0.8562 0.9253
No log 14.8333 178 0.7900 0.6227 0.7900 0.8888
No log 15.0 180 0.8210 0.5821 0.8210 0.9061
No log 15.1667 182 0.7877 0.6127 0.7877 0.8875
No log 15.3333 184 0.7352 0.6176 0.7352 0.8575
No log 15.5 186 0.7195 0.6146 0.7195 0.8483
No log 15.6667 188 0.7849 0.6526 0.7849 0.8859
No log 15.8333 190 0.9012 0.6360 0.9012 0.9493
No log 16.0 192 0.8825 0.6556 0.8825 0.9394
No log 16.1667 194 0.8373 0.6444 0.8373 0.9150
No log 16.3333 196 0.8205 0.6453 0.8205 0.9058
No log 16.5 198 0.9014 0.6351 0.9014 0.9494
No log 16.6667 200 1.0800 0.6026 1.0800 1.0392
No log 16.8333 202 1.0161 0.5967 1.0161 1.0080
No log 17.0 204 0.8218 0.6017 0.8218 0.9065
No log 17.1667 206 0.7148 0.5752 0.7148 0.8455
No log 17.3333 208 0.7070 0.5861 0.7070 0.8408
No log 17.5 210 0.7325 0.6127 0.7325 0.8559
No log 17.6667 212 0.8973 0.6167 0.8973 0.9473
No log 17.8333 214 1.0769 0.5953 1.0769 1.0377
No log 18.0 216 1.0215 0.5972 1.0215 1.0107
No log 18.1667 218 0.8159 0.6351 0.8159 0.9033
No log 18.3333 220 0.7078 0.5970 0.7078 0.8413
No log 18.5 222 0.7507 0.6114 0.7507 0.8664
No log 18.6667 224 0.7014 0.6431 0.7014 0.8375
No log 18.8333 226 0.7829 0.6684 0.7829 0.8848
No log 19.0 228 1.2211 0.5586 1.2211 1.1050
No log 19.1667 230 1.4890 0.4963 1.4890 1.2203
No log 19.3333 232 1.3057 0.5411 1.3057 1.1427
No log 19.5 234 0.9265 0.648 0.9265 0.9625
No log 19.6667 236 0.7299 0.6691 0.7299 0.8544
No log 19.8333 238 0.7115 0.6542 0.7115 0.8435
No log 20.0 240 0.7283 0.6731 0.7283 0.8534
No log 20.1667 242 0.8032 0.6742 0.8032 0.8962
No log 20.3333 244 0.8344 0.7008 0.8344 0.9134
No log 20.5 246 0.7581 0.6985 0.7581 0.8707
No log 20.6667 248 0.7211 0.6903 0.7211 0.8492
No log 20.8333 250 0.7397 0.6528 0.7397 0.8601
No log 21.0 252 0.7688 0.6419 0.7688 0.8768
No log 21.1667 254 0.7668 0.6249 0.7668 0.8756
No log 21.3333 256 0.7501 0.6224 0.7501 0.8661
No log 21.5 258 0.7629 0.6352 0.7629 0.8735
No log 21.6667 260 0.8111 0.6475 0.8111 0.9006
No log 21.8333 262 0.8077 0.6570 0.8077 0.8987
No log 22.0 264 0.7757 0.6606 0.7757 0.8807
No log 22.1667 266 0.8636 0.6637 0.8636 0.9293
No log 22.3333 268 0.9557 0.6686 0.9557 0.9776
No log 22.5 270 0.9315 0.6517 0.9315 0.9652
No log 22.6667 272 0.9106 0.6268 0.9106 0.9543
No log 22.8333 274 0.8221 0.6366 0.8221 0.9067
No log 23.0 276 0.7651 0.6431 0.7651 0.8747
No log 23.1667 278 0.7526 0.6514 0.7526 0.8675
No log 23.3333 280 0.7784 0.6557 0.7784 0.8823
No log 23.5 282 0.8640 0.6396 0.8640 0.9295
No log 23.6667 284 0.8859 0.6486 0.8859 0.9412
No log 23.8333 286 0.8251 0.6396 0.8251 0.9084
No log 24.0 288 0.8425 0.6393 0.8425 0.9179
No log 24.1667 290 0.9056 0.6419 0.9056 0.9516
No log 24.3333 292 0.8323 0.6273 0.8323 0.9123
No log 24.5 294 0.8018 0.6392 0.8018 0.8954
No log 24.6667 296 0.8282 0.6233 0.8282 0.9100
No log 24.8333 298 0.8217 0.6486 0.8217 0.9065
No log 25.0 300 0.8609 0.6442 0.8609 0.9279
No log 25.1667 302 0.9537 0.6143 0.9537 0.9766
No log 25.3333 304 0.8545 0.6442 0.8545 0.9244
No log 25.5 306 0.7813 0.6479 0.7813 0.8839
No log 25.6667 308 0.7496 0.6432 0.7496 0.8658
No log 25.8333 310 0.7254 0.6362 0.7254 0.8517
No log 26.0 312 0.7936 0.6407 0.7936 0.8908
No log 26.1667 314 0.9139 0.5988 0.9139 0.9560
No log 26.3333 316 0.8838 0.6337 0.8838 0.9401
No log 26.5 318 0.7670 0.6517 0.7670 0.8758
No log 26.6667 320 0.7416 0.6567 0.7416 0.8612
No log 26.8333 322 0.7671 0.6636 0.7671 0.8758
No log 27.0 324 0.9179 0.6367 0.9179 0.9581
No log 27.1667 326 0.9995 0.6086 0.9995 0.9997
No log 27.3333 328 0.9356 0.6203 0.9356 0.9672
No log 27.5 330 0.8945 0.6438 0.8945 0.9458
No log 27.6667 332 0.7788 0.6645 0.7788 0.8825
No log 27.8333 334 0.7337 0.6356 0.7337 0.8565
No log 28.0 336 0.7443 0.6262 0.7443 0.8627
No log 28.1667 338 0.8254 0.6511 0.8254 0.9085
No log 28.3333 340 0.9640 0.6093 0.9640 0.9819
No log 28.5 342 0.9721 0.6011 0.9721 0.9860
No log 28.6667 344 0.8353 0.6638 0.8353 0.9139
No log 28.8333 346 0.7628 0.6493 0.7628 0.8734
No log 29.0 348 0.7456 0.6700 0.7456 0.8635
No log 29.1667 350 0.7705 0.6700 0.7705 0.8778
No log 29.3333 352 0.8748 0.6620 0.8748 0.9353
No log 29.5 354 1.0122 0.6222 1.0122 1.0061
No log 29.6667 356 0.9348 0.6492 0.9348 0.9669
No log 29.8333 358 0.7874 0.6590 0.7874 0.8874
No log 30.0 360 0.7374 0.6399 0.7374 0.8587
No log 30.1667 362 0.7610 0.6567 0.7610 0.8724
No log 30.3333 364 0.8762 0.6415 0.8762 0.9360
No log 30.5 366 0.9108 0.6198 0.9108 0.9544
No log 30.6667 368 0.8642 0.6410 0.8642 0.9296
No log 30.8333 370 0.8402 0.6410 0.8402 0.9166
No log 31.0 372 0.8755 0.6393 0.8755 0.9357
No log 31.1667 374 0.8285 0.6561 0.8285 0.9102
No log 31.3333 376 0.8397 0.6561 0.8397 0.9164
No log 31.5 378 0.9151 0.6320 0.9151 0.9566
No log 31.6667 380 0.8990 0.6360 0.8990 0.9482
No log 31.8333 382 0.8182 0.6609 0.8182 0.9045
No log 32.0 384 0.8631 0.6511 0.8631 0.9290
No log 32.1667 386 0.9332 0.6438 0.9332 0.9660
No log 32.3333 388 1.0336 0.6357 1.0336 1.0167
No log 32.5 390 1.2033 0.5591 1.2033 1.0970
No log 32.6667 392 1.1941 0.5660 1.1941 1.0927
No log 32.8333 394 1.0611 0.6096 1.0611 1.0301
No log 33.0 396 0.9100 0.6136 0.9100 0.9539
No log 33.1667 398 0.8516 0.6292 0.8516 0.9228
No log 33.3333 400 0.7857 0.6357 0.7857 0.8864
No log 33.5 402 0.7607 0.6455 0.7607 0.8722
No log 33.6667 404 0.7659 0.6576 0.7659 0.8752
No log 33.8333 406 0.8066 0.6778 0.8066 0.8981
No log 34.0 408 0.8325 0.6854 0.8325 0.9124
No log 34.1667 410 0.7670 0.6828 0.7670 0.8758
No log 34.3333 412 0.7520 0.7005 0.7520 0.8672
No log 34.5 414 0.7342 0.7037 0.7342 0.8569
No log 34.6667 416 0.7658 0.6874 0.7658 0.8751
No log 34.8333 418 0.8302 0.6853 0.8302 0.9112
No log 35.0 420 0.9670 0.6252 0.9670 0.9834
No log 35.1667 422 1.0155 0.6090 1.0155 1.0077
No log 35.3333 424 0.9033 0.6444 0.9033 0.9504
No log 35.5 426 0.8376 0.6476 0.8376 0.9152
No log 35.6667 428 0.7639 0.6910 0.7639 0.8740
No log 35.8333 430 0.7063 0.6596 0.7063 0.8404
No log 36.0 432 0.7169 0.6814 0.7169 0.8467
No log 36.1667 434 0.7803 0.6683 0.7803 0.8833
No log 36.3333 436 0.9561 0.6407 0.9561 0.9778
No log 36.5 438 1.0185 0.6140 1.0185 1.0092
No log 36.6667 440 0.9695 0.6498 0.9695 0.9846
No log 36.8333 442 0.8734 0.6771 0.8734 0.9345
No log 37.0 444 0.7610 0.6791 0.7610 0.8723
No log 37.1667 446 0.7155 0.6993 0.7155 0.8458
No log 37.3333 448 0.6982 0.6989 0.6982 0.8356
No log 37.5 450 0.6979 0.6925 0.6979 0.8354
No log 37.6667 452 0.7702 0.6872 0.7702 0.8776
No log 37.8333 454 0.9205 0.6325 0.9205 0.9594
No log 38.0 456 1.0027 0.6141 1.0027 1.0013
No log 38.1667 458 0.9260 0.6686 0.9260 0.9623
No log 38.3333 460 0.8329 0.6666 0.8329 0.9126
No log 38.5 462 0.7393 0.6981 0.7393 0.8598
No log 38.6667 464 0.7168 0.6652 0.7168 0.8466
No log 38.8333 466 0.7202 0.6633 0.7202 0.8486
No log 39.0 468 0.7268 0.6774 0.7268 0.8525
No log 39.1667 470 0.7383 0.6803 0.7383 0.8593
No log 39.3333 472 0.7693 0.6862 0.7693 0.8771
No log 39.5 474 0.8764 0.6555 0.8764 0.9361
No log 39.6667 476 0.9390 0.6568 0.9390 0.9690
No log 39.8333 478 0.9624 0.6505 0.9624 0.9810
No log 40.0 480 1.0119 0.6247 1.0119 1.0059
No log 40.1667 482 0.9411 0.6522 0.9411 0.9701
No log 40.3333 484 0.8371 0.6714 0.8371 0.9149
No log 40.5 486 0.7937 0.6848 0.7937 0.8909
No log 40.6667 488 0.7352 0.7183 0.7352 0.8574
No log 40.8333 490 0.7248 0.6967 0.7248 0.8514
No log 41.0 492 0.8125 0.6667 0.8125 0.9014
No log 41.1667 494 0.9792 0.6430 0.9792 0.9896
No log 41.3333 496 1.0686 0.5945 1.0686 1.0337
No log 41.5 498 1.0098 0.6274 1.0098 1.0049
0.3368 41.6667 500 0.9036 0.6523 0.9036 0.9506
0.3368 41.8333 502 0.8122 0.6542 0.8122 0.9012
0.3368 42.0 504 0.8294 0.6534 0.8294 0.9107
0.3368 42.1667 506 0.8724 0.6601 0.8724 0.9340
0.3368 42.3333 508 0.9899 0.6315 0.9899 0.9949
0.3368 42.5 510 1.0085 0.6159 1.0085 1.0042
0.3368 42.6667 512 0.9027 0.6667 0.9027 0.9501
0.3368 42.8333 514 0.8149 0.6647 0.8149 0.9027
0.3368 43.0 516 0.7466 0.6933 0.7466 0.8640
0.3368 43.1667 518 0.7514 0.6933 0.7514 0.8668
0.3368 43.3333 520 0.7599 0.7073 0.7599 0.8717
0.3368 43.5 522 0.7667 0.6790 0.7667 0.8756
0.3368 43.6667 524 0.7772 0.6922 0.7772 0.8816
0.3368 43.8333 526 0.8380 0.6648 0.8380 0.9154
0.3368 44.0 528 0.8624 0.6638 0.8624 0.9287
0.3368 44.1667 530 0.7941 0.6550 0.7941 0.8911
0.3368 44.3333 532 0.7631 0.6645 0.7631 0.8735
0.3368 44.5 534 0.7861 0.6714 0.7861 0.8866
0.3368 44.6667 536 0.8524 0.6761 0.8524 0.9232
0.3368 44.8333 538 1.0248 0.6281 1.0248 1.0123
0.3368 45.0 540 1.1723 0.5839 1.1723 1.0827
0.3368 45.1667 542 1.1665 0.5946 1.1665 1.0801
0.3368 45.3333 544 1.0576 0.6159 1.0576 1.0284
0.3368 45.5 546 0.9093 0.6530 0.9093 0.9536
0.3368 45.6667 548 0.8250 0.6845 0.8250 0.9083
0.3368 45.8333 550 0.7732 0.6908 0.7732 0.8793
0.3368 46.0 552 0.7518 0.6948 0.7518 0.8670
0.3368 46.1667 554 0.7637 0.6948 0.7637 0.8739
0.3368 46.3333 556 0.7396 0.7051 0.7396 0.8600
0.3368 46.5 558 0.7596 0.7051 0.7596 0.8716
0.3368 46.6667 560 0.7330 0.6989 0.7330 0.8561
0.3368 46.8333 562 0.7142 0.6989 0.7142 0.8451
0.3368 47.0 564 0.6808 0.6681 0.6808 0.8251
0.3368 47.1667 566 0.6848 0.6565 0.6848 0.8275
0.3368 47.3333 568 0.7022 0.6605 0.7022 0.8380
0.3368 47.5 570 0.7311 0.6849 0.7311 0.8550
0.3368 47.6667 572 0.7913 0.6735 0.7913 0.8895
0.3368 47.8333 574 0.8931 0.6356 0.8931 0.9451
0.3368 48.0 576 0.9744 0.6423 0.9744 0.9871
0.3368 48.1667 578 0.9903 0.6423 0.9903 0.9951
0.3368 48.3333 580 0.9509 0.6472 0.9509 0.9751
0.3368 48.5 582 0.8937 0.6593 0.8937 0.9454
0.3368 48.6667 584 0.8391 0.6593 0.8391 0.9160
0.3368 48.8333 586 0.7856 0.6543 0.7856 0.8863
0.3368 49.0 588 0.7661 0.6654 0.7661 0.8753
0.3368 49.1667 590 0.7728 0.6765 0.7728 0.8791
0.3368 49.3333 592 0.8256 0.6520 0.8256 0.9086
0.3368 49.5 594 0.9562 0.6481 0.9562 0.9779
0.3368 49.6667 596 1.1159 0.5938 1.1159 1.0564
0.3368 49.8333 598 1.2273 0.5680 1.2273 1.1079
0.3368 50.0 600 1.2322 0.5774 1.2322 1.1101

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
6
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run3_AugV5_k2_task1_organization

Finetuned
(4222)
this model