ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k12_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6998
  • Qwk: 0.7673
  • Mse: 0.6998
  • Rmse: 0.8366

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0351 2 6.5625 0.0176 6.5625 2.5617
No log 0.0702 4 4.3884 0.0365 4.3884 2.0949
No log 0.1053 6 3.3757 -0.0104 3.3757 1.8373
No log 0.1404 8 2.3341 0.1642 2.3341 1.5278
No log 0.1754 10 1.6780 0.2609 1.6780 1.2954
No log 0.2105 12 1.6113 0.3415 1.6113 1.2694
No log 0.2456 14 1.4977 0.2881 1.4977 1.2238
No log 0.2807 16 1.4148 0.3186 1.4148 1.1895
No log 0.3158 18 1.4438 0.2857 1.4438 1.2016
No log 0.3509 20 1.6134 0.2712 1.6134 1.2702
No log 0.3860 22 2.0392 0.2290 2.0392 1.4280
No log 0.4211 24 2.2438 0.1277 2.2438 1.4979
No log 0.4561 26 2.0222 0.2535 2.0222 1.4221
No log 0.4912 28 1.4910 0.3770 1.4910 1.2211
No log 0.5263 30 1.2963 0.4375 1.2963 1.1385
No log 0.5614 32 1.4446 0.4348 1.4446 1.2019
No log 0.5965 34 1.9517 0.3602 1.9517 1.3970
No log 0.6316 36 1.8664 0.4156 1.8664 1.3661
No log 0.6667 38 1.9441 0.3396 1.9441 1.3943
No log 0.7018 40 2.2716 0.2703 2.2716 1.5072
No log 0.7368 42 2.1771 0.2967 2.1771 1.4755
No log 0.7719 44 2.0757 0.3523 2.0757 1.4407
No log 0.8070 46 1.6857 0.4908 1.6857 1.2984
No log 0.8421 48 1.1777 0.6087 1.1777 1.0852
No log 0.8772 50 1.2544 0.625 1.2544 1.1200
No log 0.9123 52 1.2065 0.6196 1.2065 1.0984
No log 0.9474 54 0.9310 0.7368 0.9310 0.9649
No log 0.9825 56 0.9072 0.6832 0.9072 0.9525
No log 1.0175 58 1.0462 0.6144 1.0462 1.0228
No log 1.0526 60 0.9409 0.6667 0.9409 0.9700
No log 1.0877 62 0.8808 0.7 0.8808 0.9385
No log 1.1228 64 0.7836 0.7389 0.7836 0.8852
No log 1.1579 66 0.7350 0.7211 0.7350 0.8573
No log 1.1930 68 0.7230 0.7211 0.7230 0.8503
No log 1.2281 70 1.0334 0.6879 1.0334 1.0166
No log 1.2632 72 1.2971 0.5679 1.2971 1.1389
No log 1.2982 74 1.3005 0.5823 1.3005 1.1404
No log 1.3333 76 0.9073 0.6711 0.9073 0.9525
No log 1.3684 78 0.8090 0.7376 0.8090 0.8995
No log 1.4035 80 0.9917 0.5954 0.9917 0.9959
No log 1.4386 82 0.9844 0.6383 0.9844 0.9922
No log 1.4737 84 1.1188 0.6056 1.1188 1.0577
No log 1.5088 86 1.2515 0.5556 1.2515 1.1187
No log 1.5439 88 1.1825 0.5765 1.1825 1.0874
No log 1.5789 90 1.7593 0.5572 1.7593 1.3264
No log 1.6140 92 2.0835 0.4977 2.0835 1.4434
No log 1.6491 94 1.5068 0.6 1.5068 1.2275
No log 1.6842 96 0.8941 0.7219 0.8941 0.9456
No log 1.7193 98 0.8405 0.7261 0.8405 0.9168
No log 1.7544 100 0.7725 0.7320 0.7725 0.8789
No log 1.7895 102 0.8883 0.6667 0.8883 0.9425
No log 1.8246 104 1.0062 0.6203 1.0062 1.0031
No log 1.8596 106 1.0456 0.6194 1.0456 1.0225
No log 1.8947 108 0.8241 0.6622 0.8241 0.9078
No log 1.9298 110 0.6938 0.7586 0.6938 0.8330
No log 1.9649 112 0.7265 0.7361 0.7265 0.8524
No log 2.0 114 0.7167 0.7703 0.7167 0.8466
No log 2.0351 116 0.7528 0.7381 0.7528 0.8677
No log 2.0702 118 1.1389 0.6404 1.1389 1.0672
No log 2.1053 120 1.3427 0.5889 1.3427 1.1587
No log 2.1404 122 1.0594 0.6316 1.0594 1.0293
No log 2.1754 124 0.7863 0.7643 0.7863 0.8867
No log 2.2105 126 0.8462 0.7162 0.8462 0.9199
No log 2.2456 128 0.8647 0.6846 0.8647 0.9299
No log 2.2807 130 0.8147 0.7516 0.8147 0.9026
No log 2.3158 132 0.8743 0.7143 0.8743 0.9351
No log 2.3509 134 1.0170 0.6404 1.0170 1.0085
No log 2.3860 136 0.9496 0.6919 0.9496 0.9745
No log 2.4211 138 0.7594 0.7886 0.7594 0.8714
No log 2.4561 140 0.6940 0.7545 0.6940 0.8330
No log 2.4912 142 0.7558 0.7059 0.7558 0.8694
No log 2.5263 144 0.7168 0.7067 0.7168 0.8466
No log 2.5614 146 0.6771 0.7682 0.6771 0.8229
No log 2.5965 148 0.6758 0.76 0.6758 0.8221
No log 2.6316 150 0.7098 0.7703 0.7098 0.8425
No log 2.6667 152 0.7199 0.7534 0.7199 0.8485
No log 2.7018 154 0.6949 0.7651 0.6949 0.8336
No log 2.7368 156 0.7448 0.7651 0.7448 0.8630
No log 2.7719 158 0.7117 0.7733 0.7117 0.8437
No log 2.8070 160 0.5989 0.7733 0.5989 0.7739
No log 2.8421 162 0.6177 0.8132 0.6177 0.7859
No log 2.8772 164 0.6793 0.8 0.6793 0.8242
No log 2.9123 166 0.6272 0.8087 0.6272 0.7920
No log 2.9474 168 0.6331 0.7927 0.6331 0.7957
No log 2.9825 170 0.8834 0.6928 0.8834 0.9399
No log 3.0175 172 0.9156 0.6533 0.9156 0.9569
No log 3.0526 174 0.7389 0.7020 0.7389 0.8596
No log 3.0877 176 0.5771 0.7875 0.5771 0.7597
No log 3.1228 178 0.6446 0.8362 0.6446 0.8029
No log 3.1579 180 0.6625 0.8070 0.6625 0.8139
No log 3.1930 182 0.7043 0.8 0.7043 0.8392
No log 3.2281 184 0.6803 0.7871 0.6803 0.8248
No log 3.2632 186 0.7071 0.7815 0.7071 0.8409
No log 3.2982 188 0.7253 0.7733 0.7253 0.8516
No log 3.3333 190 0.6941 0.7733 0.6941 0.8332
No log 3.3684 192 0.6640 0.7815 0.6640 0.8149
No log 3.4035 194 0.6532 0.7632 0.6532 0.8082
No log 3.4386 196 0.6668 0.7815 0.6668 0.8166
No log 3.4737 198 0.7193 0.7671 0.7193 0.8481
No log 3.5088 200 0.7688 0.7183 0.7688 0.8768
No log 3.5439 202 0.7855 0.7183 0.7855 0.8863
No log 3.5789 204 0.7424 0.7183 0.7424 0.8616
No log 3.6140 206 0.6929 0.7632 0.6929 0.8324
No log 3.6491 208 0.7495 0.7701 0.7495 0.8658
No log 3.6842 210 0.6681 0.8023 0.6681 0.8173
No log 3.7193 212 0.6232 0.7848 0.6232 0.7894
No log 3.7544 214 0.6545 0.7516 0.6545 0.8090
No log 3.7895 216 0.6338 0.7662 0.6338 0.7961
No log 3.8246 218 0.6814 0.7895 0.6814 0.8255
No log 3.8596 220 0.6966 0.7703 0.6966 0.8346
No log 3.8947 222 0.6837 0.7785 0.6837 0.8269
No log 3.9298 224 0.6744 0.7703 0.6744 0.8212
No log 3.9649 226 0.6876 0.8050 0.6876 0.8292
No log 4.0 228 0.7958 0.7574 0.7958 0.8921
No log 4.0351 230 0.7685 0.7977 0.7685 0.8766
No log 4.0702 232 0.6955 0.8047 0.6955 0.8339
No log 4.1053 234 0.6580 0.7952 0.6580 0.8112
No log 4.1404 236 0.6713 0.8 0.6713 0.8193
No log 4.1754 238 0.7068 0.8249 0.7068 0.8407
No log 4.2105 240 0.8478 0.7742 0.8478 0.9208
No log 4.2456 242 0.8615 0.75 0.8615 0.9282
No log 4.2807 244 0.7116 0.8 0.7116 0.8436
No log 4.3158 246 0.6356 0.7925 0.6356 0.7973
No log 4.3509 248 0.6635 0.7949 0.6635 0.8145
No log 4.3860 250 0.7351 0.7907 0.7351 0.8574
No log 4.4211 252 0.8628 0.7314 0.8628 0.9289
No log 4.4561 254 0.8608 0.7386 0.8608 0.9278
No log 4.4912 256 0.7108 0.7975 0.7108 0.8431
No log 4.5263 258 0.7990 0.7333 0.7990 0.8939
No log 4.5614 260 0.9727 0.6301 0.9727 0.9863
No log 4.5965 262 0.9159 0.6483 0.9159 0.9570
No log 4.6316 264 0.8434 0.7133 0.8434 0.9184
No log 4.6667 266 0.8165 0.7517 0.8165 0.9036
No log 4.7018 268 0.8003 0.7682 0.8003 0.8946
No log 4.7368 270 0.7720 0.7712 0.7720 0.8786
No log 4.7719 272 0.7102 0.7898 0.7102 0.8427
No log 4.8070 274 0.6657 0.7925 0.6657 0.8159
No log 4.8421 276 0.6404 0.7904 0.6404 0.8003
No log 4.8772 278 0.7161 0.7892 0.7161 0.8462
No log 4.9123 280 0.9313 0.7766 0.9313 0.9650
No log 4.9474 282 1.1536 0.6333 1.1536 1.0741
No log 4.9825 284 0.9897 0.6667 0.9897 0.9948
No log 5.0175 286 0.6533 0.8229 0.6533 0.8083
No log 5.0526 288 0.5864 0.7898 0.5864 0.7658
No log 5.0877 290 0.6057 0.7922 0.6057 0.7783
No log 5.1228 292 0.6525 0.7853 0.6525 0.8078
No log 5.1579 294 0.7044 0.8140 0.7044 0.8393
No log 5.1930 296 0.6802 0.8208 0.6802 0.8247
No log 5.2281 298 0.6369 0.8171 0.6369 0.7981
No log 5.2632 300 0.6495 0.8075 0.6495 0.8059
No log 5.2982 302 0.6794 0.8125 0.6794 0.8242
No log 5.3333 304 0.6934 0.8125 0.6934 0.8327
No log 5.3684 306 0.6841 0.8098 0.6841 0.8271
No log 5.4035 308 0.6917 0.7901 0.6917 0.8317
No log 5.4386 310 0.6612 0.8024 0.6612 0.8131
No log 5.4737 312 0.6761 0.8263 0.6761 0.8223
No log 5.5088 314 0.7206 0.8087 0.7206 0.8489
No log 5.5439 316 0.6927 0.8222 0.6927 0.8323
No log 5.5789 318 0.6409 0.8161 0.6409 0.8005
No log 5.6140 320 0.6697 0.8229 0.6697 0.8184
No log 5.6491 322 0.6591 0.8214 0.6591 0.8119
No log 5.6842 324 0.6367 0.7815 0.6367 0.7979
No log 5.7193 326 0.6702 0.7347 0.6702 0.8187
No log 5.7544 328 0.6619 0.7211 0.6619 0.8136
No log 5.7895 330 0.6299 0.7211 0.6299 0.7937
No log 5.8246 332 0.5852 0.8052 0.5852 0.7650
No log 5.8596 334 0.6472 0.8140 0.6472 0.8045
No log 5.8947 336 0.8815 0.6893 0.8815 0.9389
No log 5.9298 338 0.9257 0.6705 0.9257 0.9621
No log 5.9649 340 0.8771 0.7303 0.8771 0.9365
No log 6.0 342 0.8005 0.7614 0.8005 0.8947
No log 6.0351 344 0.7298 0.8070 0.7298 0.8543
No log 6.0702 346 0.7388 0.7799 0.7388 0.8595
No log 6.1053 348 0.7613 0.7742 0.7613 0.8725
No log 6.1404 350 0.7935 0.7792 0.7935 0.8908
No log 6.1754 352 0.8259 0.6897 0.8259 0.9088
No log 6.2105 354 0.8393 0.6622 0.8393 0.9161
No log 6.2456 356 0.8146 0.7134 0.8146 0.9025
No log 6.2807 358 0.7268 0.7545 0.7268 0.8525
No log 6.3158 360 0.6397 0.7784 0.6397 0.7998
No log 6.3509 362 0.6175 0.7898 0.6175 0.7858
No log 6.3860 364 0.6025 0.7898 0.6025 0.7762
No log 6.4211 366 0.6197 0.7843 0.6197 0.7872
No log 6.4561 368 0.6248 0.7712 0.6248 0.7904
No log 6.4912 370 0.6317 0.7692 0.6317 0.7948
No log 6.5263 372 0.6769 0.7742 0.6769 0.8227
No log 6.5614 374 0.6645 0.7848 0.6645 0.8151
No log 6.5965 376 0.6417 0.7848 0.6417 0.8010
No log 6.6316 378 0.6335 0.7975 0.6335 0.7959
No log 6.6667 380 0.6534 0.7647 0.6534 0.8084
No log 6.7018 382 0.6813 0.7630 0.6813 0.8254
No log 6.7368 384 0.7013 0.7574 0.7013 0.8374
No log 6.7719 386 0.7293 0.7771 0.7293 0.8540
No log 6.8070 388 0.7499 0.75 0.7499 0.8660
No log 6.8421 390 0.7384 0.75 0.7384 0.8593
No log 6.8772 392 0.7026 0.7550 0.7026 0.8382
No log 6.9123 394 0.6674 0.7692 0.6674 0.8169
No log 6.9474 396 0.6653 0.7910 0.6653 0.8156
No log 6.9825 398 0.7364 0.7650 0.7364 0.8581
No log 7.0175 400 0.8144 0.7568 0.8144 0.9025
No log 7.0526 402 0.7325 0.8065 0.7325 0.8558
No log 7.0877 404 0.6245 0.7933 0.6245 0.7902
No log 7.1228 406 0.6155 0.7882 0.6155 0.7845
No log 7.1579 408 0.6812 0.7977 0.6812 0.8254
No log 7.1930 410 0.7847 0.7543 0.7847 0.8859
No log 7.2281 412 0.8035 0.7356 0.8035 0.8964
No log 7.2632 414 0.7366 0.7797 0.7366 0.8582
No log 7.2982 416 0.6394 0.7904 0.6394 0.7997
No log 7.3333 418 0.6254 0.7712 0.6254 0.7908
No log 7.3684 420 0.6466 0.7792 0.6466 0.8041
No log 7.4035 422 0.6619 0.7712 0.6619 0.8136
No log 7.4386 424 0.7198 0.7738 0.7198 0.8484
No log 7.4737 426 0.7867 0.7586 0.7867 0.8870
No log 7.5088 428 0.7612 0.7442 0.7612 0.8725
No log 7.5439 430 0.6963 0.7816 0.6963 0.8344
No log 7.5789 432 0.6520 0.7595 0.6520 0.8075
No log 7.6140 434 0.6403 0.7763 0.6403 0.8002
No log 7.6491 436 0.6324 0.7763 0.6324 0.7952
No log 7.6842 438 0.6296 0.7871 0.6296 0.7935
No log 7.7193 440 0.6250 0.8 0.6250 0.7905
No log 7.7544 442 0.6229 0.7746 0.6229 0.7893
No log 7.7895 444 0.6506 0.7841 0.6506 0.8066
No log 7.8246 446 0.6614 0.7841 0.6614 0.8133
No log 7.8596 448 0.6847 0.7758 0.6847 0.8274
No log 7.8947 450 0.7196 0.7342 0.7196 0.8483
No log 7.9298 452 0.7389 0.7342 0.7389 0.8596
No log 7.9649 454 0.7664 0.7578 0.7664 0.8754
No log 8.0 456 0.7778 0.7625 0.7778 0.8819
No log 8.0351 458 0.7552 0.7578 0.7552 0.8690
No log 8.0702 460 0.7331 0.75 0.7331 0.8562
No log 8.1053 462 0.7103 0.75 0.7103 0.8428
No log 8.1404 464 0.6852 0.7305 0.6852 0.8278
No log 8.1754 466 0.7057 0.8066 0.7057 0.8401
No log 8.2105 468 0.7466 0.7892 0.7466 0.8641
No log 8.2456 470 0.8440 0.7869 0.8440 0.9187
No log 8.2807 472 0.7838 0.7869 0.7838 0.8853
No log 8.3158 474 0.7188 0.7841 0.7188 0.8478
No log 8.3509 476 0.7009 0.7532 0.7009 0.8372
No log 8.3860 478 0.7314 0.7432 0.7314 0.8552
No log 8.4211 480 0.7457 0.7397 0.7457 0.8635
No log 8.4561 482 0.7646 0.7682 0.7646 0.8744
No log 8.4912 484 0.7929 0.7742 0.7929 0.8905
No log 8.5263 486 0.7915 0.7654 0.7915 0.8896
No log 8.5614 488 0.7749 0.7778 0.7749 0.8803
No log 8.5965 490 0.7607 0.7925 0.7607 0.8722
No log 8.6316 492 0.7323 0.7792 0.7323 0.8557
No log 8.6667 494 0.7075 0.7895 0.7075 0.8411
No log 8.7018 496 0.6886 0.8025 0.6886 0.8298
No log 8.7368 498 0.6817 0.8025 0.6817 0.8256
0.3556 8.7719 500 0.6827 0.8025 0.6827 0.8263
0.3556 8.8070 502 0.6870 0.8025 0.6870 0.8289
0.3556 8.8421 504 0.7012 0.7898 0.7012 0.8374
0.3556 8.8772 506 0.7114 0.7848 0.7114 0.8435
0.3556 8.9123 508 0.7044 0.7848 0.7044 0.8393
0.3556 8.9474 510 0.6998 0.7673 0.6998 0.8366

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
182
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k12_task1_organization

Finetuned
(4222)
this model