ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k20_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6473
  • Qwk: 0.4094
  • Mse: 0.6473
  • Rmse: 0.8045

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0194 2 4.1289 -0.0228 4.1289 2.0320
No log 0.0388 4 2.1791 0.0145 2.1791 1.4762
No log 0.0583 6 1.9670 -0.0877 1.9670 1.4025
No log 0.0777 8 1.1839 -0.0064 1.1839 1.0881
No log 0.0971 10 0.8621 0.1465 0.8621 0.9285
No log 0.1165 12 0.8701 0.1128 0.8701 0.9328
No log 0.1359 14 1.1183 -0.0572 1.1183 1.0575
No log 0.1553 16 1.1891 -0.0065 1.1891 1.0905
No log 0.1748 18 1.0494 -0.0073 1.0494 1.0244
No log 0.1942 20 0.8830 0.1610 0.8830 0.9397
No log 0.2136 22 0.7770 0.2505 0.7770 0.8815
No log 0.2330 24 0.7958 0.2805 0.7958 0.8921
No log 0.2524 26 1.1969 0.0126 1.1969 1.0940
No log 0.2718 28 1.7814 0.0884 1.7814 1.3347
No log 0.2913 30 1.4238 0.1134 1.4238 1.1932
No log 0.3107 32 0.7832 0.3540 0.7832 0.8850
No log 0.3301 34 0.7212 0.2743 0.7212 0.8493
No log 0.3495 36 0.7539 0.2868 0.7539 0.8683
No log 0.3689 38 0.7979 0.2582 0.7979 0.8932
No log 0.3883 40 0.9744 0.2669 0.9744 0.9871
No log 0.4078 42 1.1028 0.2228 1.1028 1.0501
No log 0.4272 44 0.9231 0.2447 0.9231 0.9608
No log 0.4466 46 0.7640 0.2780 0.7640 0.8741
No log 0.4660 48 0.7255 0.3028 0.7255 0.8518
No log 0.4854 50 0.8074 0.2704 0.8074 0.8986
No log 0.5049 52 0.9113 0.2157 0.9113 0.9546
No log 0.5243 54 0.8229 0.2405 0.8229 0.9072
No log 0.5437 56 0.7350 0.3196 0.7350 0.8573
No log 0.5631 58 0.7728 0.2933 0.7728 0.8791
No log 0.5825 60 0.9496 0.2329 0.9496 0.9745
No log 0.6019 62 1.2370 0.1638 1.2370 1.1122
No log 0.6214 64 1.3139 0.1462 1.3139 1.1462
No log 0.6408 66 1.1219 0.2300 1.1219 1.0592
No log 0.6602 68 1.0567 0.2685 1.0567 1.0279
No log 0.6796 70 1.1099 0.2784 1.1099 1.0535
No log 0.6990 72 0.8452 0.2975 0.8452 0.9193
No log 0.7184 74 0.7010 0.3277 0.7010 0.8373
No log 0.7379 76 0.6663 0.2998 0.6663 0.8163
No log 0.7573 78 0.6651 0.3009 0.6651 0.8155
No log 0.7767 80 0.7700 0.4175 0.7700 0.8775
No log 0.7961 82 1.2867 0.2577 1.2867 1.1343
No log 0.8155 84 1.4881 0.1169 1.4881 1.2199
No log 0.8350 86 1.0902 0.3049 1.0902 1.0441
No log 0.8544 88 0.9484 0.3014 0.9484 0.9738
No log 0.8738 90 0.9761 0.2958 0.9761 0.9880
No log 0.8932 92 0.9417 0.2755 0.9417 0.9704
No log 0.9126 94 0.8397 0.3480 0.8397 0.9164
No log 0.9320 96 0.7792 0.3438 0.7792 0.8827
No log 0.9515 98 0.7154 0.3903 0.7154 0.8458
No log 0.9709 100 0.7122 0.3894 0.7122 0.8439
No log 0.9903 102 0.6866 0.3756 0.6866 0.8286
No log 1.0097 104 0.6854 0.3240 0.6854 0.8279
No log 1.0291 106 0.6854 0.3028 0.6854 0.8279
No log 1.0485 108 0.8306 0.3195 0.8306 0.9114
No log 1.0680 110 0.8204 0.3616 0.8204 0.9058
No log 1.0874 112 0.6734 0.3840 0.6734 0.8206
No log 1.1068 114 0.7453 0.2634 0.7453 0.8633
No log 1.1262 116 0.7316 0.2842 0.7316 0.8553
No log 1.1456 118 0.6606 0.3941 0.6606 0.8128
No log 1.1650 120 0.7020 0.4401 0.7020 0.8378
No log 1.1845 122 0.6821 0.4328 0.6821 0.8259
No log 1.2039 124 0.6849 0.4022 0.6849 0.8276
No log 1.2233 126 0.7168 0.3947 0.7168 0.8466
No log 1.2427 128 0.8001 0.3395 0.8001 0.8945
No log 1.2621 130 1.0001 0.3064 1.0001 1.0000
No log 1.2816 132 0.8546 0.3528 0.8546 0.9245
No log 1.3010 134 0.7107 0.3756 0.7107 0.8430
No log 1.3204 136 0.7016 0.3428 0.7016 0.8376
No log 1.3398 138 0.6895 0.4506 0.6895 0.8304
No log 1.3592 140 0.8374 0.3558 0.8374 0.9151
No log 1.3786 142 0.8927 0.3013 0.8927 0.9448
No log 1.3981 144 0.7682 0.4115 0.7682 0.8765
No log 1.4175 146 0.8556 0.3226 0.8556 0.9250
No log 1.4369 148 0.9912 0.3551 0.9912 0.9956
No log 1.4563 150 0.8884 0.3454 0.8884 0.9425
No log 1.4757 152 0.7026 0.3837 0.7026 0.8382
No log 1.4951 154 0.7196 0.3373 0.7196 0.8483
No log 1.5146 156 0.7206 0.3694 0.7207 0.8489
No log 1.5340 158 0.7022 0.3827 0.7022 0.8380
No log 1.5534 160 0.7665 0.3989 0.7665 0.8755
No log 1.5728 162 0.7266 0.3821 0.7266 0.8524
No log 1.5922 164 0.7520 0.3415 0.7520 0.8672
No log 1.6117 166 0.8693 0.3428 0.8693 0.9324
No log 1.6311 168 0.8121 0.4186 0.8121 0.9011
No log 1.6505 170 0.7620 0.4348 0.7620 0.8729
No log 1.6699 172 0.7491 0.4212 0.7491 0.8655
No log 1.6893 174 0.8223 0.4544 0.8223 0.9068
No log 1.7087 176 0.8921 0.4199 0.8921 0.9445
No log 1.7282 178 0.9972 0.3972 0.9972 0.9986
No log 1.7476 180 0.8679 0.4161 0.8679 0.9316
No log 1.7670 182 0.7117 0.4916 0.7117 0.8436
No log 1.7864 184 0.6305 0.4840 0.6305 0.7941
No log 1.8058 186 0.6351 0.4508 0.6351 0.7969
No log 1.8252 188 0.6015 0.4014 0.6015 0.7756
No log 1.8447 190 0.6736 0.3612 0.6736 0.8207
No log 1.8641 192 0.8373 0.3167 0.8373 0.9150
No log 1.8835 194 0.9050 0.3088 0.9050 0.9513
No log 1.9029 196 0.7791 0.3764 0.7791 0.8826
No log 1.9223 198 0.6566 0.4397 0.6566 0.8103
No log 1.9417 200 0.6211 0.4590 0.6211 0.7881
No log 1.9612 202 0.6264 0.4691 0.6264 0.7915
No log 1.9806 204 0.6982 0.4379 0.6982 0.8356
No log 2.0 206 0.9727 0.3860 0.9727 0.9863
No log 2.0194 208 1.4482 0.2098 1.4482 1.2034
No log 2.0388 210 1.3560 0.2142 1.3560 1.1645
No log 2.0583 212 1.0564 0.3783 1.0564 1.0278
No log 2.0777 214 0.7344 0.4260 0.7344 0.8569
No log 2.0971 216 0.6735 0.3894 0.6735 0.8207
No log 2.1165 218 0.6623 0.4229 0.6623 0.8138
No log 2.1359 220 0.7009 0.3939 0.7009 0.8372
No log 2.1553 222 0.7625 0.4142 0.7625 0.8732
No log 2.1748 224 0.7043 0.4701 0.7043 0.8393
No log 2.1942 226 0.6780 0.4240 0.6780 0.8234
No log 2.2136 228 0.7023 0.4384 0.7023 0.8380
No log 2.2330 230 0.7970 0.3900 0.7970 0.8927
No log 2.2524 232 0.9161 0.3998 0.9161 0.9571
No log 2.2718 234 0.8438 0.4187 0.8438 0.9186
No log 2.2913 236 0.8187 0.4243 0.8187 0.9048
No log 2.3107 238 0.6980 0.4255 0.6980 0.8354
No log 2.3301 240 0.6678 0.4373 0.6678 0.8172
No log 2.3495 242 0.6821 0.4465 0.6821 0.8259
No log 2.3689 244 0.7019 0.4548 0.7019 0.8378
No log 2.3883 246 0.6421 0.4697 0.6421 0.8013
No log 2.4078 248 0.6318 0.4690 0.6318 0.7948
No log 2.4272 250 0.6369 0.5277 0.6369 0.7981
No log 2.4466 252 0.6333 0.4898 0.6333 0.7958
No log 2.4660 254 0.6498 0.4964 0.6498 0.8061
No log 2.4854 256 0.6496 0.5050 0.6496 0.8060
No log 2.5049 258 0.6606 0.5010 0.6606 0.8128
No log 2.5243 260 0.6622 0.5402 0.6622 0.8137
No log 2.5437 262 0.6564 0.5280 0.6564 0.8102
No log 2.5631 264 0.6674 0.5134 0.6674 0.8169
No log 2.5825 266 0.6750 0.5234 0.6750 0.8216
No log 2.6019 268 0.6890 0.5109 0.6890 0.8301
No log 2.6214 270 0.7017 0.5119 0.7017 0.8377
No log 2.6408 272 0.6415 0.4299 0.6415 0.8009
No log 2.6602 274 0.6436 0.4163 0.6436 0.8022
No log 2.6796 276 0.6425 0.4188 0.6425 0.8016
No log 2.6990 278 0.6440 0.4573 0.6440 0.8025
No log 2.7184 280 0.7073 0.4639 0.7073 0.8410
No log 2.7379 282 0.7718 0.3968 0.7718 0.8785
No log 2.7573 284 0.6795 0.4878 0.6795 0.8243
No log 2.7767 286 0.6951 0.4206 0.6951 0.8337
No log 2.7961 288 0.7125 0.4318 0.7125 0.8441
No log 2.8155 290 0.6424 0.4452 0.6424 0.8015
No log 2.8350 292 0.6529 0.4590 0.6529 0.8080
No log 2.8544 294 0.8419 0.4109 0.8419 0.9176
No log 2.8738 296 0.9501 0.4090 0.9501 0.9747
No log 2.8932 298 0.8674 0.4194 0.8674 0.9313
No log 2.9126 300 0.7297 0.4545 0.7297 0.8543
No log 2.9320 302 0.6604 0.4076 0.6604 0.8126
No log 2.9515 304 0.6828 0.4135 0.6828 0.8263
No log 2.9709 306 0.6822 0.4402 0.6822 0.8260
No log 2.9903 308 0.6412 0.4436 0.6412 0.8007
No log 3.0097 310 0.6621 0.4674 0.6621 0.8137
No log 3.0291 312 0.6825 0.4797 0.6825 0.8262
No log 3.0485 314 0.6506 0.4355 0.6506 0.8066
No log 3.0680 316 0.6225 0.3956 0.6225 0.7890
No log 3.0874 318 0.6083 0.4112 0.6083 0.7799
No log 3.1068 320 0.6208 0.4355 0.6208 0.7879
No log 3.1262 322 0.6391 0.4575 0.6391 0.7995
No log 3.1456 324 0.6123 0.4089 0.6123 0.7825
No log 3.1650 326 0.5848 0.5467 0.5848 0.7647
No log 3.1845 328 0.6074 0.5048 0.6074 0.7794
No log 3.2039 330 0.5828 0.5459 0.5828 0.7634
No log 3.2233 332 0.6501 0.4222 0.6501 0.8063
No log 3.2427 334 0.6950 0.4538 0.6950 0.8337
No log 3.2621 336 0.6512 0.4374 0.6512 0.8069
No log 3.2816 338 0.5843 0.4670 0.5843 0.7644
No log 3.3010 340 0.5953 0.5143 0.5953 0.7715
No log 3.3204 342 0.5905 0.4635 0.5905 0.7685
No log 3.3398 344 0.6400 0.4710 0.6400 0.8000
No log 3.3592 346 0.7336 0.4750 0.7336 0.8565
No log 3.3786 348 0.6971 0.4779 0.6971 0.8349
No log 3.3981 350 0.6765 0.4920 0.6765 0.8225
No log 3.4175 352 0.6890 0.4400 0.6890 0.8300
No log 3.4369 354 0.6718 0.4051 0.6718 0.8196
No log 3.4563 356 0.6238 0.3354 0.6238 0.7898
No log 3.4757 358 0.6089 0.3855 0.6089 0.7803
No log 3.4951 360 0.6195 0.3855 0.6195 0.7871
No log 3.5146 362 0.6528 0.3836 0.6528 0.8080
No log 3.5340 364 0.6555 0.3977 0.6555 0.8096
No log 3.5534 366 0.6027 0.4130 0.6027 0.7763
No log 3.5728 368 0.5951 0.4690 0.5951 0.7714
No log 3.5922 370 0.6102 0.4928 0.6102 0.7812
No log 3.6117 372 0.6236 0.4999 0.6236 0.7897
No log 3.6311 374 0.6050 0.4632 0.6050 0.7778
No log 3.6505 376 0.6586 0.4769 0.6586 0.8115
No log 3.6699 378 0.7028 0.4452 0.7028 0.8383
No log 3.6893 380 0.6534 0.4639 0.6534 0.8083
No log 3.7087 382 0.6045 0.4530 0.6045 0.7775
No log 3.7282 384 0.6420 0.4605 0.6420 0.8013
No log 3.7476 386 0.6662 0.4265 0.6662 0.8162
No log 3.7670 388 0.6335 0.4269 0.6335 0.7959
No log 3.7864 390 0.6346 0.4402 0.6346 0.7966
No log 3.8058 392 0.7722 0.3485 0.7722 0.8788
No log 3.8252 394 0.8704 0.3328 0.8704 0.9329
No log 3.8447 396 0.7767 0.3312 0.7767 0.8813
No log 3.8641 398 0.6560 0.3573 0.6560 0.8099
No log 3.8835 400 0.6669 0.4530 0.6669 0.8166
No log 3.9029 402 0.7187 0.4288 0.7187 0.8478
No log 3.9223 404 0.6715 0.4552 0.6715 0.8194
No log 3.9417 406 0.6333 0.3975 0.6333 0.7958
No log 3.9612 408 0.7501 0.4531 0.7501 0.8661
No log 3.9806 410 0.8964 0.3813 0.8964 0.9468
No log 4.0 412 0.8708 0.3782 0.8708 0.9332
No log 4.0194 414 0.7345 0.4169 0.7345 0.8571
No log 4.0388 416 0.6622 0.3829 0.6622 0.8137
No log 4.0583 418 0.6611 0.4032 0.6611 0.8131
No log 4.0777 420 0.6608 0.3902 0.6608 0.8129
No log 4.0971 422 0.6426 0.3792 0.6426 0.8016
No log 4.1165 424 0.6225 0.3597 0.6225 0.7890
No log 4.1359 426 0.6234 0.4033 0.6234 0.7896
No log 4.1553 428 0.6240 0.3710 0.6240 0.7899
No log 4.1748 430 0.6187 0.4079 0.6187 0.7866
No log 4.1942 432 0.6270 0.3555 0.6270 0.7918
No log 4.2136 434 0.6345 0.3822 0.6345 0.7965
No log 4.2330 436 0.6495 0.3668 0.6495 0.8059
No log 4.2524 438 0.6924 0.3752 0.6924 0.8321
No log 4.2718 440 0.7407 0.3892 0.7407 0.8606
No log 4.2913 442 0.7020 0.3884 0.7020 0.8378
No log 4.3107 444 0.6714 0.4358 0.6714 0.8194
No log 4.3301 446 0.6731 0.4307 0.6731 0.8204
No log 4.3495 448 0.6704 0.4253 0.6704 0.8188
No log 4.3689 450 0.6660 0.4292 0.6660 0.8161
No log 4.3883 452 0.6822 0.4295 0.6822 0.8260
No log 4.4078 454 0.6768 0.4236 0.6768 0.8227
No log 4.4272 456 0.6308 0.4289 0.6308 0.7942
No log 4.4466 458 0.6129 0.4108 0.6129 0.7829
No log 4.4660 460 0.6371 0.4191 0.6371 0.7982
No log 4.4854 462 0.6495 0.4343 0.6495 0.8059
No log 4.5049 464 0.6334 0.4347 0.6334 0.7959
No log 4.5243 466 0.6566 0.4543 0.6566 0.8103
No log 4.5437 468 0.6861 0.4444 0.6861 0.8283
No log 4.5631 470 0.6588 0.4855 0.6588 0.8117
No log 4.5825 472 0.6396 0.4418 0.6396 0.7997
No log 4.6019 474 0.6555 0.4403 0.6555 0.8096
No log 4.6214 476 0.6478 0.4345 0.6478 0.8048
No log 4.6408 478 0.6460 0.4539 0.6460 0.8038
No log 4.6602 480 0.6339 0.4302 0.6339 0.7962
No log 4.6796 482 0.6552 0.4339 0.6552 0.8095
No log 4.6990 484 0.6536 0.4472 0.6536 0.8085
No log 4.7184 486 0.6250 0.4433 0.6250 0.7906
No log 4.7379 488 0.6222 0.4270 0.6222 0.7888
No log 4.7573 490 0.6338 0.4027 0.6338 0.7961
No log 4.7767 492 0.6479 0.4465 0.6479 0.8049
No log 4.7961 494 0.6326 0.3705 0.6326 0.7953
No log 4.8155 496 0.6550 0.4512 0.6550 0.8093
No log 4.8350 498 0.6659 0.4385 0.6659 0.8160
0.3936 4.8544 500 0.6361 0.4465 0.6361 0.7976
0.3936 4.8738 502 0.6234 0.4038 0.6234 0.7896
0.3936 4.8932 504 0.6461 0.5086 0.6461 0.8038
0.3936 4.9126 506 0.6709 0.5 0.6709 0.8191
0.3936 4.9320 508 0.6479 0.4956 0.6479 0.8049
0.3936 4.9515 510 0.6279 0.4436 0.6279 0.7924
0.3936 4.9709 512 0.6812 0.4268 0.6812 0.8253
0.3936 4.9903 514 0.7281 0.4102 0.7281 0.8533
0.3936 5.0097 516 0.6684 0.4270 0.6684 0.8175
0.3936 5.0291 518 0.6473 0.4094 0.6473 0.8045

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
6
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k20_task2_organization

Finetuned
(4222)
this model