ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k11_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4509
  • Qwk: 0.5840
  • Mse: 0.4509
  • Rmse: 0.6715

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0690 2 2.4713 -0.0262 2.4713 1.5720
No log 0.1379 4 1.3112 0.0998 1.3112 1.1451
No log 0.2069 6 0.8126 0.0535 0.8126 0.9014
No log 0.2759 8 0.9384 0.0927 0.9384 0.9687
No log 0.3448 10 0.9361 0.2939 0.9361 0.9675
No log 0.4138 12 0.9504 0.2601 0.9504 0.9749
No log 0.4828 14 0.8084 0.2285 0.8084 0.8991
No log 0.5517 16 0.7364 -0.0027 0.7364 0.8582
No log 0.6207 18 0.7324 0.0898 0.7324 0.8558
No log 0.6897 20 0.7246 0.2494 0.7246 0.8513
No log 0.7586 22 0.6418 0.1277 0.6418 0.8011
No log 0.8276 24 0.6349 0.1604 0.6349 0.7968
No log 0.8966 26 0.6258 0.2002 0.6258 0.7911
No log 0.9655 28 0.6073 0.2783 0.6073 0.7793
No log 1.0345 30 0.6091 0.3426 0.6091 0.7804
No log 1.1034 32 0.5988 0.3377 0.5988 0.7738
No log 1.1724 34 0.7307 0.2886 0.7307 0.8548
No log 1.2414 36 0.8048 0.3094 0.8048 0.8971
No log 1.3103 38 0.7674 0.2817 0.7674 0.8760
No log 1.3793 40 0.6644 0.3999 0.6644 0.8151
No log 1.4483 42 0.5600 0.5361 0.5600 0.7483
No log 1.5172 44 0.6354 0.1935 0.6354 0.7971
No log 1.5862 46 0.8653 0.0717 0.8653 0.9302
No log 1.6552 48 0.8344 0.1268 0.8344 0.9135
No log 1.7241 50 0.6253 0.2995 0.6253 0.7907
No log 1.7931 52 0.5520 0.4938 0.5520 0.7429
No log 1.8621 54 0.6968 0.4280 0.6968 0.8348
No log 1.9310 56 0.9318 0.2009 0.9318 0.9653
No log 2.0 58 0.7647 0.3484 0.7647 0.8745
No log 2.0690 60 0.5696 0.4229 0.5696 0.7547
No log 2.1379 62 0.7541 0.3606 0.7541 0.8684
No log 2.2069 64 0.7939 0.3473 0.7939 0.8910
No log 2.2759 66 0.6665 0.2027 0.6665 0.8164
No log 2.3448 68 0.6484 0.2892 0.6484 0.8052
No log 2.4138 70 0.6588 0.2892 0.6588 0.8116
No log 2.4828 72 0.6602 0.2374 0.6602 0.8125
No log 2.5517 74 0.6708 0.2506 0.6708 0.8190
No log 2.6207 76 0.7577 0.2692 0.7577 0.8704
No log 2.6897 78 0.8452 0.3042 0.8452 0.9194
No log 2.7586 80 0.7794 0.3302 0.7794 0.8828
No log 2.8276 82 0.7342 0.2995 0.7342 0.8568
No log 2.8966 84 0.6643 0.3947 0.6643 0.8151
No log 2.9655 86 0.6658 0.4448 0.6658 0.8160
No log 3.0345 88 0.6493 0.4019 0.6493 0.8058
No log 3.1034 90 0.7082 0.4134 0.7082 0.8416
No log 3.1724 92 0.6083 0.5345 0.6083 0.7799
No log 3.2414 94 0.5783 0.4898 0.5783 0.7605
No log 3.3103 96 0.6343 0.4698 0.6343 0.7964
No log 3.3793 98 0.5718 0.4494 0.5718 0.7562
No log 3.4483 100 0.6080 0.4905 0.6080 0.7797
No log 3.5172 102 0.7659 0.4462 0.7659 0.8751
No log 3.5862 104 0.7333 0.4462 0.7333 0.8564
No log 3.6552 106 0.5705 0.5067 0.5705 0.7553
No log 3.7241 108 0.5439 0.4918 0.5439 0.7375
No log 3.7931 110 0.6864 0.5146 0.6864 0.8285
No log 3.8621 112 0.7508 0.4784 0.7508 0.8665
No log 3.9310 114 0.6642 0.5704 0.6642 0.8150
No log 4.0 116 0.6147 0.5285 0.6147 0.7840
No log 4.0690 118 0.5934 0.5397 0.5934 0.7703
No log 4.1379 120 0.5864 0.5646 0.5864 0.7658
No log 4.2069 122 0.5621 0.4696 0.5621 0.7497
No log 4.2759 124 0.5438 0.5617 0.5438 0.7375
No log 4.3448 126 0.5358 0.5184 0.5358 0.7320
No log 4.4138 128 0.5188 0.4478 0.5188 0.7203
No log 4.4828 130 0.5471 0.5587 0.5471 0.7397
No log 4.5517 132 0.5155 0.4762 0.5155 0.7180
No log 4.6207 134 0.5186 0.4828 0.5186 0.7201
No log 4.6897 136 0.5290 0.4828 0.5290 0.7273
No log 4.7586 138 0.5445 0.5157 0.5445 0.7379
No log 4.8276 140 0.5838 0.5195 0.5838 0.7641
No log 4.8966 142 0.5495 0.5157 0.5495 0.7413
No log 4.9655 144 0.5287 0.5184 0.5287 0.7271
No log 5.0345 146 0.5225 0.5413 0.5225 0.7229
No log 5.1034 148 0.5280 0.4596 0.5280 0.7267
No log 5.1724 150 0.5247 0.5798 0.5247 0.7244
No log 5.2414 152 0.5274 0.5714 0.5274 0.7263
No log 5.3103 154 0.5339 0.5475 0.5339 0.7307
No log 5.3793 156 0.5509 0.4147 0.5509 0.7423
No log 5.4483 158 0.6433 0.4243 0.6433 0.8020
No log 5.5172 160 0.6569 0.4243 0.6569 0.8105
No log 5.5862 162 0.5680 0.4379 0.5680 0.7537
No log 5.6552 164 0.5734 0.5286 0.5734 0.7572
No log 5.7241 166 0.5865 0.5697 0.5865 0.7659
No log 5.7931 168 0.6037 0.5190 0.6037 0.7770
No log 5.8621 170 0.6297 0.5220 0.6297 0.7935
No log 5.9310 172 0.6042 0.5220 0.6042 0.7773
No log 6.0 174 0.5347 0.5596 0.5347 0.7312
No log 6.0690 176 0.5285 0.5413 0.5285 0.7270
No log 6.1379 178 0.5218 0.5672 0.5218 0.7223
No log 6.2069 180 0.5374 0.5205 0.5374 0.7331
No log 6.2759 182 0.5950 0.5244 0.5950 0.7713
No log 6.3448 184 0.5973 0.5273 0.5973 0.7729
No log 6.4138 186 0.6052 0.5483 0.6052 0.7779
No log 6.4828 188 0.5962 0.5855 0.5962 0.7722
No log 6.5517 190 0.5579 0.5143 0.5579 0.7469
No log 6.6207 192 0.7073 0.3973 0.7073 0.8410
No log 6.6897 194 1.0740 0.3455 1.0740 1.0363
No log 6.7586 196 1.1042 0.2601 1.1042 1.0508
No log 6.8276 198 0.8808 0.3269 0.8808 0.9385
No log 6.8966 200 0.6467 0.3399 0.6467 0.8042
No log 6.9655 202 0.5957 0.4345 0.5957 0.7718
No log 7.0345 204 0.6242 0.4821 0.6242 0.7901
No log 7.1034 206 0.5822 0.4590 0.5822 0.7630
No log 7.1724 208 0.6002 0.3665 0.6002 0.7747
No log 7.2414 210 0.6332 0.3936 0.6332 0.7958
No log 7.3103 212 0.6462 0.4836 0.6462 0.8039
No log 7.3793 214 0.5909 0.3936 0.5909 0.7687
No log 7.4483 216 0.5351 0.4591 0.5351 0.7315
No log 7.5172 218 0.5461 0.4821 0.5461 0.7390
No log 7.5862 220 0.5641 0.5065 0.5641 0.7510
No log 7.6552 222 0.5200 0.5571 0.5200 0.7211
No log 7.7241 224 0.5077 0.5596 0.5077 0.7125
No log 7.7931 226 0.5054 0.5953 0.5054 0.7109
No log 7.8621 228 0.5169 0.5796 0.5169 0.7190
No log 7.9310 230 0.5128 0.5796 0.5128 0.7161
No log 8.0 232 0.4932 0.5796 0.4932 0.7023
No log 8.0690 234 0.4858 0.5985 0.4858 0.6970
No log 8.1379 236 0.4888 0.5826 0.4888 0.6991
No log 8.2069 238 0.4989 0.6465 0.4989 0.7063
No log 8.2759 240 0.5164 0.6361 0.5164 0.7186
No log 8.3448 242 0.5089 0.6130 0.5089 0.7134
No log 8.4138 244 0.4949 0.6020 0.4949 0.7035
No log 8.4828 246 0.4993 0.5283 0.4993 0.7066
No log 8.5517 248 0.4868 0.5356 0.4868 0.6977
No log 8.6207 250 0.4807 0.5556 0.4807 0.6933
No log 8.6897 252 0.4887 0.5986 0.4887 0.6991
No log 8.7586 254 0.5041 0.5819 0.5041 0.7100
No log 8.8276 256 0.4887 0.5550 0.4887 0.6991
No log 8.8966 258 0.4855 0.5714 0.4855 0.6968
No log 8.9655 260 0.4876 0.5265 0.4876 0.6983
No log 9.0345 262 0.4951 0.5432 0.4951 0.7037
No log 9.1034 264 0.4906 0.5672 0.4906 0.7004
No log 9.1724 266 0.5003 0.6142 0.5003 0.7073
No log 9.2414 268 0.5218 0.5708 0.5218 0.7223
No log 9.3103 270 0.5462 0.5166 0.5462 0.7390
No log 9.3793 272 0.5210 0.5034 0.5210 0.7218
No log 9.4483 274 0.5220 0.5104 0.5220 0.7225
No log 9.5172 276 0.5066 0.5123 0.5066 0.7118
No log 9.5862 278 0.4823 0.6648 0.4823 0.6945
No log 9.6552 280 0.4846 0.6068 0.4846 0.6961
No log 9.7241 282 0.4823 0.5861 0.4823 0.6945
No log 9.7931 284 0.4958 0.5796 0.4958 0.7042
No log 9.8621 286 0.5088 0.5512 0.5088 0.7133
No log 9.9310 288 0.5458 0.5161 0.5458 0.7388
No log 10.0 290 0.5141 0.5438 0.5141 0.7170
No log 10.0690 292 0.5042 0.5587 0.5042 0.7101
No log 10.1379 294 0.5065 0.5587 0.5065 0.7117
No log 10.2069 296 0.4964 0.5510 0.4964 0.7046
No log 10.2759 298 0.4900 0.5731 0.4900 0.7000
No log 10.3448 300 0.4822 0.5324 0.4822 0.6944
No log 10.4138 302 0.4786 0.5574 0.4786 0.6918
No log 10.4828 304 0.4759 0.5798 0.4759 0.6898
No log 10.5517 306 0.4828 0.5874 0.4828 0.6949
No log 10.6207 308 0.4704 0.5930 0.4704 0.6858
No log 10.6897 310 0.4689 0.6068 0.4689 0.6848
No log 10.7586 312 0.4673 0.5517 0.4673 0.6836
No log 10.8276 314 0.4878 0.5195 0.4878 0.6984
No log 10.8966 316 0.4869 0.4684 0.4869 0.6978
No log 10.9655 318 0.4836 0.5104 0.4836 0.6954
No log 11.0345 320 0.4508 0.5738 0.4508 0.6714
No log 11.1034 322 0.4464 0.6184 0.4464 0.6681
No log 11.1724 324 0.4821 0.6248 0.4821 0.6943
No log 11.2414 326 0.4681 0.6235 0.4681 0.6842
No log 11.3103 328 0.4407 0.6046 0.4407 0.6638
No log 11.3793 330 0.5660 0.4892 0.5660 0.7524
No log 11.4483 332 0.6588 0.4925 0.6588 0.8116
No log 11.5172 334 0.5982 0.4707 0.5982 0.7735
No log 11.5862 336 0.4797 0.6060 0.4797 0.6926
No log 11.6552 338 0.4562 0.6462 0.4562 0.6754
No log 11.7241 340 0.5094 0.5368 0.5094 0.7137
No log 11.7931 342 0.4989 0.6047 0.4989 0.7063
No log 11.8621 344 0.4456 0.6579 0.4456 0.6675
No log 11.9310 346 0.4363 0.5912 0.4363 0.6606
No log 12.0 348 0.4382 0.5912 0.4382 0.6619
No log 12.0690 350 0.4329 0.6060 0.4329 0.6579
No log 12.1379 352 0.4473 0.6068 0.4473 0.6688
No log 12.2069 354 0.4470 0.6068 0.4470 0.6685
No log 12.2759 356 0.4463 0.5455 0.4463 0.6680
No log 12.3448 358 0.4937 0.5481 0.4937 0.7027
No log 12.4138 360 0.5888 0.5292 0.5888 0.7673
No log 12.4828 362 0.6437 0.4808 0.6437 0.8023
No log 12.5517 364 0.5760 0.5292 0.5760 0.7590
No log 12.6207 366 0.4886 0.5177 0.4886 0.6990
No log 12.6897 368 0.4686 0.4375 0.4686 0.6846
No log 12.7586 370 0.4685 0.4441 0.4685 0.6844
No log 12.8276 372 0.4747 0.4375 0.4747 0.6890
No log 12.8966 374 0.5241 0.5233 0.5241 0.7239
No log 12.9655 376 0.6221 0.4728 0.6221 0.7887
No log 13.0345 378 0.6447 0.5072 0.6447 0.8029
No log 13.1034 380 0.5659 0.4815 0.5659 0.7523
No log 13.1724 382 0.4750 0.5177 0.4750 0.6892
No log 13.2414 384 0.4575 0.5567 0.4575 0.6764
No log 13.3103 386 0.4554 0.5567 0.4554 0.6748
No log 13.3793 388 0.4585 0.6101 0.4585 0.6771
No log 13.4483 390 0.4728 0.6101 0.4728 0.6876
No log 13.5172 392 0.4706 0.6101 0.4706 0.6860
No log 13.5862 394 0.4470 0.6313 0.4470 0.6686
No log 13.6552 396 0.4468 0.6477 0.4468 0.6684
No log 13.7241 398 0.5096 0.6061 0.5096 0.7139
No log 13.7931 400 0.5128 0.5855 0.5128 0.7161
No log 13.8621 402 0.4682 0.6556 0.4682 0.6843
No log 13.9310 404 0.4491 0.6106 0.4491 0.6702
No log 14.0 406 0.4531 0.6200 0.4531 0.6732
No log 14.0690 408 0.4537 0.6200 0.4537 0.6736
No log 14.1379 410 0.4466 0.5960 0.4466 0.6683
No log 14.2069 412 0.4472 0.6106 0.4472 0.6687
No log 14.2759 414 0.4481 0.6321 0.4481 0.6694
No log 14.3448 416 0.4573 0.5930 0.4573 0.6762
No log 14.4138 418 0.4618 0.5672 0.4618 0.6795
No log 14.4828 420 0.4558 0.5252 0.4558 0.6751
No log 14.5517 422 0.4553 0.5593 0.4553 0.6748
No log 14.6207 424 0.4481 0.6334 0.4481 0.6694
No log 14.6897 426 0.4497 0.6068 0.4497 0.6706
No log 14.7586 428 0.4540 0.6317 0.4540 0.6738
No log 14.8276 430 0.4339 0.6243 0.4339 0.6587
No log 14.8966 432 0.4278 0.6641 0.4278 0.6541
No log 14.9655 434 0.4372 0.5386 0.4372 0.6612
No log 15.0345 436 0.4379 0.5587 0.4379 0.6617
No log 15.1034 438 0.4349 0.5836 0.4349 0.6595
No log 15.1724 440 0.4278 0.6661 0.4278 0.6541
No log 15.2414 442 0.4186 0.6279 0.4186 0.6470
No log 15.3103 444 0.4186 0.6279 0.4186 0.6470
No log 15.3793 446 0.4180 0.6383 0.4180 0.6465
No log 15.4483 448 0.4284 0.6683 0.4284 0.6545
No log 15.5172 450 0.4710 0.6169 0.4710 0.6863
No log 15.5862 452 0.4719 0.6169 0.4719 0.6869
No log 15.6552 454 0.4526 0.6705 0.4526 0.6727
No log 15.7241 456 0.4213 0.6771 0.4213 0.6490
No log 15.7931 458 0.4120 0.6852 0.4120 0.6418
No log 15.8621 460 0.4127 0.6946 0.4127 0.6424
No log 15.9310 462 0.4272 0.6873 0.4272 0.6536
No log 16.0 464 0.4304 0.6873 0.4304 0.6560
No log 16.0690 466 0.4323 0.6873 0.4323 0.6575
No log 16.1379 468 0.4168 0.6771 0.4168 0.6456
No log 16.2069 470 0.4094 0.6563 0.4094 0.6398
No log 16.2759 472 0.4140 0.6641 0.4140 0.6434
No log 16.3448 474 0.4098 0.6553 0.4098 0.6401
No log 16.4138 476 0.4080 0.6741 0.4080 0.6388
No log 16.4828 478 0.4088 0.6553 0.4088 0.6394
No log 16.5517 480 0.4128 0.6443 0.4128 0.6425
No log 16.6207 482 0.4167 0.6326 0.4167 0.6455
No log 16.6897 484 0.4182 0.6267 0.4182 0.6467
No log 16.7586 486 0.4214 0.6279 0.4214 0.6492
No log 16.8276 488 0.4253 0.6129 0.4253 0.6521
No log 16.8966 490 0.4392 0.6506 0.4392 0.6627
No log 16.9655 492 0.4418 0.5980 0.4418 0.6647
No log 17.0345 494 0.4250 0.5926 0.4250 0.6519
No log 17.1034 496 0.4276 0.6395 0.4276 0.6539
No log 17.1724 498 0.4424 0.5929 0.4424 0.6651
0.3437 17.2414 500 0.4542 0.6130 0.4542 0.6739
0.3437 17.3103 502 0.4464 0.6105 0.4464 0.6682
0.3437 17.3793 504 0.4468 0.6727 0.4468 0.6684
0.3437 17.4483 506 0.4523 0.5707 0.4523 0.6725
0.3437 17.5172 508 0.4554 0.6040 0.4554 0.6749
0.3437 17.5862 510 0.4568 0.6040 0.4568 0.6759
0.3437 17.6552 512 0.4597 0.6130 0.4597 0.6780
0.3437 17.7241 514 0.4509 0.5840 0.4509 0.6715

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
8
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k11_task7_organization

Finetuned
(4223)
this model