ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k15_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6211
  • Qwk: 0.4252
  • Mse: 0.6211
  • Rmse: 0.7881

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0256 2 4.2420 -0.0191 4.2420 2.0596
No log 0.0513 4 2.4511 0.0354 2.4511 1.5656
No log 0.0769 6 1.4855 0.0476 1.4855 1.2188
No log 0.1026 8 2.0208 -0.0487 2.0208 1.4215
No log 0.1282 10 1.6882 0.0056 1.6882 1.2993
No log 0.1538 12 1.6185 0.0231 1.6185 1.2722
No log 0.1795 14 1.1733 0.0343 1.1733 1.0832
No log 0.2051 16 1.2153 -0.0071 1.2153 1.1024
No log 0.2308 18 1.1298 0.0084 1.1298 1.0629
No log 0.2564 20 1.0810 0.0628 1.0810 1.0397
No log 0.2821 22 0.9292 0.1142 0.9292 0.9639
No log 0.3077 24 0.7597 0.2253 0.7597 0.8716
No log 0.3333 26 0.8461 0.0640 0.8461 0.9198
No log 0.3590 28 0.8501 0.1806 0.8501 0.9220
No log 0.3846 30 0.7204 0.2391 0.7204 0.8487
No log 0.4103 32 0.9483 0.1388 0.9483 0.9738
No log 0.4359 34 1.5415 0.0633 1.5415 1.2416
No log 0.4615 36 1.4443 0.0473 1.4443 1.2018
No log 0.4872 38 1.1535 0.1152 1.1535 1.0740
No log 0.5128 40 0.9933 0.2007 0.9933 0.9966
No log 0.5385 42 0.8288 0.2937 0.8288 0.9104
No log 0.5641 44 0.7997 0.3055 0.7997 0.8943
No log 0.5897 46 0.8727 0.2757 0.8727 0.9342
No log 0.6154 48 1.1296 0.2497 1.1296 1.0628
No log 0.6410 50 1.1650 0.2409 1.1650 1.0793
No log 0.6667 52 1.0912 0.2332 1.0912 1.0446
No log 0.6923 54 0.8542 0.3366 0.8542 0.9242
No log 0.7179 56 0.6075 0.4641 0.6075 0.7794
No log 0.7436 58 0.6852 0.4515 0.6852 0.8278
No log 0.7692 60 1.0603 0.1416 1.0603 1.0297
No log 0.7949 62 1.0827 0.1387 1.0827 1.0405
No log 0.8205 64 0.8400 0.2970 0.8400 0.9165
No log 0.8462 66 0.5829 0.4198 0.5829 0.7635
No log 0.8718 68 0.6961 0.3206 0.6960 0.8343
No log 0.8974 70 1.0970 0.1603 1.0970 1.0474
No log 0.9231 72 1.2072 0.1304 1.2072 1.0987
No log 0.9487 74 1.0020 0.2133 1.0020 1.0010
No log 0.9744 76 0.8352 0.3441 0.8352 0.9139
No log 1.0 78 0.7032 0.3291 0.7032 0.8386
No log 1.0256 80 0.6270 0.4100 0.6270 0.7918
No log 1.0513 82 0.6404 0.3798 0.6404 0.8003
No log 1.0769 84 0.6762 0.3810 0.6762 0.8223
No log 1.1026 86 0.7279 0.4180 0.7279 0.8532
No log 1.1282 88 1.1147 0.3505 1.1147 1.0558
No log 1.1538 90 1.4353 0.2419 1.4353 1.1981
No log 1.1795 92 1.2663 0.2886 1.2663 1.1253
No log 1.2051 94 0.7584 0.4407 0.7584 0.8709
No log 1.2308 96 0.5843 0.5034 0.5843 0.7644
No log 1.2564 98 0.5927 0.4630 0.5927 0.7699
No log 1.2821 100 0.6105 0.3954 0.6105 0.7813
No log 1.3077 102 0.6866 0.3704 0.6866 0.8286
No log 1.3333 104 0.7174 0.3911 0.7174 0.8470
No log 1.3590 106 0.7152 0.3797 0.7152 0.8457
No log 1.3846 108 0.6530 0.3875 0.6530 0.8081
No log 1.4103 110 0.6404 0.5135 0.6404 0.8002
No log 1.4359 112 0.8209 0.4444 0.8209 0.9061
No log 1.4615 114 0.9126 0.4077 0.9126 0.9553
No log 1.4872 116 0.7673 0.4404 0.7673 0.8760
No log 1.5128 118 0.6861 0.5077 0.6861 0.8283
No log 1.5385 120 0.8210 0.4927 0.8209 0.9061
No log 1.5641 122 0.9520 0.4229 0.9520 0.9757
No log 1.5897 124 0.7999 0.5312 0.7999 0.8944
No log 1.6154 126 0.7159 0.5006 0.7159 0.8461
No log 1.6410 128 0.7888 0.4118 0.7888 0.8881
No log 1.6667 130 0.7152 0.4618 0.7152 0.8457
No log 1.6923 132 0.6499 0.4610 0.6499 0.8061
No log 1.7179 134 0.8814 0.4209 0.8814 0.9388
No log 1.7436 136 0.9923 0.4004 0.9923 0.9961
No log 1.7692 138 0.8358 0.4301 0.8358 0.9142
No log 1.7949 140 0.6478 0.4817 0.6478 0.8049
No log 1.8205 142 0.6362 0.4683 0.6362 0.7976
No log 1.8462 144 0.6350 0.5126 0.6350 0.7968
No log 1.8718 146 0.6588 0.4712 0.6588 0.8117
No log 1.8974 148 0.7985 0.4336 0.7985 0.8936
No log 1.9231 150 1.0670 0.3548 1.0670 1.0329
No log 1.9487 152 1.0504 0.4031 1.0504 1.0249
No log 1.9744 154 0.7867 0.4614 0.7867 0.8869
No log 2.0 156 0.6853 0.5143 0.6853 0.8278
No log 2.0256 158 0.7575 0.4539 0.7575 0.8704
No log 2.0513 160 0.6949 0.5074 0.6949 0.8336
No log 2.0769 162 0.6222 0.4998 0.6222 0.7888
No log 2.1026 164 0.7379 0.4399 0.7379 0.8590
No log 2.1282 166 0.8758 0.4301 0.8758 0.9358
No log 2.1538 168 0.9882 0.4031 0.9882 0.9941
No log 2.1795 170 0.8323 0.4475 0.8323 0.9123
No log 2.2051 172 0.6634 0.5018 0.6634 0.8145
No log 2.2308 174 0.6665 0.4542 0.6665 0.8164
No log 2.2564 176 0.6621 0.4542 0.6621 0.8137
No log 2.2821 178 0.6366 0.4930 0.6366 0.7978
No log 2.3077 180 0.6546 0.4551 0.6546 0.8090
No log 2.3333 182 0.6637 0.4983 0.6637 0.8147
No log 2.3590 184 0.6542 0.4895 0.6542 0.8088
No log 2.3846 186 0.6664 0.5037 0.6664 0.8164
No log 2.4103 188 0.6821 0.5272 0.6821 0.8259
No log 2.4359 190 0.7118 0.5171 0.7118 0.8437
No log 2.4615 192 0.7557 0.5090 0.7557 0.8693
No log 2.4872 194 0.7042 0.5060 0.7042 0.8392
No log 2.5128 196 0.6711 0.5169 0.6711 0.8192
No log 2.5385 198 0.6573 0.4985 0.6573 0.8107
No log 2.5641 200 0.6574 0.4832 0.6574 0.8108
No log 2.5897 202 0.6738 0.4476 0.6738 0.8208
No log 2.6154 204 0.6893 0.5010 0.6893 0.8302
No log 2.6410 206 0.7302 0.4817 0.7302 0.8545
No log 2.6667 208 0.7643 0.4087 0.7643 0.8742
No log 2.6923 210 0.7015 0.4899 0.7015 0.8375
No log 2.7179 212 0.7545 0.5003 0.7545 0.8686
No log 2.7436 214 0.8563 0.4325 0.8563 0.9253
No log 2.7692 216 0.7688 0.5190 0.7688 0.8768
No log 2.7949 218 0.6519 0.4296 0.6519 0.8074
No log 2.8205 220 0.6386 0.4443 0.6386 0.7991
No log 2.8462 222 0.6354 0.4133 0.6354 0.7971
No log 2.8718 224 0.6355 0.3982 0.6355 0.7972
No log 2.8974 226 0.6688 0.4700 0.6688 0.8178
No log 2.9231 228 0.6561 0.4871 0.6561 0.8100
No log 2.9487 230 0.6652 0.4282 0.6652 0.8156
No log 2.9744 232 0.7719 0.4501 0.7719 0.8786
No log 3.0 234 0.8142 0.4403 0.8142 0.9023
No log 3.0256 236 0.7294 0.4486 0.7294 0.8541
No log 3.0513 238 0.7032 0.5498 0.7032 0.8386
No log 3.0769 240 0.7425 0.5353 0.7425 0.8617
No log 3.1026 242 0.7575 0.5191 0.7575 0.8703
No log 3.1282 244 0.7052 0.5142 0.7052 0.8397
No log 3.1538 246 0.6793 0.5271 0.6793 0.8242
No log 3.1795 248 0.6517 0.4655 0.6517 0.8073
No log 3.2051 250 0.6502 0.4299 0.6502 0.8063
No log 3.2308 252 0.7126 0.4765 0.7126 0.8442
No log 3.2564 254 0.7048 0.4765 0.7048 0.8395
No log 3.2821 256 0.6975 0.4678 0.6975 0.8352
No log 3.3077 258 0.6504 0.4359 0.6504 0.8065
No log 3.3333 260 0.6572 0.4657 0.6572 0.8107
No log 3.3590 262 0.6570 0.4796 0.6570 0.8106
No log 3.3846 264 0.6350 0.4229 0.6350 0.7969
No log 3.4103 266 0.6639 0.4488 0.6639 0.8148
No log 3.4359 268 0.8447 0.4495 0.8447 0.9191
No log 3.4615 270 0.8754 0.4345 0.8754 0.9356
No log 3.4872 272 0.7471 0.5484 0.7471 0.8644
No log 3.5128 274 0.6430 0.5017 0.6430 0.8019
No log 3.5385 276 0.6447 0.5146 0.6447 0.8029
No log 3.5641 278 0.6384 0.5175 0.6384 0.7990
No log 3.5897 280 0.6091 0.5012 0.6091 0.7805
No log 3.6154 282 0.5988 0.5053 0.5988 0.7738
No log 3.6410 284 0.6071 0.4352 0.6071 0.7792
No log 3.6667 286 0.6496 0.5195 0.6496 0.8060
No log 3.6923 288 0.6506 0.5003 0.6506 0.8066
No log 3.7179 290 0.6250 0.4564 0.6250 0.7906
No log 3.7436 292 0.6340 0.5025 0.6340 0.7962
No log 3.7692 294 0.6274 0.4830 0.6274 0.7921
No log 3.7949 296 0.6093 0.4632 0.6093 0.7806
No log 3.8205 298 0.6483 0.5208 0.6483 0.8052
No log 3.8462 300 0.6949 0.5223 0.6949 0.8336
No log 3.8718 302 0.6636 0.5329 0.6636 0.8146
No log 3.8974 304 0.6160 0.5129 0.6160 0.7849
No log 3.9231 306 0.6641 0.5043 0.6641 0.8149
No log 3.9487 308 0.7218 0.5167 0.7218 0.8496
No log 3.9744 310 0.6944 0.5 0.6944 0.8333
No log 4.0 312 0.6582 0.5011 0.6582 0.8113
No log 4.0256 314 0.6934 0.4966 0.6934 0.8327
No log 4.0513 316 0.7495 0.5293 0.7495 0.8657
No log 4.0769 318 0.7162 0.5286 0.7162 0.8463
No log 4.1026 320 0.6489 0.4721 0.6489 0.8055
No log 4.1282 322 0.6209 0.4591 0.6209 0.7880
No log 4.1538 324 0.6196 0.4702 0.6196 0.7871
No log 4.1795 326 0.6134 0.4601 0.6134 0.7832
No log 4.2051 328 0.6065 0.4467 0.6065 0.7788
No log 4.2308 330 0.5879 0.5422 0.5879 0.7667
No log 4.2564 332 0.5903 0.5663 0.5903 0.7683
No log 4.2821 334 0.5766 0.5150 0.5766 0.7593
No log 4.3077 336 0.5798 0.5150 0.5798 0.7614
No log 4.3333 338 0.5972 0.4906 0.5972 0.7728
No log 4.3590 340 0.8061 0.5089 0.8061 0.8978
No log 4.3846 342 1.0248 0.3831 1.0248 1.0123
No log 4.4103 344 0.9511 0.4374 0.9511 0.9752
No log 4.4359 346 0.7146 0.5816 0.7146 0.8453
No log 4.4615 348 0.6310 0.5194 0.6310 0.7944
No log 4.4872 350 0.6450 0.5211 0.6450 0.8031
No log 4.5128 352 0.6327 0.5379 0.6327 0.7954
No log 4.5385 354 0.6296 0.5724 0.6296 0.7934
No log 4.5641 356 0.6755 0.5885 0.6755 0.8219
No log 4.5897 358 0.7163 0.5187 0.7163 0.8464
No log 4.6154 360 0.6753 0.5181 0.6753 0.8218
No log 4.6410 362 0.6261 0.5880 0.6261 0.7912
No log 4.6667 364 0.6246 0.5787 0.6246 0.7903
No log 4.6923 366 0.6178 0.5188 0.6178 0.7860
No log 4.7179 368 0.5835 0.5778 0.5835 0.7639
No log 4.7436 370 0.5802 0.4990 0.5802 0.7617
No log 4.7692 372 0.6173 0.4534 0.6173 0.7857
No log 4.7949 374 0.5947 0.4227 0.5947 0.7712
No log 4.8205 376 0.5724 0.4341 0.5724 0.7566
No log 4.8462 378 0.6133 0.4886 0.6133 0.7831
No log 4.8718 380 0.6509 0.4728 0.6509 0.8068
No log 4.8974 382 0.6302 0.4791 0.6302 0.7938
No log 4.9231 384 0.6178 0.4979 0.6178 0.7860
No log 4.9487 386 0.6159 0.5462 0.6159 0.7848
No log 4.9744 388 0.6156 0.5713 0.6156 0.7846
No log 5.0 390 0.5978 0.5195 0.5978 0.7732
No log 5.0256 392 0.5821 0.4664 0.5821 0.7629
No log 5.0513 394 0.5741 0.4585 0.5741 0.7577
No log 5.0769 396 0.5724 0.4575 0.5724 0.7565
No log 5.1026 398 0.5780 0.4970 0.5780 0.7603
No log 5.1282 400 0.5833 0.5631 0.5833 0.7637
No log 5.1538 402 0.6153 0.5455 0.6153 0.7844
No log 5.1795 404 0.6952 0.5276 0.6952 0.8338
No log 5.2051 406 0.6602 0.5402 0.6602 0.8125
No log 5.2308 408 0.5906 0.5856 0.5906 0.7685
No log 5.2564 410 0.5925 0.5310 0.5925 0.7698
No log 5.2821 412 0.5698 0.5592 0.5698 0.7549
No log 5.3077 414 0.5847 0.4621 0.5847 0.7646
No log 5.3333 416 0.6291 0.4736 0.6291 0.7932
No log 5.3590 418 0.6042 0.4598 0.6042 0.7773
No log 5.3846 420 0.5804 0.4623 0.5804 0.7618
No log 5.4103 422 0.5803 0.5019 0.5803 0.7618
No log 5.4359 424 0.6022 0.5374 0.6022 0.7760
No log 5.4615 426 0.6450 0.5727 0.6450 0.8031
No log 5.4872 428 0.7258 0.5414 0.7258 0.8520
No log 5.5128 430 0.7430 0.5724 0.7430 0.8620
No log 5.5385 432 0.7179 0.5396 0.7179 0.8473
No log 5.5641 434 0.7184 0.5005 0.7184 0.8476
No log 5.5897 436 0.6935 0.4867 0.6935 0.8327
No log 5.6154 438 0.6883 0.4917 0.6883 0.8296
No log 5.6410 440 0.6693 0.4744 0.6693 0.8181
No log 5.6667 442 0.6964 0.5035 0.6964 0.8345
No log 5.6923 444 0.7166 0.5074 0.7166 0.8465
No log 5.7179 446 0.7966 0.5067 0.7966 0.8926
No log 5.7436 448 0.8322 0.4261 0.8322 0.9123
No log 5.7692 450 0.7668 0.4198 0.7668 0.8757
No log 5.7949 452 0.6913 0.3735 0.6913 0.8314
No log 5.8205 454 0.6394 0.4005 0.6394 0.7996
No log 5.8462 456 0.6444 0.4920 0.6444 0.8027
No log 5.8718 458 0.6520 0.4920 0.6520 0.8074
No log 5.8974 460 0.6373 0.4740 0.6373 0.7983
No log 5.9231 462 0.6721 0.4609 0.6721 0.8198
No log 5.9487 464 0.7174 0.4409 0.7174 0.8470
No log 5.9744 466 0.7351 0.4331 0.7351 0.8574
No log 6.0 468 0.6812 0.5 0.6812 0.8254
No log 6.0256 470 0.6333 0.5105 0.6333 0.7958
No log 6.0513 472 0.6238 0.5203 0.6238 0.7898
No log 6.0769 474 0.6349 0.4667 0.6349 0.7968
No log 6.1026 476 0.6862 0.4810 0.6862 0.8284
No log 6.1282 478 0.6816 0.5012 0.6816 0.8256
No log 6.1538 480 0.6311 0.4237 0.6311 0.7944
No log 6.1795 482 0.6002 0.4433 0.6002 0.7748
No log 6.2051 484 0.6094 0.4869 0.6094 0.7806
No log 6.2308 486 0.5990 0.4497 0.5990 0.7739
No log 6.2564 488 0.6189 0.4749 0.6189 0.7867
No log 6.2821 490 0.7245 0.4493 0.7245 0.8511
No log 6.3077 492 0.8030 0.4429 0.8030 0.8961
No log 6.3333 494 0.7800 0.4475 0.7800 0.8832
No log 6.3590 496 0.7057 0.5063 0.7057 0.8400
No log 6.3846 498 0.6973 0.5152 0.6973 0.8350
0.3895 6.4103 500 0.7036 0.4822 0.7036 0.8388
0.3895 6.4359 502 0.7003 0.4838 0.7003 0.8369
0.3895 6.4615 504 0.7101 0.3854 0.7101 0.8427
0.3895 6.4872 506 0.7319 0.3684 0.7319 0.8555
0.3895 6.5128 508 0.6808 0.3553 0.6808 0.8251
0.3895 6.5385 510 0.6485 0.4428 0.6485 0.8053
0.3895 6.5641 512 0.6211 0.4252 0.6211 0.7881

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
6
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k15_task2_organization

Finetuned
(4222)
this model