ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k10_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6844
  • Qwk: 0.4299
  • Mse: 0.6844
  • Rmse: 0.8273

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0377 2 4.2204 -0.0141 4.2204 2.0544
No log 0.0755 4 2.3438 0.0747 2.3438 1.5309
No log 0.1132 6 1.9534 -0.0359 1.9534 1.3976
No log 0.1509 8 1.5515 0.0061 1.5515 1.2456
No log 0.1887 10 1.5735 -0.0321 1.5735 1.2544
No log 0.2264 12 1.8069 -0.1025 1.8069 1.3442
No log 0.2642 14 1.7228 -0.0776 1.7228 1.3126
No log 0.3019 16 1.1867 0.0339 1.1867 1.0894
No log 0.3396 18 1.0190 0.0296 1.0190 1.0094
No log 0.3774 20 1.0160 0.0408 1.0160 1.0080
No log 0.4151 22 0.9917 0.1170 0.9917 0.9959
No log 0.4528 24 0.7645 0.3283 0.7645 0.8743
No log 0.4906 26 0.7364 0.2818 0.7364 0.8581
No log 0.5283 28 0.8240 0.2620 0.8240 0.9078
No log 0.5660 30 0.7949 0.2620 0.7949 0.8916
No log 0.6038 32 0.8790 0.2492 0.8790 0.9375
No log 0.6415 34 1.7811 0.1303 1.7811 1.3346
No log 0.6792 36 2.7615 0.0918 2.7615 1.6618
No log 0.7170 38 2.1983 0.1180 2.1983 1.4827
No log 0.7547 40 1.1532 0.3212 1.1532 1.0739
No log 0.7925 42 0.7626 0.2919 0.7626 0.8733
No log 0.8302 44 0.6893 0.3812 0.6893 0.8302
No log 0.8679 46 0.6754 0.4148 0.6754 0.8218
No log 0.9057 48 0.7908 0.3292 0.7908 0.8893
No log 0.9434 50 0.9290 0.1349 0.9290 0.9639
No log 0.9811 52 0.8995 0.1857 0.8995 0.9484
No log 1.0189 54 0.7703 0.3340 0.7703 0.8777
No log 1.0566 56 0.6653 0.3425 0.6653 0.8157
No log 1.0943 58 0.6693 0.2900 0.6693 0.8181
No log 1.1321 60 0.6754 0.3361 0.6754 0.8218
No log 1.1698 62 0.7725 0.3764 0.7725 0.8789
No log 1.2075 64 1.2004 0.1149 1.2004 1.0956
No log 1.2453 66 1.2931 0.1149 1.2931 1.1371
No log 1.2830 68 1.0779 0.1242 1.0779 1.0382
No log 1.3208 70 0.8488 0.2859 0.8488 0.9213
No log 1.3585 72 0.7250 0.3793 0.7250 0.8515
No log 1.3962 74 0.6443 0.4135 0.6443 0.8027
No log 1.4340 76 0.6563 0.3069 0.6563 0.8101
No log 1.4717 78 0.6383 0.3679 0.6383 0.7989
No log 1.5094 80 0.6408 0.3989 0.6408 0.8005
No log 1.5472 82 1.0071 0.3497 1.0071 1.0035
No log 1.5849 84 1.3337 0.2126 1.3337 1.1549
No log 1.6226 86 1.0818 0.2847 1.0818 1.0401
No log 1.6604 88 0.7015 0.3704 0.7015 0.8376
No log 1.6981 90 0.6564 0.3702 0.6564 0.8102
No log 1.7358 92 0.6819 0.3294 0.6819 0.8258
No log 1.7736 94 0.6703 0.3425 0.6703 0.8187
No log 1.8113 96 0.7473 0.3804 0.7473 0.8644
No log 1.8491 98 0.8536 0.3254 0.8536 0.9239
No log 1.8868 100 1.0022 0.2124 1.0022 1.0011
No log 1.9245 102 1.0108 0.2375 1.0108 1.0054
No log 1.9623 104 1.0184 0.3117 1.0184 1.0092
No log 2.0 106 0.8023 0.3783 0.8023 0.8957
No log 2.0377 108 0.6540 0.4590 0.6540 0.8087
No log 2.0755 110 0.6359 0.5065 0.6359 0.7974
No log 2.1132 112 0.6419 0.4439 0.6419 0.8012
No log 2.1509 114 0.6743 0.4933 0.6743 0.8212
No log 2.1887 116 0.6730 0.5134 0.6730 0.8204
No log 2.2264 118 0.6962 0.5225 0.6962 0.8344
No log 2.2642 120 0.7798 0.4678 0.7798 0.8831
No log 2.3019 122 0.7306 0.4889 0.7306 0.8547
No log 2.3396 124 0.6849 0.4267 0.6849 0.8276
No log 2.3774 126 0.6629 0.4941 0.6629 0.8142
No log 2.4151 128 0.7222 0.4978 0.7222 0.8498
No log 2.4528 130 0.7934 0.4844 0.7934 0.8907
No log 2.4906 132 0.9339 0.4034 0.9339 0.9664
No log 2.5283 134 1.0057 0.3508 1.0057 1.0028
No log 2.5660 136 0.7483 0.4761 0.7483 0.8651
No log 2.6038 138 0.6275 0.5165 0.6275 0.7922
No log 2.6415 140 0.6754 0.4793 0.6754 0.8218
No log 2.6792 142 0.7056 0.4934 0.7056 0.8400
No log 2.7170 144 0.6153 0.4568 0.6153 0.7844
No log 2.7547 146 0.6132 0.5262 0.6132 0.7831
No log 2.7925 148 0.6629 0.4612 0.6629 0.8142
No log 2.8302 150 0.6950 0.4991 0.6950 0.8337
No log 2.8679 152 0.8200 0.4687 0.8200 0.9055
No log 2.9057 154 0.7971 0.4848 0.7971 0.8928
No log 2.9434 156 0.7289 0.5026 0.7289 0.8538
No log 2.9811 158 0.6748 0.4938 0.6748 0.8215
No log 3.0189 160 0.6716 0.5398 0.6716 0.8195
No log 3.0566 162 0.7022 0.5197 0.7022 0.8380
No log 3.0943 164 0.9971 0.3519 0.9971 0.9986
No log 3.1321 166 1.1278 0.3338 1.1278 1.0620
No log 3.1698 168 0.8664 0.3898 0.8664 0.9308
No log 3.2075 170 0.6329 0.4418 0.6329 0.7956
No log 3.2453 172 0.7246 0.4606 0.7246 0.8512
No log 3.2830 174 0.7171 0.4598 0.7171 0.8468
No log 3.3208 176 0.6169 0.3995 0.6169 0.7854
No log 3.3585 178 0.6477 0.4829 0.6477 0.8048
No log 3.3962 180 0.6870 0.4996 0.6870 0.8288
No log 3.4340 182 0.6496 0.5217 0.6496 0.8060
No log 3.4717 184 0.6283 0.5149 0.6283 0.7926
No log 3.5094 186 0.6837 0.4912 0.6837 0.8269
No log 3.5472 188 0.6713 0.4627 0.6713 0.8193
No log 3.5849 190 0.6162 0.5100 0.6162 0.7850
No log 3.6226 192 0.6646 0.4864 0.6646 0.8152
No log 3.6604 194 0.7058 0.5253 0.7058 0.8401
No log 3.6981 196 0.6521 0.5662 0.6521 0.8075
No log 3.7358 198 0.5932 0.5404 0.5932 0.7702
No log 3.7736 200 0.6576 0.5261 0.6576 0.8109
No log 3.8113 202 0.6738 0.5187 0.6738 0.8209
No log 3.8491 204 0.6344 0.5078 0.6344 0.7965
No log 3.8868 206 0.6404 0.5211 0.6404 0.8003
No log 3.9245 208 0.6710 0.5078 0.6710 0.8191
No log 3.9623 210 0.6183 0.5122 0.6183 0.7863
No log 4.0 212 0.5944 0.5584 0.5944 0.7710
No log 4.0377 214 0.5914 0.5624 0.5914 0.7690
No log 4.0755 216 0.5819 0.5019 0.5819 0.7628
No log 4.1132 218 0.6311 0.5104 0.6311 0.7944
No log 4.1509 220 0.7068 0.4824 0.7068 0.8407
No log 4.1887 222 0.6658 0.4853 0.6658 0.8160
No log 4.2264 224 0.6575 0.5080 0.6575 0.8109
No log 4.2642 226 0.6301 0.5383 0.6301 0.7938
No log 4.3019 228 0.6415 0.6134 0.6415 0.8009
No log 4.3396 230 0.6259 0.5817 0.6259 0.7912
No log 4.3774 232 0.6253 0.5751 0.6253 0.7907
No log 4.4151 234 0.6218 0.5241 0.6218 0.7885
No log 4.4528 236 0.5938 0.5323 0.5938 0.7706
No log 4.4906 238 0.5889 0.5549 0.5889 0.7674
No log 4.5283 240 0.5995 0.5259 0.5995 0.7742
No log 4.5660 242 0.7058 0.4608 0.7058 0.8401
No log 4.6038 244 0.9825 0.4382 0.9825 0.9912
No log 4.6415 246 1.0785 0.4116 1.0785 1.0385
No log 4.6792 248 0.8231 0.4508 0.8231 0.9072
No log 4.7170 250 0.6316 0.5062 0.6316 0.7947
No log 4.7547 252 0.5994 0.5235 0.5994 0.7742
No log 4.7925 254 0.5840 0.5516 0.5840 0.7642
No log 4.8302 256 0.5832 0.4984 0.5832 0.7637
No log 4.8679 258 0.6170 0.5123 0.6170 0.7855
No log 4.9057 260 0.6192 0.5171 0.6192 0.7869
No log 4.9434 262 0.5952 0.4888 0.5952 0.7715
No log 4.9811 264 0.6239 0.4904 0.6239 0.7899
No log 5.0189 266 0.6255 0.4834 0.6255 0.7909
No log 5.0566 268 0.6387 0.5334 0.6387 0.7992
No log 5.0943 270 0.7539 0.5007 0.7539 0.8683
No log 5.1321 272 0.7486 0.4127 0.7486 0.8652
No log 5.1698 274 0.6680 0.4474 0.6680 0.8173
No log 5.2075 276 0.6356 0.4474 0.6356 0.7973
No log 5.2453 278 0.6601 0.4226 0.6601 0.8124
No log 5.2830 280 0.6381 0.4611 0.6381 0.7988
No log 5.3208 282 0.6605 0.5173 0.6605 0.8127
No log 5.3585 284 0.6874 0.4658 0.6874 0.8291
No log 5.3962 286 0.6159 0.4963 0.6159 0.7848
No log 5.4340 288 0.6020 0.4674 0.6020 0.7759
No log 5.4717 290 0.7295 0.4550 0.7295 0.8541
No log 5.5094 292 0.7469 0.4562 0.7469 0.8642
No log 5.5472 294 0.6272 0.5060 0.6272 0.7920
No log 5.5849 296 0.6024 0.5218 0.6024 0.7762
No log 5.6226 298 0.6047 0.5017 0.6047 0.7776
No log 5.6604 300 0.6065 0.4053 0.6065 0.7788
No log 5.6981 302 0.5980 0.4755 0.5980 0.7733
No log 5.7358 304 0.6121 0.4759 0.6121 0.7824
No log 5.7736 306 0.6226 0.4882 0.6226 0.7890
No log 5.8113 308 0.6701 0.4971 0.6701 0.8186
No log 5.8491 310 0.6810 0.4996 0.6810 0.8252
No log 5.8868 312 0.6513 0.4843 0.6513 0.8070
No log 5.9245 314 0.6432 0.4677 0.6432 0.8020
No log 5.9623 316 0.6509 0.4355 0.6509 0.8068
No log 6.0 318 0.6675 0.4930 0.6675 0.8170
No log 6.0377 320 0.7117 0.5032 0.7117 0.8437
No log 6.0755 322 0.6845 0.4825 0.6845 0.8273
No log 6.1132 324 0.6759 0.5005 0.6759 0.8221
No log 6.1509 326 0.6680 0.5258 0.6680 0.8173
No log 6.1887 328 0.6722 0.4514 0.6722 0.8199
No log 6.2264 330 0.6913 0.4435 0.6913 0.8315
No log 6.2642 332 0.6382 0.4820 0.6382 0.7989
No log 6.3019 334 0.6379 0.4792 0.6379 0.7987
No log 6.3396 336 0.6310 0.5140 0.6310 0.7944
No log 6.3774 338 0.6342 0.4730 0.6342 0.7964
No log 6.4151 340 0.6355 0.4514 0.6355 0.7972
No log 6.4528 342 0.6860 0.4339 0.6860 0.8283
No log 6.4906 344 0.6721 0.4722 0.6721 0.8198
No log 6.5283 346 0.6503 0.4353 0.6503 0.8064
No log 6.5660 348 0.6552 0.4583 0.6552 0.8094
No log 6.6038 350 0.6449 0.4547 0.6449 0.8030
No log 6.6415 352 0.6387 0.4366 0.6387 0.7992
No log 6.6792 354 0.6175 0.4252 0.6175 0.7858
No log 6.7170 356 0.6005 0.4524 0.6005 0.7749
No log 6.7547 358 0.6276 0.4586 0.6276 0.7922
No log 6.7925 360 0.6632 0.4716 0.6632 0.8144
No log 6.8302 362 0.6188 0.4489 0.6188 0.7866
No log 6.8679 364 0.6176 0.4390 0.6176 0.7859
No log 6.9057 366 0.6296 0.4700 0.6296 0.7935
No log 6.9434 368 0.6506 0.4755 0.6506 0.8066
No log 6.9811 370 0.6478 0.4971 0.6478 0.8048
No log 7.0189 372 0.6154 0.4371 0.6154 0.7845
No log 7.0566 374 0.6140 0.4801 0.6140 0.7836
No log 7.0943 376 0.6075 0.4788 0.6075 0.7794
No log 7.1321 378 0.5874 0.4311 0.5874 0.7664
No log 7.1698 380 0.6298 0.4579 0.6298 0.7936
No log 7.2075 382 0.7025 0.4078 0.7025 0.8382
No log 7.2453 384 0.6519 0.4589 0.6519 0.8074
No log 7.2830 386 0.6032 0.4363 0.6032 0.7767
No log 7.3208 388 0.6481 0.5196 0.6481 0.8051
No log 7.3585 390 0.6342 0.5397 0.6342 0.7963
No log 7.3962 392 0.6035 0.4966 0.6035 0.7769
No log 7.4340 394 0.6074 0.5306 0.6074 0.7793
No log 7.4717 396 0.6102 0.5158 0.6102 0.7811
No log 7.5094 398 0.6022 0.5333 0.6022 0.7760
No log 7.5472 400 0.5871 0.4770 0.5871 0.7662
No log 7.5849 402 0.5914 0.4426 0.5914 0.7690
No log 7.6226 404 0.5868 0.4207 0.5868 0.7660
No log 7.6604 406 0.6010 0.4370 0.6010 0.7752
No log 7.6981 408 0.6733 0.4835 0.6733 0.8205
No log 7.7358 410 0.7155 0.4876 0.7155 0.8459
No log 7.7736 412 0.6502 0.4741 0.6502 0.8063
No log 7.8113 414 0.6236 0.4553 0.6236 0.7897
No log 7.8491 416 0.6667 0.4968 0.6667 0.8165
No log 7.8868 418 0.6655 0.4687 0.6655 0.8158
No log 7.9245 420 0.6216 0.3917 0.6216 0.7884
No log 7.9623 422 0.6077 0.4407 0.6077 0.7796
No log 8.0 424 0.6730 0.3978 0.6730 0.8203
No log 8.0377 426 0.6930 0.4078 0.6930 0.8324
No log 8.0755 428 0.6349 0.4298 0.6349 0.7968
No log 8.1132 430 0.5913 0.4177 0.5913 0.7689
No log 8.1509 432 0.5926 0.4406 0.5926 0.7698
No log 8.1887 434 0.5965 0.4539 0.5965 0.7723
No log 8.2264 436 0.6234 0.4334 0.6234 0.7896
No log 8.2642 438 0.6344 0.4542 0.6344 0.7965
No log 8.3019 440 0.6239 0.4551 0.6239 0.7899
No log 8.3396 442 0.6180 0.4592 0.6180 0.7861
No log 8.3774 444 0.6139 0.4271 0.6139 0.7835
No log 8.4151 446 0.6176 0.4241 0.6176 0.7859
No log 8.4528 448 0.6414 0.4472 0.6414 0.8009
No log 8.4906 450 0.6256 0.4337 0.6256 0.7909
No log 8.5283 452 0.6187 0.4529 0.6187 0.7866
No log 8.5660 454 0.6632 0.5039 0.6632 0.8144
No log 8.6038 456 0.6739 0.4872 0.6739 0.8209
No log 8.6415 458 0.6255 0.4872 0.6255 0.7909
No log 8.6792 460 0.6528 0.4481 0.6528 0.8080
No log 8.7170 462 0.7071 0.4371 0.7071 0.8409
No log 8.7547 464 0.6679 0.4399 0.6679 0.8173
No log 8.7925 466 0.6106 0.4302 0.6106 0.7814
No log 8.8302 468 0.6646 0.4882 0.6646 0.8152
No log 8.8679 470 0.6825 0.4945 0.6825 0.8261
No log 8.9057 472 0.6391 0.4273 0.6391 0.7994
No log 8.9434 474 0.6348 0.4528 0.6348 0.7967
No log 8.9811 476 0.6838 0.4357 0.6838 0.8269
No log 9.0189 478 0.6704 0.4275 0.6704 0.8188
No log 9.0566 480 0.6392 0.4174 0.6392 0.7995
No log 9.0943 482 0.6417 0.4601 0.6417 0.8010
No log 9.1321 484 0.6398 0.4291 0.6398 0.7998
No log 9.1698 486 0.6296 0.4038 0.6296 0.7935
No log 9.2075 488 0.6407 0.4447 0.6407 0.8004
No log 9.2453 490 0.6381 0.4514 0.6381 0.7988
No log 9.2830 492 0.6400 0.4514 0.6400 0.8000
No log 9.3208 494 0.6468 0.4579 0.6468 0.8042
No log 9.3585 496 0.6354 0.4465 0.6354 0.7971
No log 9.3962 498 0.6557 0.4476 0.6557 0.8098
0.4152 9.4340 500 0.6660 0.4499 0.6660 0.8161
0.4152 9.4717 502 0.6704 0.4406 0.6704 0.8188
0.4152 9.5094 504 0.7284 0.4844 0.7284 0.8535
0.4152 9.5472 506 0.7666 0.4267 0.7666 0.8756
0.4152 9.5849 508 0.7002 0.4808 0.7002 0.8368
0.4152 9.6226 510 0.6518 0.4188 0.6518 0.8073
0.4152 9.6604 512 0.6437 0.4579 0.6437 0.8023
0.4152 9.6981 514 0.6340 0.4418 0.6340 0.7962
0.4152 9.7358 516 0.6602 0.4211 0.6602 0.8126
0.4152 9.7736 518 0.7127 0.4014 0.7127 0.8442
0.4152 9.8113 520 0.7329 0.4045 0.7329 0.8561
0.4152 9.8491 522 0.6844 0.4299 0.6844 0.8273

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
5
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k10_task2_organization

Finetuned
(4222)
this model