ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k9_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6213
  • Qwk: 0.4444
  • Mse: 0.6213
  • Rmse: 0.7882

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0426 2 4.2940 -0.0240 4.2940 2.0722
No log 0.0851 4 2.1996 0.0227 2.1996 1.4831
No log 0.1277 6 1.2778 0.0361 1.2778 1.1304
No log 0.1702 8 1.0320 0.0128 1.0320 1.0159
No log 0.2128 10 0.8502 0.1833 0.8502 0.9221
No log 0.2553 12 0.7788 0.1140 0.7788 0.8825
No log 0.2979 14 0.9022 0.1321 0.9022 0.9499
No log 0.3404 16 1.2240 0.0 1.2240 1.1063
No log 0.3830 18 1.1183 0.0205 1.1183 1.0575
No log 0.4255 20 1.1744 -0.0116 1.1744 1.0837
No log 0.4681 22 1.1505 -0.0175 1.1505 1.0726
No log 0.5106 24 0.9106 0.2267 0.9106 0.9542
No log 0.5532 26 0.8668 0.2267 0.8668 0.9310
No log 0.5957 28 0.8445 0.2859 0.8445 0.9190
No log 0.6383 30 0.8508 0.2267 0.8508 0.9224
No log 0.6809 32 0.8025 0.3140 0.8025 0.8958
No log 0.7234 34 0.7830 0.3096 0.7830 0.8849
No log 0.7660 36 0.7781 0.3030 0.7781 0.8821
No log 0.8085 38 0.8751 0.2075 0.8751 0.9354
No log 0.8511 40 0.8501 0.2122 0.8501 0.9220
No log 0.8936 42 0.8445 0.2122 0.8445 0.9190
No log 0.9362 44 1.0070 0.1996 1.0070 1.0035
No log 0.9787 46 0.9945 0.1892 0.9945 0.9972
No log 1.0213 48 0.8695 0.2583 0.8695 0.9325
No log 1.0638 50 0.8173 0.3139 0.8173 0.9040
No log 1.1064 52 0.8032 0.3117 0.8032 0.8962
No log 1.1489 54 0.8043 0.3011 0.8043 0.8968
No log 1.1915 56 0.8232 0.2927 0.8232 0.9073
No log 1.2340 58 0.8785 0.1676 0.8785 0.9373
No log 1.2766 60 0.9037 0.1242 0.9037 0.9506
No log 1.3191 62 0.9784 0.1135 0.9784 0.9891
No log 1.3617 64 0.9726 0.1132 0.9726 0.9862
No log 1.4043 66 1.0216 0.0924 1.0216 1.0107
No log 1.4468 68 1.0131 0.1334 1.0131 1.0065
No log 1.4894 70 0.8271 0.2545 0.8271 0.9094
No log 1.5319 72 0.6669 0.4179 0.6669 0.8166
No log 1.5745 74 0.6350 0.4824 0.6350 0.7969
No log 1.6170 76 0.6680 0.4692 0.6680 0.8173
No log 1.6596 78 0.7392 0.3954 0.7392 0.8598
No log 1.7021 80 0.8193 0.3283 0.8193 0.9051
No log 1.7447 82 0.7965 0.3587 0.7965 0.8925
No log 1.7872 84 0.7277 0.4263 0.7277 0.8531
No log 1.8298 86 0.5967 0.4626 0.5967 0.7724
No log 1.8723 88 0.5653 0.5027 0.5653 0.7518
No log 1.9149 90 0.5731 0.4289 0.5731 0.7570
No log 1.9574 92 0.5826 0.4513 0.5826 0.7633
No log 2.0 94 0.6307 0.4496 0.6307 0.7942
No log 2.0426 96 0.5992 0.4740 0.5992 0.7741
No log 2.0851 98 0.5744 0.5391 0.5744 0.7579
No log 2.1277 100 0.5764 0.5329 0.5764 0.7592
No log 2.1702 102 0.6227 0.5247 0.6227 0.7891
No log 2.2128 104 0.6920 0.5112 0.6920 0.8318
No log 2.2553 106 0.6959 0.4867 0.6959 0.8342
No log 2.2979 108 0.6371 0.5500 0.6371 0.7982
No log 2.3404 110 0.6623 0.5496 0.6623 0.8138
No log 2.3830 112 0.6392 0.5341 0.6392 0.7995
No log 2.4255 114 0.5962 0.5209 0.5962 0.7721
No log 2.4681 116 0.6419 0.4754 0.6419 0.8012
No log 2.5106 118 0.9088 0.3408 0.9088 0.9533
No log 2.5532 120 0.9332 0.3079 0.9332 0.9660
No log 2.5957 122 0.6915 0.4406 0.6915 0.8316
No log 2.6383 124 0.6043 0.4878 0.6043 0.7774
No log 2.6809 126 0.6338 0.5236 0.6338 0.7961
No log 2.7234 128 0.6292 0.5100 0.6292 0.7932
No log 2.7660 130 0.7555 0.4949 0.7555 0.8692
No log 2.8085 132 0.7946 0.4303 0.7946 0.8914
No log 2.8511 134 0.8824 0.4106 0.8824 0.9394
No log 2.8936 136 0.7224 0.5303 0.7224 0.8499
No log 2.9362 138 0.6603 0.5296 0.6603 0.8126
No log 2.9787 140 0.6367 0.4750 0.6367 0.7979
No log 3.0213 142 0.6185 0.4349 0.6185 0.7865
No log 3.0638 144 0.6285 0.4469 0.6285 0.7928
No log 3.1064 146 0.6611 0.4308 0.6611 0.8131
No log 3.1489 148 0.5928 0.5148 0.5928 0.7699
No log 3.1915 150 0.6401 0.4424 0.6401 0.8001
No log 3.2340 152 0.7774 0.4081 0.7774 0.8817
No log 3.2766 154 0.8070 0.4264 0.8070 0.8984
No log 3.3191 156 0.6427 0.4792 0.6427 0.8017
No log 3.3617 158 0.6283 0.5356 0.6283 0.7927
No log 3.4043 160 0.7900 0.4222 0.7900 0.8888
No log 3.4468 162 0.8019 0.4475 0.8019 0.8955
No log 3.4894 164 0.6526 0.5080 0.6526 0.8078
No log 3.5319 166 0.6230 0.5040 0.6230 0.7893
No log 3.5745 168 0.6512 0.4954 0.6512 0.8070
No log 3.6170 170 0.6245 0.5393 0.6245 0.7902
No log 3.6596 172 0.7220 0.4666 0.7220 0.8497
No log 3.7021 174 0.7629 0.5038 0.7629 0.8734
No log 3.7447 176 0.6632 0.5237 0.6632 0.8143
No log 3.7872 178 0.6519 0.5507 0.6519 0.8074
No log 3.8298 180 0.6577 0.5724 0.6577 0.8110
No log 3.8723 182 0.6961 0.4755 0.6961 0.8343
No log 3.9149 184 0.7599 0.4208 0.7599 0.8717
No log 3.9574 186 0.6826 0.4795 0.6826 0.8262
No log 4.0 188 0.6551 0.5126 0.6551 0.8094
No log 4.0426 190 0.6879 0.4965 0.6879 0.8294
No log 4.0851 192 0.6716 0.5069 0.6716 0.8195
No log 4.1277 194 0.6580 0.5147 0.6580 0.8112
No log 4.1702 196 0.6823 0.4889 0.6823 0.8260
No log 4.2128 198 0.6661 0.4635 0.6661 0.8162
No log 4.2553 200 0.6743 0.4736 0.6743 0.8212
No log 4.2979 202 0.6815 0.4603 0.6815 0.8255
No log 4.3404 204 0.6583 0.4576 0.6583 0.8114
No log 4.3830 206 0.6719 0.4334 0.6719 0.8197
No log 4.4255 208 0.6611 0.4351 0.6611 0.8131
No log 4.4681 210 0.6340 0.3933 0.6340 0.7963
No log 4.5106 212 0.6526 0.4520 0.6526 0.8078
No log 4.5532 214 0.6503 0.4612 0.6503 0.8064
No log 4.5957 216 0.6093 0.4212 0.6093 0.7806
No log 4.6383 218 0.6457 0.3569 0.6457 0.8036
No log 4.6809 220 0.7381 0.3397 0.7381 0.8592
No log 4.7234 222 0.7279 0.2695 0.7279 0.8532
No log 4.7660 224 0.6895 0.2732 0.6895 0.8304
No log 4.8085 226 0.6398 0.3205 0.6398 0.7999
No log 4.8511 228 0.6535 0.4411 0.6535 0.8084
No log 4.8936 230 0.7089 0.4110 0.7089 0.8419
No log 4.9362 232 0.7972 0.3999 0.7972 0.8929
No log 4.9787 234 0.7253 0.4337 0.7253 0.8517
No log 5.0213 236 0.6180 0.4223 0.6180 0.7861
No log 5.0638 238 0.5938 0.4659 0.5938 0.7706
No log 5.1064 240 0.6197 0.4605 0.6197 0.7872
No log 5.1489 242 0.6194 0.4605 0.6194 0.7870
No log 5.1915 244 0.5857 0.5109 0.5857 0.7653
No log 5.2340 246 0.6318 0.4835 0.6318 0.7949
No log 5.2766 248 0.8165 0.4475 0.8165 0.9036
No log 5.3191 250 0.8488 0.4539 0.8488 0.9213
No log 5.3617 252 0.7171 0.4820 0.7171 0.8468
No log 5.4043 254 0.6160 0.5243 0.6160 0.7848
No log 5.4468 256 0.5941 0.4788 0.5941 0.7708
No log 5.4894 258 0.5874 0.4489 0.5874 0.7664
No log 5.5319 260 0.5950 0.4805 0.5950 0.7713
No log 5.5745 262 0.6248 0.5176 0.6248 0.7904
No log 5.6170 264 0.6033 0.4537 0.6033 0.7767
No log 5.6596 266 0.6038 0.4888 0.6038 0.7771
No log 5.7021 268 0.6436 0.5170 0.6436 0.8023
No log 5.7447 270 0.7202 0.4411 0.7202 0.8487
No log 5.7872 272 0.6862 0.5093 0.6862 0.8284
No log 5.8298 274 0.6242 0.5074 0.6242 0.7901
No log 5.8723 276 0.6119 0.4960 0.6119 0.7822
No log 5.9149 278 0.6142 0.5048 0.6142 0.7837
No log 5.9574 280 0.6099 0.5201 0.6099 0.7810
No log 6.0 282 0.6121 0.5283 0.6121 0.7824
No log 6.0426 284 0.6084 0.5221 0.6084 0.7800
No log 6.0851 286 0.5981 0.4875 0.5981 0.7734
No log 6.1277 288 0.5915 0.4793 0.5915 0.7691
No log 6.1702 290 0.5965 0.4739 0.5965 0.7723
No log 6.2128 292 0.5912 0.4623 0.5912 0.7689
No log 6.2553 294 0.5937 0.4427 0.5937 0.7705
No log 6.2979 296 0.6072 0.3926 0.6072 0.7792
No log 6.3404 298 0.6130 0.4439 0.6130 0.7829
No log 6.3830 300 0.6310 0.4262 0.6310 0.7943
No log 6.4255 302 0.6740 0.5010 0.6740 0.8210
No log 6.4681 304 0.6686 0.4565 0.6686 0.8177
No log 6.5106 306 0.6636 0.4117 0.6636 0.8146
No log 6.5532 308 0.7837 0.3898 0.7837 0.8853
No log 6.5957 310 0.8428 0.3898 0.8428 0.9181
No log 6.6383 312 0.8194 0.3593 0.8194 0.9052
No log 6.6809 314 0.7468 0.3548 0.7468 0.8641
No log 6.7234 316 0.6742 0.3859 0.6742 0.8211
No log 6.7660 318 0.6674 0.3729 0.6674 0.8169
No log 6.8085 320 0.6630 0.3729 0.6630 0.8143
No log 6.8511 322 0.6605 0.4120 0.6605 0.8127
No log 6.8936 324 0.6703 0.4214 0.6703 0.8187
No log 6.9362 326 0.6836 0.4175 0.6836 0.8268
No log 6.9787 328 0.6828 0.3907 0.6828 0.8263
No log 7.0213 330 0.6523 0.3933 0.6523 0.8077
No log 7.0638 332 0.6550 0.3886 0.6550 0.8093
No log 7.1064 334 0.6549 0.3948 0.6549 0.8093
No log 7.1489 336 0.6529 0.4081 0.6529 0.8080
No log 7.1915 338 0.6566 0.4465 0.6566 0.8103
No log 7.2340 340 0.6846 0.4161 0.6846 0.8274
No log 7.2766 342 0.6965 0.4198 0.6965 0.8346
No log 7.3191 344 0.7323 0.3833 0.7323 0.8557
No log 7.3617 346 0.7449 0.4417 0.7449 0.8631
No log 7.4043 348 0.6932 0.4421 0.6932 0.8326
No log 7.4468 350 0.6737 0.4060 0.6737 0.8208
No log 7.4894 352 0.6674 0.3998 0.6674 0.8169
No log 7.5319 354 0.6514 0.3702 0.6514 0.8071
No log 7.5745 356 0.6800 0.4427 0.6800 0.8246
No log 7.6170 358 0.6744 0.4252 0.6744 0.8212
No log 7.6596 360 0.6420 0.3747 0.6420 0.8012
No log 7.7021 362 0.6713 0.4019 0.6713 0.8193
No log 7.7447 364 0.7185 0.4073 0.7185 0.8477
No log 7.7872 366 0.7058 0.4452 0.7058 0.8401
No log 7.8298 368 0.6727 0.4086 0.6727 0.8202
No log 7.8723 370 0.6561 0.4301 0.6561 0.8100
No log 7.9149 372 0.6463 0.4059 0.6463 0.8039
No log 7.9574 374 0.6543 0.4222 0.6543 0.8089
No log 8.0 376 0.6514 0.4356 0.6514 0.8071
No log 8.0426 378 0.6386 0.4200 0.6386 0.7991
No log 8.0851 380 0.6377 0.4102 0.6377 0.7986
No log 8.1277 382 0.6420 0.4174 0.6420 0.8012
No log 8.1702 384 0.6478 0.4235 0.6478 0.8049
No log 8.2128 386 0.6430 0.4196 0.6430 0.8018
No log 8.2553 388 0.6428 0.4131 0.6428 0.8017
No log 8.2979 390 0.6408 0.4204 0.6408 0.8005
No log 8.3404 392 0.6376 0.3896 0.6376 0.7985
No log 8.3830 394 0.6531 0.3759 0.6531 0.8081
No log 8.4255 396 0.6525 0.3745 0.6525 0.8078
No log 8.4681 398 0.6423 0.3999 0.6423 0.8015
No log 8.5106 400 0.6461 0.3605 0.6461 0.8038
No log 8.5532 402 0.6515 0.3887 0.6515 0.8072
No log 8.5957 404 0.6564 0.4377 0.6564 0.8102
No log 8.6383 406 0.6439 0.4156 0.6439 0.8024
No log 8.6809 408 0.6479 0.3934 0.6479 0.8049
No log 8.7234 410 0.6524 0.4077 0.6524 0.8077
No log 8.7660 412 0.6384 0.3380 0.6384 0.7990
No log 8.8085 414 0.6404 0.3433 0.6404 0.8002
No log 8.8511 416 0.6395 0.3512 0.6395 0.7997
No log 8.8936 418 0.6516 0.4310 0.6516 0.8072
No log 8.9362 420 0.7281 0.4388 0.7281 0.8533
No log 8.9787 422 0.7658 0.4448 0.7658 0.8751
No log 9.0213 424 0.7172 0.4500 0.7172 0.8469
No log 9.0638 426 0.6581 0.5005 0.6581 0.8113
No log 9.1064 428 0.6660 0.4569 0.6660 0.8161
No log 9.1489 430 0.6670 0.4425 0.6670 0.8167
No log 9.1915 432 0.6637 0.4431 0.6637 0.8147
No log 9.2340 434 0.6683 0.4 0.6683 0.8175
No log 9.2766 436 0.6566 0.4573 0.6566 0.8103
No log 9.3191 438 0.6721 0.4 0.6721 0.8198
No log 9.3617 440 0.7021 0.4260 0.7021 0.8379
No log 9.4043 442 0.7671 0.4576 0.7671 0.8758
No log 9.4468 444 0.7410 0.4510 0.7410 0.8608
No log 9.4894 446 0.7258 0.4853 0.7258 0.8520
No log 9.5319 448 0.7792 0.4657 0.7792 0.8827
No log 9.5745 450 0.7743 0.4934 0.7743 0.8799
No log 9.6170 452 0.7020 0.4878 0.7020 0.8378
No log 9.6596 454 0.6834 0.5041 0.6834 0.8267
No log 9.7021 456 0.6811 0.4585 0.6811 0.8253
No log 9.7447 458 0.6647 0.4073 0.6647 0.8153
No log 9.7872 460 0.6321 0.4473 0.6321 0.7950
No log 9.8298 462 0.6323 0.4506 0.6323 0.7952
No log 9.8723 464 0.6369 0.4874 0.6369 0.7980
No log 9.9149 466 0.6376 0.4884 0.6376 0.7985
No log 9.9574 468 0.6617 0.4299 0.6617 0.8134
No log 10.0 470 0.7284 0.4515 0.7284 0.8535
No log 10.0426 472 0.7610 0.4054 0.7610 0.8723
No log 10.0851 474 0.7008 0.4381 0.7008 0.8371
No log 10.1277 476 0.6399 0.4814 0.6399 0.8000
No log 10.1702 478 0.6454 0.5134 0.6454 0.8034
No log 10.2128 480 0.6372 0.5097 0.6372 0.7983
No log 10.2553 482 0.6360 0.4316 0.6360 0.7975
No log 10.2979 484 0.6688 0.4476 0.6688 0.8178
No log 10.3404 486 0.6582 0.4499 0.6582 0.8113
No log 10.3830 488 0.6250 0.4620 0.6250 0.7905
No log 10.4255 490 0.6398 0.4805 0.6398 0.7999
No log 10.4681 492 0.6849 0.4697 0.6849 0.8276
No log 10.5106 494 0.6891 0.4630 0.6891 0.8301
No log 10.5532 496 0.6436 0.5100 0.6436 0.8023
No log 10.5957 498 0.6317 0.4848 0.6317 0.7948
0.3669 10.6383 500 0.6351 0.4935 0.6351 0.7969
0.3669 10.6809 502 0.6312 0.4859 0.6312 0.7945
0.3669 10.7234 504 0.6223 0.4887 0.6223 0.7889
0.3669 10.7660 506 0.6248 0.4843 0.6248 0.7904
0.3669 10.8085 508 0.6431 0.5193 0.6431 0.8019
0.3669 10.8511 510 0.6311 0.5004 0.6311 0.7944
0.3669 10.8936 512 0.6195 0.4286 0.6195 0.7871
0.3669 10.9362 514 0.6146 0.4506 0.6146 0.7840
0.3669 10.9787 516 0.6136 0.4163 0.6136 0.7834
0.3669 11.0213 518 0.6188 0.4076 0.6188 0.7866
0.3669 11.0638 520 0.6213 0.4444 0.6213 0.7882

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
6
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k9_task2_organization

Finetuned
(4222)
this model