ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k19_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6496
  • Qwk: 0.5093
  • Mse: 0.6496
  • Rmse: 0.8059

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.02 2 4.1587 -0.0232 4.1587 2.0393
No log 0.04 4 1.9922 0.0475 1.9922 1.4114
No log 0.06 6 1.3448 0.0107 1.3448 1.1596
No log 0.08 8 1.0618 0.0038 1.0618 1.0304
No log 0.1 10 1.1816 -0.1194 1.1816 1.0870
No log 0.12 12 1.4062 -0.1269 1.4062 1.1859
No log 0.14 14 1.0683 0.0092 1.0683 1.0336
No log 0.16 16 0.9267 0.2027 0.9267 0.9627
No log 0.18 18 0.8841 0.2237 0.8841 0.9403
No log 0.2 20 0.9374 0.1209 0.9374 0.9682
No log 0.22 22 0.8632 0.2366 0.8632 0.9291
No log 0.24 24 0.8272 0.2097 0.8272 0.9095
No log 0.26 26 0.8992 0.1265 0.8992 0.9482
No log 0.28 28 0.8495 0.2982 0.8495 0.9217
No log 0.3 30 0.8051 0.2469 0.8051 0.8973
No log 0.32 32 0.7971 0.2662 0.7971 0.8928
No log 0.34 34 0.9799 0.1025 0.9799 0.9899
No log 0.36 36 1.1943 0.0628 1.1943 1.0928
No log 0.38 38 1.1500 0.1052 1.1500 1.0724
No log 0.4 40 0.8574 0.2135 0.8574 0.9259
No log 0.42 42 0.7215 0.3276 0.7215 0.8494
No log 0.44 44 0.8193 0.1732 0.8193 0.9051
No log 0.46 46 0.7448 0.2498 0.7448 0.8630
No log 0.48 48 0.7052 0.3361 0.7052 0.8398
No log 0.5 50 0.8272 0.2110 0.8272 0.9095
No log 0.52 52 0.7759 0.3152 0.7759 0.8808
No log 0.54 54 0.7041 0.2923 0.7041 0.8391
No log 0.56 56 0.7568 0.1431 0.7568 0.8699
No log 0.58 58 0.8821 0.0704 0.8821 0.9392
No log 0.6 60 0.8676 0.0784 0.8676 0.9314
No log 0.62 62 0.7573 0.1210 0.7573 0.8702
No log 0.64 64 0.8333 0.2148 0.8333 0.9128
No log 0.66 66 0.9730 0.1186 0.9730 0.9864
No log 0.68 68 1.0129 0.1070 1.0129 1.0064
No log 0.7 70 0.9322 0.1528 0.9322 0.9655
No log 0.72 72 0.8297 0.2085 0.8297 0.9109
No log 0.74 74 0.7801 0.2827 0.7801 0.8833
No log 0.76 76 0.7419 0.2834 0.7419 0.8613
No log 0.78 78 0.7310 0.2253 0.7310 0.8550
No log 0.8 80 0.7332 0.2237 0.7332 0.8562
No log 0.82 82 0.7369 0.2859 0.7369 0.8584
No log 0.84 84 0.7248 0.3319 0.7248 0.8514
No log 0.86 86 0.7744 0.2787 0.7744 0.8800
No log 0.88 88 0.7629 0.2831 0.7629 0.8735
No log 0.9 90 0.7132 0.4879 0.7132 0.8445
No log 0.92 92 0.7347 0.4955 0.7347 0.8571
No log 0.94 94 0.7026 0.5012 0.7026 0.8382
No log 0.96 96 0.6880 0.4965 0.6880 0.8295
No log 0.98 98 0.6889 0.5159 0.6889 0.8300
No log 1.0 100 0.8050 0.4173 0.8050 0.8972
No log 1.02 102 0.9470 0.3643 0.9470 0.9731
No log 1.04 104 1.0043 0.3456 1.0043 1.0021
No log 1.06 106 0.7540 0.4100 0.7540 0.8683
No log 1.08 108 0.6892 0.4426 0.6892 0.8302
No log 1.1 110 0.7383 0.4192 0.7383 0.8593
No log 1.12 112 0.7766 0.3989 0.7766 0.8813
No log 1.1400 114 0.8093 0.3821 0.8093 0.8996
No log 1.16 116 0.8138 0.4334 0.8138 0.9021
No log 1.18 118 0.9039 0.3954 0.9039 0.9507
No log 1.2 120 0.9803 0.3563 0.9803 0.9901
No log 1.22 122 0.8824 0.4841 0.8824 0.9394
No log 1.24 124 0.7652 0.4971 0.7652 0.8747
No log 1.26 126 0.8030 0.4382 0.8030 0.8961
No log 1.28 128 0.7681 0.4044 0.7681 0.8764
No log 1.3 130 0.7967 0.4496 0.7967 0.8926
No log 1.32 132 0.7290 0.4245 0.7290 0.8538
No log 1.34 134 0.7227 0.3984 0.7227 0.8501
No log 1.3600 136 0.7839 0.3832 0.7839 0.8854
No log 1.38 138 0.8621 0.3604 0.8621 0.9285
No log 1.4 140 0.7481 0.3592 0.7481 0.8650
No log 1.42 142 1.0943 0.2856 1.0943 1.0461
No log 1.44 144 1.0866 0.2658 1.0866 1.0424
No log 1.46 146 0.7458 0.3238 0.7458 0.8636
No log 1.48 148 0.7603 0.3880 0.7603 0.8719
No log 1.5 150 0.7323 0.3888 0.7323 0.8557
No log 1.52 152 0.7189 0.4023 0.7189 0.8479
No log 1.54 154 0.7487 0.3944 0.7487 0.8653
No log 1.56 156 0.6687 0.4244 0.6687 0.8177
No log 1.58 158 0.6963 0.4631 0.6963 0.8345
No log 1.6 160 0.7107 0.5443 0.7107 0.8431
No log 1.62 162 0.8274 0.4334 0.8274 0.9096
No log 1.6400 164 1.0123 0.3878 1.0123 1.0061
No log 1.6600 166 0.9420 0.4096 0.9420 0.9706
No log 1.6800 168 0.7719 0.4575 0.7719 0.8786
No log 1.7 170 0.7772 0.5399 0.7772 0.8816
No log 1.72 172 0.7682 0.5534 0.7682 0.8765
No log 1.74 174 0.8171 0.4542 0.8171 0.9039
No log 1.76 176 0.9070 0.4239 0.9070 0.9523
No log 1.78 178 0.7991 0.4606 0.7991 0.8939
No log 1.8 180 0.7140 0.5648 0.7140 0.8450
No log 1.8200 182 0.7983 0.5117 0.7983 0.8935
No log 1.8400 184 0.7324 0.5305 0.7324 0.8558
No log 1.8600 186 0.6537 0.5443 0.6537 0.8085
No log 1.88 188 0.7053 0.4487 0.7053 0.8398
No log 1.9 190 0.7653 0.4382 0.7653 0.8748
No log 1.92 192 0.7197 0.4459 0.7197 0.8484
No log 1.94 194 0.6705 0.5071 0.6705 0.8188
No log 1.96 196 0.6542 0.4885 0.6542 0.8088
No log 1.98 198 0.6744 0.4611 0.6744 0.8212
No log 2.0 200 0.7076 0.5183 0.7076 0.8412
No log 2.02 202 0.7302 0.5246 0.7302 0.8545
No log 2.04 204 0.7040 0.4925 0.7040 0.8390
No log 2.06 206 0.6884 0.4575 0.6884 0.8297
No log 2.08 208 0.7344 0.4690 0.7344 0.8570
No log 2.1 210 0.7793 0.3925 0.7793 0.8828
No log 2.12 212 0.6962 0.4573 0.6962 0.8344
No log 2.14 214 0.6763 0.4130 0.6763 0.8224
No log 2.16 216 0.7646 0.4007 0.7646 0.8744
No log 2.18 218 0.8488 0.3334 0.8488 0.9213
No log 2.2 220 0.7077 0.3685 0.7077 0.8412
No log 2.22 222 0.7163 0.3659 0.7163 0.8464
No log 2.24 224 0.6758 0.3760 0.6758 0.8221
No log 2.26 226 0.7164 0.4135 0.7164 0.8464
No log 2.2800 228 0.7954 0.3695 0.7954 0.8919
No log 2.3 230 0.7106 0.4359 0.7106 0.8430
No log 2.32 232 0.6221 0.4653 0.6221 0.7887
No log 2.34 234 0.6158 0.4737 0.6158 0.7847
No log 2.36 236 0.6431 0.4759 0.6431 0.8019
No log 2.38 238 0.6557 0.4891 0.6557 0.8097
No log 2.4 240 0.6321 0.4857 0.6321 0.7950
No log 2.42 242 0.6517 0.4506 0.6517 0.8073
No log 2.44 244 0.6738 0.4986 0.6738 0.8209
No log 2.46 246 0.6866 0.5044 0.6866 0.8286
No log 2.48 248 0.6869 0.5101 0.6869 0.8288
No log 2.5 250 0.6638 0.5063 0.6638 0.8147
No log 2.52 252 0.7404 0.4906 0.7404 0.8604
No log 2.54 254 0.7091 0.4639 0.7091 0.8421
No log 2.56 256 0.6136 0.5150 0.6136 0.7833
No log 2.58 258 0.6537 0.4980 0.6537 0.8085
No log 2.6 260 0.6313 0.4783 0.6313 0.7945
No log 2.62 262 0.6440 0.4978 0.6440 0.8025
No log 2.64 264 0.6800 0.4650 0.6800 0.8246
No log 2.66 266 0.7124 0.4837 0.7124 0.8440
No log 2.68 268 0.7398 0.4514 0.7398 0.8601
No log 2.7 270 0.7582 0.4625 0.7582 0.8708
No log 2.7200 272 0.7878 0.4625 0.7878 0.8876
No log 2.74 274 0.8279 0.4399 0.8279 0.9099
No log 2.76 276 0.8957 0.3608 0.8957 0.9464
No log 2.7800 278 0.9242 0.3613 0.9242 0.9614
No log 2.8 280 1.1036 0.3400 1.1036 1.0505
No log 2.82 282 1.0397 0.3142 1.0397 1.0197
No log 2.84 284 0.7932 0.4304 0.7932 0.8906
No log 2.86 286 0.6950 0.4453 0.6950 0.8337
No log 2.88 288 0.6994 0.4558 0.6994 0.8363
No log 2.9 290 0.7853 0.3935 0.7853 0.8862
No log 2.92 292 0.8670 0.4239 0.8670 0.9311
No log 2.94 294 0.7880 0.4698 0.7880 0.8877
No log 2.96 296 0.6993 0.4974 0.6993 0.8362
No log 2.98 298 0.7275 0.4514 0.7275 0.8529
No log 3.0 300 0.7054 0.4675 0.7054 0.8399
No log 3.02 302 0.7257 0.4365 0.7257 0.8519
No log 3.04 304 0.6816 0.4722 0.6816 0.8256
No log 3.06 306 0.6431 0.4913 0.6431 0.8019
No log 3.08 308 0.6572 0.4549 0.6572 0.8107
No log 3.1 310 0.6552 0.5033 0.6552 0.8095
No log 3.12 312 0.7676 0.4270 0.7676 0.8761
No log 3.14 314 0.8480 0.4474 0.8480 0.9209
No log 3.16 316 0.7437 0.4689 0.7437 0.8624
No log 3.18 318 0.6397 0.5020 0.6397 0.7998
No log 3.2 320 0.6130 0.4991 0.6130 0.7829
No log 3.22 322 0.6280 0.4114 0.6280 0.7925
No log 3.24 324 0.6143 0.4598 0.6143 0.7838
No log 3.26 326 0.6193 0.4559 0.6193 0.7869
No log 3.2800 328 0.6290 0.4430 0.6290 0.7931
No log 3.3 330 0.6413 0.5403 0.6413 0.8008
No log 3.32 332 0.6490 0.4829 0.6490 0.8056
No log 3.34 334 0.6667 0.4638 0.6667 0.8165
No log 3.36 336 0.6668 0.4821 0.6668 0.8166
No log 3.38 338 0.6709 0.4906 0.6709 0.8191
No log 3.4 340 0.6497 0.4645 0.6497 0.8060
No log 3.42 342 0.6433 0.4763 0.6433 0.8021
No log 3.44 344 0.6396 0.4646 0.6396 0.7997
No log 3.46 346 0.6334 0.4642 0.6334 0.7958
No log 3.48 348 0.6469 0.4965 0.6469 0.8043
No log 3.5 350 0.6331 0.4835 0.6331 0.7957
No log 3.52 352 0.6653 0.4651 0.6653 0.8156
No log 3.54 354 0.6957 0.4266 0.6957 0.8341
No log 3.56 356 0.6261 0.4928 0.6261 0.7913
No log 3.58 358 0.7666 0.3822 0.7666 0.8756
No log 3.6 360 0.7906 0.3875 0.7906 0.8892
No log 3.62 362 0.6615 0.4893 0.6615 0.8133
No log 3.64 364 0.6703 0.4122 0.6703 0.8187
No log 3.66 366 0.9155 0.3218 0.9155 0.9568
No log 3.68 368 0.9203 0.3218 0.9203 0.9593
No log 3.7 370 0.7017 0.3959 0.7017 0.8377
No log 3.7200 372 0.6380 0.4576 0.6380 0.7988
No log 3.74 374 0.6880 0.3614 0.6880 0.8294
No log 3.76 376 0.6575 0.4313 0.6575 0.8108
No log 3.7800 378 0.6223 0.4073 0.6223 0.7889
No log 3.8 380 0.6447 0.4061 0.6447 0.8029
No log 3.82 382 0.6554 0.4223 0.6554 0.8095
No log 3.84 384 0.6301 0.5396 0.6301 0.7938
No log 3.86 386 0.6894 0.4779 0.6894 0.8303
No log 3.88 388 0.7350 0.4680 0.7350 0.8573
No log 3.9 390 0.7466 0.4919 0.7466 0.8641
No log 3.92 392 0.7139 0.4953 0.7139 0.8449
No log 3.94 394 0.6687 0.5208 0.6687 0.8177
No log 3.96 396 0.6742 0.4831 0.6742 0.8211
No log 3.98 398 0.7105 0.4977 0.7105 0.8429
No log 4.0 400 0.8683 0.4301 0.8683 0.9318
No log 4.02 402 0.9678 0.3796 0.9678 0.9838
No log 4.04 404 0.9231 0.4175 0.9231 0.9608
No log 4.06 406 0.7402 0.4803 0.7402 0.8604
No log 4.08 408 0.6170 0.5284 0.6170 0.7855
No log 4.1 410 0.6583 0.4843 0.6583 0.8113
No log 4.12 412 0.6644 0.4843 0.6644 0.8151
No log 4.14 414 0.6380 0.5354 0.6380 0.7988
No log 4.16 416 0.7425 0.4971 0.7425 0.8617
No log 4.18 418 0.8432 0.4514 0.8432 0.9182
No log 4.2 420 0.8287 0.4130 0.8287 0.9103
No log 4.22 422 0.6900 0.4926 0.6900 0.8306
No log 4.24 424 0.6035 0.5097 0.6035 0.7768
No log 4.26 426 0.6248 0.4257 0.6248 0.7904
No log 4.28 428 0.6166 0.4603 0.6166 0.7853
No log 4.3 430 0.6102 0.5440 0.6102 0.7811
No log 4.32 432 0.6723 0.5045 0.6723 0.8200
No log 4.34 434 0.7594 0.4610 0.7594 0.8714
No log 4.36 436 0.7165 0.5313 0.7165 0.8465
No log 4.38 438 0.6538 0.5475 0.6538 0.8086
No log 4.4 440 0.6190 0.4965 0.6190 0.7868
No log 4.42 442 0.6262 0.5065 0.6262 0.7913
No log 4.44 444 0.6429 0.5051 0.6429 0.8018
No log 4.46 446 0.6469 0.5051 0.6469 0.8043
No log 4.48 448 0.6258 0.5059 0.6258 0.7911
No log 4.5 450 0.6198 0.4835 0.6198 0.7872
No log 4.52 452 0.6131 0.4609 0.6131 0.7830
No log 4.54 454 0.6127 0.4630 0.6127 0.7827
No log 4.5600 456 0.6250 0.5019 0.6250 0.7906
No log 4.58 458 0.6380 0.5139 0.6380 0.7988
No log 4.6 460 0.6340 0.5176 0.6340 0.7963
No log 4.62 462 0.6413 0.5292 0.6413 0.8008
No log 4.64 464 0.6336 0.5017 0.6336 0.7960
No log 4.66 466 0.6225 0.5284 0.6225 0.7890
No log 4.68 468 0.6358 0.5298 0.6358 0.7974
No log 4.7 470 0.6334 0.5200 0.6334 0.7959
No log 4.72 472 0.6274 0.4596 0.6274 0.7921
No log 4.74 474 0.6040 0.4424 0.6040 0.7772
No log 4.76 476 0.6091 0.4281 0.6091 0.7804
No log 4.78 478 0.6014 0.4696 0.6014 0.7755
No log 4.8 480 0.6225 0.4712 0.6225 0.7890
No log 4.82 482 0.6251 0.5105 0.6251 0.7906
No log 4.84 484 0.6575 0.5246 0.6575 0.8109
No log 4.86 486 0.6424 0.4972 0.6424 0.8015
No log 4.88 488 0.6380 0.4972 0.6380 0.7988
No log 4.9 490 0.6526 0.5124 0.6526 0.8078
No log 4.92 492 0.6969 0.5462 0.6969 0.8348
No log 4.9400 494 0.7455 0.5275 0.7455 0.8634
No log 4.96 496 0.7211 0.5251 0.7211 0.8492
No log 4.98 498 0.7025 0.5280 0.7025 0.8381
0.4106 5.0 500 0.7017 0.5400 0.7017 0.8377
0.4106 5.02 502 0.6904 0.5321 0.6904 0.8309
0.4106 5.04 504 0.6578 0.5371 0.6578 0.8111
0.4106 5.06 506 0.6378 0.5224 0.6378 0.7986
0.4106 5.08 508 0.6351 0.5329 0.6351 0.7969
0.4106 5.1 510 0.6496 0.5093 0.6496 0.8059

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k19_task2_organization

Finetuned
(4206)
this model