ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k12_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7994
  • Qwk: 0.5618
  • Mse: 0.7994
  • Rmse: 0.8941

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0455 2 4.5493 0.0010 4.5493 2.1329
No log 0.0909 4 2.5039 0.0332 2.5039 1.5824
No log 0.1364 6 1.7782 0.0062 1.7782 1.3335
No log 0.1818 8 1.5483 0.0372 1.5483 1.2443
No log 0.2273 10 1.3628 0.0792 1.3628 1.1674
No log 0.2727 12 1.2425 0.2328 1.2425 1.1147
No log 0.3182 14 1.2417 0.1347 1.2417 1.1143
No log 0.3636 16 1.3334 0.1115 1.3334 1.1547
No log 0.4091 18 1.4216 0.1277 1.4216 1.1923
No log 0.4545 20 1.4764 0.1277 1.4764 1.2151
No log 0.5 22 1.6812 0.1026 1.6812 1.2966
No log 0.5455 24 1.6508 0.1247 1.6508 1.2848
No log 0.5909 26 1.6180 0.1169 1.6180 1.2720
No log 0.6364 28 1.5411 0.1028 1.5411 1.2414
No log 0.6818 30 1.3429 0.1395 1.3429 1.1588
No log 0.7273 32 1.3322 0.1570 1.3322 1.1542
No log 0.7727 34 1.5419 0.0749 1.5419 1.2417
No log 0.8182 36 1.6007 0.0889 1.6007 1.2652
No log 0.8636 38 1.3934 0.0622 1.3934 1.1804
No log 0.9091 40 1.2171 0.2161 1.2171 1.1032
No log 0.9545 42 1.0207 0.3487 1.0207 1.0103
No log 1.0 44 1.0242 0.3189 1.0242 1.0120
No log 1.0455 46 1.0970 0.2984 1.0970 1.0474
No log 1.0909 48 1.4195 0.0818 1.4195 1.1914
No log 1.1364 50 1.7634 0.0789 1.7634 1.3279
No log 1.1818 52 1.7358 0.1223 1.7358 1.3175
No log 1.2273 54 1.7019 0.1223 1.7019 1.3046
No log 1.2727 56 1.4203 0.1522 1.4203 1.1918
No log 1.3182 58 1.0576 0.2769 1.0576 1.0284
No log 1.3636 60 0.9597 0.3330 0.9597 0.9796
No log 1.4091 62 0.9210 0.4557 0.9210 0.9597
No log 1.4545 64 0.9220 0.4773 0.9220 0.9602
No log 1.5 66 0.9176 0.4773 0.9176 0.9579
No log 1.5455 68 0.9088 0.5045 0.9088 0.9533
No log 1.5909 70 0.9023 0.4074 0.9023 0.9499
No log 1.6364 72 0.8758 0.4074 0.8758 0.9358
No log 1.6818 74 0.9864 0.4391 0.9864 0.9932
No log 1.7273 76 1.0327 0.4391 1.0327 1.0162
No log 1.7727 78 1.0000 0.4885 1.0000 1.0000
No log 1.8182 80 1.0767 0.4617 1.0767 1.0376
No log 1.8636 82 1.5119 0.3419 1.5119 1.2296
No log 1.9091 84 1.9062 0.2274 1.9062 1.3807
No log 1.9545 86 1.7692 0.2631 1.7692 1.3301
No log 2.0 88 1.2363 0.3807 1.2363 1.1119
No log 2.0455 90 0.9941 0.3198 0.9941 0.9970
No log 2.0909 92 0.8099 0.3806 0.8099 0.8999
No log 2.1364 94 0.8095 0.4466 0.8095 0.8997
No log 2.1818 96 0.8300 0.4406 0.8300 0.9110
No log 2.2273 98 0.9027 0.3781 0.9027 0.9501
No log 2.2727 100 1.0408 0.3006 1.0408 1.0202
No log 2.3182 102 1.1125 0.3372 1.1125 1.0548
No log 2.3636 104 0.9868 0.3640 0.9868 0.9934
No log 2.4091 106 0.7532 0.5351 0.7532 0.8679
No log 2.4545 108 0.8466 0.4979 0.8466 0.9201
No log 2.5 110 1.0462 0.2978 1.0462 1.0229
No log 2.5455 112 0.9616 0.4257 0.9616 0.9806
No log 2.5909 114 0.7663 0.5150 0.7663 0.8754
No log 2.6364 116 0.8564 0.4550 0.8564 0.9254
No log 2.6818 118 1.0029 0.4758 1.0029 1.0014
No log 2.7273 120 0.8890 0.5115 0.8890 0.9428
No log 2.7727 122 0.7202 0.5647 0.7202 0.8487
No log 2.8182 124 0.7273 0.4942 0.7273 0.8528
No log 2.8636 126 0.7391 0.4579 0.7391 0.8597
No log 2.9091 128 0.7323 0.5027 0.7323 0.8558
No log 2.9545 130 0.7826 0.4624 0.7826 0.8847
No log 3.0 132 0.8668 0.5007 0.8668 0.9310
No log 3.0455 134 0.8663 0.5007 0.8663 0.9307
No log 3.0909 136 0.8170 0.3648 0.8170 0.9039
No log 3.1364 138 0.9107 0.4428 0.9107 0.9543
No log 3.1818 140 0.9561 0.4280 0.9561 0.9778
No log 3.2273 142 0.7987 0.4995 0.7987 0.8937
No log 3.2727 144 0.7516 0.5027 0.7516 0.8670
No log 3.3182 146 0.7366 0.5622 0.7366 0.8582
No log 3.3636 148 0.7487 0.5725 0.7487 0.8653
No log 3.4091 150 0.7115 0.5295 0.7115 0.8435
No log 3.4545 152 0.7339 0.4292 0.7339 0.8567
No log 3.5 154 0.7467 0.5451 0.7467 0.8641
No log 3.5455 156 0.8047 0.5704 0.8047 0.8971
No log 3.5909 158 0.7677 0.5569 0.7677 0.8762
No log 3.6364 160 0.7878 0.4541 0.7878 0.8876
No log 3.6818 162 0.8421 0.4475 0.8421 0.9177
No log 3.7273 164 0.7953 0.4944 0.7953 0.8918
No log 3.7727 166 0.8138 0.5472 0.8138 0.9021
No log 3.8182 168 0.8800 0.4686 0.8800 0.9381
No log 3.8636 170 0.8977 0.4824 0.8977 0.9475
No log 3.9091 172 0.8461 0.5012 0.8461 0.9198
No log 3.9545 174 0.8574 0.4781 0.8574 0.9260
No log 4.0 176 0.8957 0.4244 0.8957 0.9464
No log 4.0455 178 0.9583 0.3840 0.9583 0.9789
No log 4.0909 180 0.9121 0.4488 0.9121 0.9551
No log 4.1364 182 0.9123 0.4824 0.9123 0.9551
No log 4.1818 184 0.9350 0.4304 0.9350 0.9669
No log 4.2273 186 0.8723 0.5065 0.8723 0.9340
No log 4.2727 188 0.8078 0.4157 0.8078 0.8988
No log 4.3182 190 0.8403 0.3704 0.8403 0.9167
No log 4.3636 192 0.8457 0.3704 0.8457 0.9196
No log 4.4091 194 0.8474 0.4012 0.8474 0.9205
No log 4.4545 196 0.8389 0.4012 0.8389 0.9159
No log 4.5 198 0.8508 0.3457 0.8508 0.9224
No log 4.5455 200 0.8810 0.4115 0.8810 0.9386
No log 4.5909 202 0.9002 0.4755 0.9002 0.9488
No log 4.6364 204 0.8271 0.4439 0.8271 0.9094
No log 4.6818 206 0.7541 0.5265 0.7541 0.8684
No log 4.7273 208 0.7523 0.5810 0.7523 0.8673
No log 4.7727 210 0.7351 0.4359 0.7351 0.8574
No log 4.8182 212 0.7410 0.4301 0.7410 0.8608
No log 4.8636 214 0.7777 0.4379 0.7777 0.8819
No log 4.9091 216 0.7524 0.4301 0.7524 0.8674
No log 4.9545 218 0.7371 0.4724 0.7371 0.8586
No log 5.0 220 0.7361 0.5265 0.7361 0.8579
No log 5.0455 222 0.7338 0.5345 0.7338 0.8566
No log 5.0909 224 0.7324 0.5352 0.7324 0.8558
No log 5.1364 226 0.7329 0.5173 0.7329 0.8561
No log 5.1818 228 0.7195 0.5606 0.7195 0.8482
No log 5.2273 230 0.7247 0.5606 0.7247 0.8513
No log 5.2727 232 0.7254 0.5534 0.7254 0.8517
No log 5.3182 234 0.7852 0.4526 0.7852 0.8861
No log 5.3636 236 0.8838 0.4059 0.8838 0.9401
No log 5.4091 238 0.8512 0.4629 0.8512 0.9226
No log 5.4545 240 0.7795 0.4686 0.7795 0.8829
No log 5.5 242 0.7776 0.5234 0.7776 0.8818
No log 5.5455 244 0.8512 0.5492 0.8512 0.9226
No log 5.5909 246 0.9384 0.4259 0.9384 0.9687
No log 5.6364 248 0.8952 0.4743 0.8952 0.9462
No log 5.6818 250 0.9047 0.4513 0.9047 0.9511
No log 5.7273 252 0.9343 0.4386 0.9343 0.9666
No log 5.7727 254 0.9486 0.4037 0.9486 0.9739
No log 5.8182 256 0.8983 0.5362 0.8983 0.9478
No log 5.8636 258 0.8112 0.4371 0.8112 0.9007
No log 5.9091 260 0.8525 0.3284 0.8525 0.9233
No log 5.9545 262 0.8804 0.3679 0.8804 0.9383
No log 6.0 264 0.8206 0.4157 0.8206 0.9059
No log 6.0455 266 0.8221 0.5362 0.8221 0.9067
No log 6.0909 268 0.9772 0.4397 0.9772 0.9885
No log 6.1364 270 0.9774 0.4302 0.9774 0.9886
No log 6.1818 272 0.8768 0.5417 0.8768 0.9364
No log 6.2273 274 0.7963 0.4349 0.7963 0.8924
No log 6.2727 276 0.9991 0.4843 0.9991 0.9996
No log 6.3182 278 1.1053 0.3416 1.1053 1.0513
No log 6.3636 280 0.9317 0.4471 0.9317 0.9652
No log 6.4091 282 0.8470 0.4455 0.8470 0.9204
No log 6.4545 284 0.9580 0.3836 0.9580 0.9788
No log 6.5 286 0.9859 0.3428 0.9859 0.9929
No log 6.5455 288 0.9315 0.3298 0.9315 0.9651
No log 6.5909 290 0.8890 0.4341 0.8890 0.9429
No log 6.6364 292 0.9176 0.3891 0.9176 0.9579
No log 6.6818 294 0.9265 0.3891 0.9265 0.9625
No log 6.7273 296 0.9121 0.4341 0.9121 0.9550
No log 6.7727 298 0.9581 0.3615 0.9581 0.9788
No log 6.8182 300 1.0360 0.3278 1.0360 1.0179
No log 6.8636 302 1.0307 0.2019 1.0307 1.0152
No log 6.9091 304 0.9431 0.3804 0.9431 0.9711
No log 6.9545 306 0.8361 0.4423 0.8361 0.9144
No log 7.0 308 0.7898 0.4701 0.7898 0.8887
No log 7.0455 310 0.7841 0.4367 0.7841 0.8855
No log 7.0909 312 0.7890 0.4008 0.7890 0.8883
No log 7.1364 314 0.8048 0.4854 0.8048 0.8971
No log 7.1818 316 0.7963 0.4198 0.7963 0.8924
No log 7.2273 318 0.7967 0.3909 0.7967 0.8926
No log 7.2727 320 0.8186 0.4466 0.8186 0.9048
No log 7.3182 322 0.8696 0.4393 0.8696 0.9325
No log 7.3636 324 0.9005 0.3537 0.9005 0.9489
No log 7.4091 326 0.8708 0.4033 0.8708 0.9332
No log 7.4545 328 0.8245 0.4260 0.8245 0.9080
No log 7.5 330 0.8244 0.4299 0.8244 0.9080
No log 7.5455 332 0.8219 0.4260 0.8219 0.9066
No log 7.5909 334 0.8165 0.4260 0.8165 0.9036
No log 7.6364 336 0.8219 0.4260 0.8219 0.9066
No log 7.6818 338 0.8367 0.4958 0.8367 0.9147
No log 7.7273 340 0.8590 0.4781 0.8590 0.9268
No log 7.7727 342 0.8895 0.3069 0.8895 0.9431
No log 7.8182 344 0.9028 0.3069 0.9028 0.9502
No log 7.8636 346 0.8691 0.3700 0.8691 0.9323
No log 7.9091 348 0.8410 0.3744 0.8410 0.9171
No log 7.9545 350 0.8403 0.4482 0.8403 0.9167
No log 8.0 352 0.8530 0.4243 0.8530 0.9236
No log 8.0455 354 0.8406 0.4769 0.8406 0.9169
No log 8.0909 356 0.8210 0.5010 0.8210 0.9061
No log 8.1364 358 0.8559 0.4275 0.8559 0.9251
No log 8.1818 360 0.8412 0.4236 0.8412 0.9172
No log 8.2273 362 0.7836 0.5561 0.7836 0.8852
No log 8.2727 364 0.7992 0.6006 0.7992 0.8940
No log 8.3182 366 0.7919 0.6108 0.7919 0.8899
No log 8.3636 368 0.7568 0.5940 0.7568 0.8700
No log 8.4091 370 0.7507 0.4671 0.7507 0.8664
No log 8.4545 372 0.7578 0.5027 0.7578 0.8705
No log 8.5 374 0.8045 0.3985 0.8045 0.8969
No log 8.5455 376 0.7983 0.3985 0.7983 0.8934
No log 8.5909 378 0.7478 0.4853 0.7478 0.8648
No log 8.6364 380 0.7109 0.6082 0.7109 0.8432
No log 8.6818 382 0.7045 0.6506 0.7045 0.8394
No log 8.7273 384 0.6955 0.6046 0.6955 0.8340
No log 8.7727 386 0.7670 0.5913 0.7670 0.8758
No log 8.8182 388 0.9075 0.3091 0.9075 0.9526
No log 8.8636 390 0.9484 0.2522 0.9484 0.9738
No log 8.9091 392 0.8904 0.2865 0.8904 0.9436
No log 8.9545 394 0.7994 0.4697 0.7994 0.8941
No log 9.0 396 0.7888 0.4344 0.7888 0.8881
No log 9.0455 398 0.8442 0.4586 0.8442 0.9188
No log 9.0909 400 0.8554 0.4578 0.8554 0.9249
No log 9.1364 402 0.7973 0.4796 0.7973 0.8929
No log 9.1818 404 0.7259 0.5690 0.7259 0.8520
No log 9.2273 406 0.7560 0.6157 0.7560 0.8695
No log 9.2727 408 0.9143 0.4166 0.9143 0.9562
No log 9.3182 410 0.9954 0.2782 0.9954 0.9977
No log 9.3636 412 0.9810 0.2972 0.9810 0.9904
No log 9.4091 414 0.8917 0.3278 0.8917 0.9443
No log 9.4545 416 0.8409 0.4439 0.8409 0.9170
No log 9.5 418 0.8155 0.4115 0.8155 0.9030
No log 9.5455 420 0.8053 0.4511 0.8053 0.8974
No log 9.5909 422 0.8088 0.4116 0.8088 0.8993
No log 9.6364 424 0.8445 0.4033 0.8445 0.9190
No log 9.6818 426 0.9196 0.3278 0.9196 0.9590
No log 9.7273 428 0.9484 0.3124 0.9484 0.9739
No log 9.7727 430 0.9016 0.3278 0.9016 0.9496
No log 9.8182 432 0.8564 0.4433 0.8564 0.9254
No log 9.8636 434 0.8500 0.4261 0.8500 0.9219
No log 9.9091 436 0.8492 0.4301 0.8492 0.9215
No log 9.9545 438 0.8409 0.4261 0.8409 0.9170
No log 10.0 440 0.8487 0.4413 0.8487 0.9213
No log 10.0455 442 0.8892 0.4201 0.8892 0.9430
No log 10.0909 444 0.8603 0.4785 0.8603 0.9275
No log 10.1364 446 0.8299 0.4701 0.8299 0.9110
No log 10.1818 448 0.8414 0.4220 0.8414 0.9173
No log 10.2273 450 0.8371 0.4116 0.8371 0.9149
No log 10.2727 452 0.8354 0.3970 0.8354 0.9140
No log 10.3182 454 0.8517 0.3437 0.8517 0.9229
No log 10.3636 456 0.8764 0.4243 0.8764 0.9362
No log 10.4091 458 0.8686 0.4385 0.8686 0.9320
No log 10.4545 460 0.8409 0.3629 0.8409 0.9170
No log 10.5 462 0.8207 0.4180 0.8207 0.9059
No log 10.5455 464 0.7968 0.4116 0.7968 0.8926
No log 10.5909 466 0.7886 0.4116 0.7886 0.8880
No log 10.6364 468 0.7956 0.4608 0.7956 0.8920
No log 10.6818 470 0.7987 0.5163 0.7987 0.8937
No log 10.7273 472 0.8038 0.5244 0.8038 0.8965
No log 10.7727 474 0.7948 0.4839 0.7948 0.8915
No log 10.8182 476 0.8045 0.4116 0.8045 0.8969
No log 10.8636 478 0.8191 0.4115 0.8191 0.9051
No log 10.9091 480 0.8253 0.4115 0.8253 0.9084
No log 10.9545 482 0.8446 0.3933 0.8446 0.9190
No log 11.0 484 0.8469 0.4757 0.8469 0.9202
No log 11.0455 486 0.8081 0.5163 0.8081 0.8990
No log 11.0909 488 0.7787 0.5505 0.7787 0.8825
No log 11.1364 490 0.7652 0.5505 0.7652 0.8747
No log 11.1818 492 0.7581 0.5634 0.7581 0.8707
No log 11.2273 494 0.7629 0.5634 0.7629 0.8734
No log 11.2727 496 0.8018 0.5692 0.8018 0.8954
No log 11.3182 498 0.8174 0.5692 0.8174 0.9041
0.3179 11.3636 500 0.8013 0.5794 0.8013 0.8952
0.3179 11.4091 502 0.7939 0.5505 0.7939 0.8910
0.3179 11.4545 504 0.8100 0.5143 0.8100 0.9000
0.3179 11.5 506 0.8081 0.5143 0.8081 0.8990
0.3179 11.5455 508 0.7933 0.5429 0.7933 0.8907
0.3179 11.5909 510 0.7994 0.5618 0.7994 0.8941

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
135M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k12_task2_organization

Finetuned
(4005)
this model