ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k6_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5830
  • Qwk: 0.6771
  • Mse: 0.5830
  • Rmse: 0.7635

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1176 2 4.0158 0.0069 4.0158 2.0039
No log 0.2353 4 2.1220 0.0631 2.1220 1.4567
No log 0.3529 6 1.2734 0.0878 1.2734 1.1285
No log 0.4706 8 1.0964 0.3666 1.0964 1.0471
No log 0.5882 10 1.0283 0.2161 1.0283 1.0140
No log 0.7059 12 1.1688 0.1101 1.1688 1.0811
No log 0.8235 14 1.2591 -0.0598 1.2591 1.1221
No log 0.9412 16 1.3923 -0.1273 1.3923 1.1800
No log 1.0588 18 1.5675 -0.0411 1.5675 1.2520
No log 1.1765 20 1.7316 0.0140 1.7316 1.3159
No log 1.2941 22 1.4309 -0.0411 1.4309 1.1962
No log 1.4118 24 1.1657 0.0249 1.1657 1.0797
No log 1.5294 26 1.1034 0.1313 1.1034 1.0504
No log 1.6471 28 1.0780 0.2236 1.0780 1.0383
No log 1.7647 30 1.0633 0.2081 1.0633 1.0312
No log 1.8824 32 1.0326 0.2135 1.0326 1.0162
No log 2.0 34 1.0321 0.3332 1.0321 1.0159
No log 2.1176 36 1.0370 0.3171 1.0370 1.0183
No log 2.2353 38 1.0014 0.3026 1.0014 1.0007
No log 2.3529 40 0.9978 0.2635 0.9978 0.9989
No log 2.4706 42 0.9627 0.2969 0.9627 0.9812
No log 2.5882 44 0.8318 0.4709 0.8318 0.9120
No log 2.7059 46 0.8235 0.4928 0.8235 0.9075
No log 2.8235 48 0.9445 0.3920 0.9445 0.9719
No log 2.9412 50 0.8550 0.4555 0.8550 0.9247
No log 3.0588 52 0.7663 0.4660 0.7663 0.8754
No log 3.1765 54 0.7738 0.5054 0.7738 0.8797
No log 3.2941 56 0.8575 0.4638 0.8575 0.9260
No log 3.4118 58 0.8239 0.4641 0.8239 0.9077
No log 3.5294 60 0.6931 0.5016 0.6931 0.8325
No log 3.6471 62 1.0092 0.4565 1.0092 1.0046
No log 3.7647 64 1.3717 0.1533 1.3717 1.1712
No log 3.8824 66 1.0712 0.3814 1.0712 1.0350
No log 4.0 68 0.7674 0.5439 0.7674 0.8760
No log 4.1176 70 0.7845 0.5971 0.7845 0.8857
No log 4.2353 72 1.1365 0.3392 1.1365 1.0661
No log 4.3529 74 1.0105 0.4787 1.0105 1.0052
No log 4.4706 76 0.6873 0.6174 0.6873 0.8291
No log 4.5882 78 0.7406 0.5735 0.7406 0.8606
No log 4.7059 80 0.7874 0.5425 0.7874 0.8874
No log 4.8235 82 0.8170 0.5745 0.8170 0.9039
No log 4.9412 84 0.7422 0.5504 0.7422 0.8615
No log 5.0588 86 0.6660 0.5809 0.6660 0.8161
No log 5.1765 88 0.6823 0.6089 0.6823 0.8260
No log 5.2941 90 0.6537 0.6104 0.6537 0.8085
No log 5.4118 92 0.6727 0.5660 0.6727 0.8202
No log 5.5294 94 0.6515 0.6196 0.6515 0.8072
No log 5.6471 96 0.7914 0.5530 0.7914 0.8896
No log 5.7647 98 0.9577 0.5541 0.9577 0.9786
No log 5.8824 100 0.8684 0.5833 0.8684 0.9319
No log 6.0 102 0.6806 0.6302 0.6806 0.8250
No log 6.1176 104 0.6658 0.6177 0.6658 0.8159
No log 6.2353 106 0.6514 0.5990 0.6514 0.8071
No log 6.3529 108 0.7160 0.5770 0.7160 0.8462
No log 6.4706 110 1.0505 0.3704 1.0505 1.0249
No log 6.5882 112 1.0780 0.3631 1.0780 1.0383
No log 6.7059 114 0.7890 0.5895 0.7890 0.8883
No log 6.8235 116 0.7366 0.4974 0.7366 0.8582
No log 6.9412 118 1.0015 0.4578 1.0015 1.0008
No log 7.0588 120 1.0136 0.4960 1.0136 1.0068
No log 7.1765 122 0.8290 0.5039 0.8290 0.9105
No log 7.2941 124 0.6957 0.5446 0.6957 0.8341
No log 7.4118 126 0.6837 0.5582 0.6837 0.8268
No log 7.5294 128 0.6923 0.5446 0.6923 0.8321
No log 7.6471 130 0.7714 0.5292 0.7714 0.8783
No log 7.7647 132 0.7679 0.4948 0.7679 0.8763
No log 7.8824 134 0.6784 0.6066 0.6784 0.8236
No log 8.0 136 0.6732 0.6334 0.6732 0.8205
No log 8.1176 138 0.6727 0.6410 0.6727 0.8202
No log 8.2353 140 0.6887 0.6418 0.6887 0.8299
No log 8.3529 142 0.7125 0.6286 0.7125 0.8441
No log 8.4706 144 0.6638 0.6452 0.6638 0.8148
No log 8.5882 146 0.6209 0.6508 0.6209 0.7880
No log 8.7059 148 0.6294 0.6045 0.6294 0.7933
No log 8.8235 150 0.7611 0.5853 0.7611 0.8724
No log 8.9412 152 0.8074 0.5378 0.8074 0.8985
No log 9.0588 154 0.6939 0.5976 0.6939 0.8330
No log 9.1765 156 0.5906 0.6078 0.5906 0.7685
No log 9.2941 158 0.5930 0.6455 0.5930 0.7701
No log 9.4118 160 0.7435 0.6493 0.7435 0.8623
No log 9.5294 162 0.6707 0.6502 0.6707 0.8189
No log 9.6471 164 0.5530 0.7308 0.5530 0.7436
No log 9.7647 166 0.5905 0.6525 0.5905 0.7685
No log 9.8824 168 0.5664 0.6164 0.5664 0.7526
No log 10.0 170 0.6169 0.6422 0.6169 0.7854
No log 10.1176 172 0.6785 0.5559 0.6785 0.8237
No log 10.2353 174 0.6734 0.5088 0.6734 0.8206
No log 10.3529 176 0.6610 0.5559 0.6610 0.8130
No log 10.4706 178 0.6006 0.5921 0.6006 0.7750
No log 10.5882 180 0.5577 0.6593 0.5577 0.7468
No log 10.7059 182 0.5408 0.7218 0.5408 0.7354
No log 10.8235 184 0.5503 0.6948 0.5503 0.7418
No log 10.9412 186 0.5600 0.6948 0.5600 0.7484
No log 11.0588 188 0.5614 0.6704 0.5614 0.7492
No log 11.1765 190 0.5789 0.6460 0.5789 0.7609
No log 11.2941 192 0.5957 0.6252 0.5957 0.7718
No log 11.4118 194 0.6386 0.5798 0.6386 0.7991
No log 11.5294 196 0.6410 0.5798 0.6410 0.8006
No log 11.6471 198 0.6444 0.5798 0.6444 0.8027
No log 11.7647 200 0.6231 0.6154 0.6231 0.7894
No log 11.8824 202 0.6005 0.5988 0.6005 0.7749
No log 12.0 204 0.5967 0.6205 0.5967 0.7724
No log 12.1176 206 0.5776 0.6896 0.5776 0.7600
No log 12.2353 208 0.6513 0.6340 0.6513 0.8070
No log 12.3529 210 0.6965 0.6301 0.6965 0.8346
No log 12.4706 212 0.6091 0.6519 0.6091 0.7805
No log 12.5882 214 0.5777 0.6636 0.5777 0.7601
No log 12.7059 216 0.5868 0.7025 0.5868 0.7660
No log 12.8235 218 0.5742 0.7124 0.5742 0.7577
No log 12.9412 220 0.6053 0.6038 0.6053 0.7780
No log 13.0588 222 0.6274 0.6128 0.6274 0.7921
No log 13.1765 224 0.5919 0.6414 0.5919 0.7694
No log 13.2941 226 0.5782 0.7286 0.5782 0.7604
No log 13.4118 228 0.5765 0.6854 0.5765 0.7593
No log 13.5294 230 0.5869 0.6217 0.5869 0.7661
No log 13.6471 232 0.6526 0.5305 0.6526 0.8078
No log 13.7647 234 0.7173 0.5521 0.7173 0.8470
No log 13.8824 236 0.6678 0.4974 0.6678 0.8172
No log 14.0 238 0.6070 0.5990 0.6070 0.7791
No log 14.1176 240 0.6090 0.6891 0.6090 0.7804
No log 14.2353 242 0.6875 0.6397 0.6875 0.8292
No log 14.3529 244 0.7045 0.6474 0.7045 0.8393
No log 14.4706 246 0.6845 0.6397 0.6845 0.8273
No log 14.5882 248 0.6221 0.7131 0.6221 0.7887
No log 14.7059 250 0.6183 0.6581 0.6183 0.7863
No log 14.8235 252 0.6592 0.6118 0.6592 0.8119
No log 14.9412 254 0.7008 0.5595 0.7008 0.8372
No log 15.0588 256 0.6738 0.5948 0.6738 0.8208
No log 15.1765 258 0.6318 0.6291 0.6318 0.7949
No log 15.2941 260 0.5851 0.6288 0.5851 0.7649
No log 15.4118 262 0.5759 0.6499 0.5759 0.7589
No log 15.5294 264 0.5693 0.6589 0.5693 0.7545
No log 15.6471 266 0.5709 0.6500 0.5709 0.7556
No log 15.7647 268 0.5810 0.6347 0.5810 0.7623
No log 15.8824 270 0.5932 0.6347 0.5932 0.7702
No log 16.0 272 0.6402 0.6118 0.6402 0.8001
No log 16.1176 274 0.6546 0.6118 0.6546 0.8091
No log 16.2353 276 0.6546 0.5832 0.6546 0.8091
No log 16.3529 278 0.6426 0.5943 0.6426 0.8016
No log 16.4706 280 0.6567 0.5710 0.6567 0.8103
No log 16.5882 282 0.6864 0.5585 0.6864 0.8285
No log 16.7059 284 0.7462 0.5400 0.7462 0.8638
No log 16.8235 286 0.7584 0.5358 0.7584 0.8709
No log 16.9412 288 0.7181 0.5686 0.7181 0.8474
No log 17.0588 290 0.6559 0.5923 0.6559 0.8099
No log 17.1765 292 0.6041 0.5948 0.6041 0.7772
No log 17.2941 294 0.5730 0.6133 0.5730 0.7569
No log 17.4118 296 0.5801 0.6276 0.5801 0.7616
No log 17.5294 298 0.6096 0.5495 0.6096 0.7808
No log 17.6471 300 0.6377 0.5259 0.6377 0.7986
No log 17.7647 302 0.6383 0.5259 0.6383 0.7989
No log 17.8824 304 0.6329 0.5605 0.6329 0.7955
No log 18.0 306 0.6350 0.5909 0.6350 0.7969
No log 18.1176 308 0.6613 0.6112 0.6613 0.8132
No log 18.2353 310 0.6691 0.5948 0.6691 0.8180
No log 18.3529 312 0.6474 0.5669 0.6474 0.8046
No log 18.4706 314 0.6472 0.5680 0.6472 0.8045
No log 18.5882 316 0.6546 0.5669 0.6546 0.8090
No log 18.7059 318 0.6830 0.5540 0.6830 0.8265
No log 18.8235 320 0.7322 0.5717 0.7322 0.8557
No log 18.9412 322 0.7315 0.5717 0.7315 0.8553
No log 19.0588 324 0.6642 0.6071 0.6642 0.8150
No log 19.1765 326 0.6141 0.6335 0.6141 0.7836
No log 19.2941 328 0.6175 0.6572 0.6175 0.7858
No log 19.4118 330 0.6271 0.5839 0.6271 0.7919
No log 19.5294 332 0.6063 0.6535 0.6063 0.7786
No log 19.6471 334 0.5957 0.6380 0.5957 0.7718
No log 19.7647 336 0.6257 0.5894 0.6257 0.7910
No log 19.8824 338 0.6359 0.5552 0.6359 0.7974
No log 20.0 340 0.6203 0.5894 0.6203 0.7876
No log 20.1176 342 0.6119 0.6597 0.6119 0.7822
No log 20.2353 344 0.6370 0.5894 0.6370 0.7981
No log 20.3529 346 0.6374 0.5917 0.6374 0.7984
No log 20.4706 348 0.6353 0.5510 0.6353 0.7971
No log 20.5882 350 0.6286 0.5982 0.6286 0.7928
No log 20.7059 352 0.6365 0.6620 0.6365 0.7978
No log 20.8235 354 0.6200 0.6435 0.6200 0.7874
No log 20.9412 356 0.6150 0.6572 0.6150 0.7842
No log 21.0588 358 0.6061 0.6470 0.6061 0.7785
No log 21.1765 360 0.6242 0.6301 0.6242 0.7900
No log 21.2941 362 0.6331 0.5797 0.6331 0.7956
No log 21.4118 364 0.6218 0.6175 0.6218 0.7885
No log 21.5294 366 0.6268 0.5328 0.6268 0.7917
No log 21.6471 368 0.6240 0.5327 0.6240 0.7899
No log 21.7647 370 0.6163 0.6078 0.6163 0.7850
No log 21.8824 372 0.6139 0.5748 0.6139 0.7835
No log 22.0 374 0.6089 0.6196 0.6089 0.7803
No log 22.1176 376 0.6414 0.5221 0.6414 0.8009
No log 22.2353 378 0.6628 0.5751 0.6628 0.8141
No log 22.3529 380 0.6361 0.5221 0.6361 0.7976
No log 22.4706 382 0.6298 0.5516 0.6298 0.7936
No log 22.5882 384 0.6323 0.5415 0.6323 0.7952
No log 22.7059 386 0.6239 0.5542 0.6239 0.7899
No log 22.8235 388 0.6163 0.5905 0.6163 0.7850
No log 22.9412 390 0.6499 0.5708 0.6499 0.8062
No log 23.0588 392 0.6756 0.5909 0.6756 0.8220
No log 23.1765 394 0.6388 0.5809 0.6388 0.7992
No log 23.2941 396 0.5960 0.6866 0.5960 0.7720
No log 23.4118 398 0.5898 0.6364 0.5898 0.7680
No log 23.5294 400 0.6051 0.6219 0.6051 0.7779
No log 23.6471 402 0.6022 0.6219 0.6022 0.7760
No log 23.7647 404 0.5787 0.6768 0.5787 0.7607
No log 23.8824 406 0.5648 0.6627 0.5648 0.7515
No log 24.0 408 0.5614 0.7041 0.5614 0.7493
No log 24.1176 410 0.5684 0.7136 0.5684 0.7539
No log 24.2353 412 0.5762 0.6673 0.5762 0.7591
No log 24.3529 414 0.5871 0.6488 0.5871 0.7662
No log 24.4706 416 0.5851 0.6606 0.5851 0.7649
No log 24.5882 418 0.5768 0.6911 0.5768 0.7595
No log 24.7059 420 0.5704 0.6911 0.5704 0.7552
No log 24.8235 422 0.5665 0.6911 0.5665 0.7527
No log 24.9412 424 0.5730 0.7026 0.5730 0.7570
No log 25.0588 426 0.5793 0.6806 0.5793 0.7611
No log 25.1765 428 0.5738 0.7143 0.5738 0.7575
No log 25.2941 430 0.5904 0.6070 0.5904 0.7683
No log 25.4118 432 0.6452 0.5909 0.6452 0.8032
No log 25.5294 434 0.6502 0.5909 0.6502 0.8064
No log 25.6471 436 0.5979 0.6184 0.5979 0.7732
No log 25.7647 438 0.5683 0.6866 0.5683 0.7538
No log 25.8824 440 0.5991 0.5657 0.5991 0.7740
No log 26.0 442 0.6095 0.5450 0.6095 0.7807
No log 26.1176 444 0.5891 0.6143 0.5891 0.7675
No log 26.2353 446 0.5841 0.6262 0.5841 0.7643
No log 26.3529 448 0.5839 0.6335 0.5839 0.7642
No log 26.4706 450 0.5829 0.6335 0.5829 0.7635
No log 26.5882 452 0.5915 0.6441 0.5915 0.7691
No log 26.7059 454 0.6161 0.6218 0.6161 0.7849
No log 26.8235 456 0.6034 0.6687 0.6034 0.7768
No log 26.9412 458 0.6088 0.6396 0.6088 0.7803
No log 27.0588 460 0.6340 0.6218 0.6340 0.7962
No log 27.1765 462 0.6384 0.6218 0.6384 0.7990
No log 27.2941 464 0.6481 0.6218 0.6481 0.8050
No log 27.4118 466 0.6914 0.6015 0.6914 0.8315
No log 27.5294 468 0.7631 0.5844 0.7631 0.8735
No log 27.6471 470 0.7516 0.5675 0.7516 0.8669
No log 27.7647 472 0.7114 0.5902 0.7114 0.8435
No log 27.8824 474 0.6616 0.5992 0.6616 0.8134
No log 28.0 476 0.6063 0.6322 0.6063 0.7787
No log 28.1176 478 0.5927 0.5542 0.5927 0.7699
No log 28.2353 480 0.5970 0.5846 0.5970 0.7727
No log 28.3529 482 0.5889 0.5983 0.5889 0.7674
No log 28.4706 484 0.6016 0.6697 0.6016 0.7757
No log 28.5882 486 0.6039 0.6432 0.6039 0.7771
No log 28.7059 488 0.5807 0.6584 0.5807 0.7620
No log 28.8235 490 0.5748 0.6872 0.5748 0.7582
No log 28.9412 492 0.5709 0.6772 0.5709 0.7555
No log 29.0588 494 0.5599 0.6546 0.5599 0.7483
No log 29.1765 496 0.5516 0.6796 0.5516 0.7427
No log 29.2941 498 0.5481 0.6796 0.5481 0.7403
0.2596 29.4118 500 0.5528 0.6886 0.5528 0.7435
0.2596 29.5294 502 0.5700 0.7051 0.5700 0.7550
0.2596 29.6471 504 0.6261 0.6209 0.6261 0.7913
0.2596 29.7647 506 0.6579 0.6099 0.6579 0.8111
0.2596 29.8824 508 0.6375 0.6099 0.6375 0.7985
0.2596 30.0 510 0.5936 0.6958 0.5936 0.7704
0.2596 30.1176 512 0.5873 0.6900 0.5873 0.7663
0.2596 30.2353 514 0.5680 0.6886 0.5680 0.7537
0.2596 30.3529 516 0.5653 0.6886 0.5653 0.7519
0.2596 30.4706 518 0.5656 0.6207 0.5656 0.7521
0.2596 30.5882 520 0.5691 0.6046 0.5691 0.7544
0.2596 30.7059 522 0.5830 0.6771 0.5830 0.7635

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
8
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k6_task5_organization

Finetuned
(4223)
this model