ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k18_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7574
  • Qwk: 0.4774
  • Mse: 0.7574
  • Rmse: 0.8703

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0435 2 4.0223 -0.0101 4.0223 2.0056
No log 0.0870 4 2.1125 0.0537 2.1125 1.4534
No log 0.1304 6 1.2801 0.0584 1.2801 1.1314
No log 0.1739 8 1.1475 0.1944 1.1475 1.0712
No log 0.2174 10 1.1905 0.2166 1.1905 1.0911
No log 0.2609 12 1.1179 0.1046 1.1179 1.0573
No log 0.3043 14 1.0891 0.1046 1.0891 1.0436
No log 0.3478 16 1.0504 0.1046 1.0504 1.0249
No log 0.3913 18 1.0780 0.2314 1.0780 1.0383
No log 0.4348 20 1.0653 0.2114 1.0653 1.0321
No log 0.4783 22 1.0592 0.2865 1.0592 1.0292
No log 0.5217 24 1.0291 0.2566 1.0291 1.0144
No log 0.5652 26 1.0021 0.1810 1.0021 1.0010
No log 0.6087 28 1.0411 0.2236 1.0411 1.0203
No log 0.6522 30 1.0538 0.2424 1.0538 1.0265
No log 0.6957 32 0.9977 0.2154 0.9977 0.9988
No log 0.7391 34 1.0371 0.2958 1.0371 1.0184
No log 0.7826 36 0.9664 0.2935 0.9664 0.9831
No log 0.8261 38 0.9398 0.1857 0.9398 0.9694
No log 0.8696 40 0.9723 0.3483 0.9723 0.9860
No log 0.9130 42 1.0721 0.3287 1.0721 1.0354
No log 0.9565 44 1.2198 0.2273 1.2198 1.1044
No log 1.0 46 0.9720 0.3483 0.9720 0.9859
No log 1.0435 48 0.8665 0.3407 0.8665 0.9309
No log 1.0870 50 0.9745 0.2567 0.9745 0.9872
No log 1.1304 52 1.0153 0.3276 1.0153 1.0076
No log 1.1739 54 1.0002 0.3885 1.0002 1.0001
No log 1.2174 56 0.9320 0.5164 0.9320 0.9654
No log 1.2609 58 0.7444 0.5192 0.7444 0.8628
No log 1.3043 60 0.7497 0.5635 0.7497 0.8658
No log 1.3478 62 0.8204 0.4518 0.8204 0.9057
No log 1.3913 64 0.7703 0.6119 0.7703 0.8777
No log 1.4348 66 0.8684 0.5298 0.8684 0.9319
No log 1.4783 68 0.7968 0.5363 0.7968 0.8926
No log 1.5217 70 0.8711 0.4624 0.8711 0.9333
No log 1.5652 72 0.8384 0.4833 0.8384 0.9156
No log 1.6087 74 0.7433 0.5844 0.7433 0.8622
No log 1.6522 76 0.7114 0.5736 0.7114 0.8434
No log 1.6957 78 0.6830 0.6390 0.6830 0.8265
No log 1.7391 80 0.6651 0.6584 0.6651 0.8156
No log 1.7826 82 0.7022 0.6497 0.7022 0.8380
No log 1.8261 84 0.6961 0.5223 0.6961 0.8343
No log 1.8696 86 0.8143 0.5266 0.8143 0.9024
No log 1.9130 88 0.7777 0.4741 0.7777 0.8819
No log 1.9565 90 0.7565 0.5024 0.7565 0.8698
No log 2.0 92 0.7726 0.5107 0.7726 0.8790
No log 2.0435 94 0.7848 0.4981 0.7848 0.8859
No log 2.0870 96 0.8970 0.4595 0.8970 0.9471
No log 2.1304 98 0.8859 0.4947 0.8859 0.9412
No log 2.1739 100 0.8136 0.4990 0.8136 0.9020
No log 2.2174 102 0.8587 0.5543 0.8587 0.9266
No log 2.2609 104 1.0936 0.3973 1.0936 1.0457
No log 2.3043 106 1.1102 0.3819 1.1102 1.0537
No log 2.3478 108 0.8536 0.4638 0.8536 0.9239
No log 2.3913 110 0.8009 0.4662 0.8009 0.8949
No log 2.4348 112 0.8346 0.4953 0.8346 0.9135
No log 2.4783 114 0.9413 0.4500 0.9413 0.9702
No log 2.5217 116 0.9657 0.4389 0.9657 0.9827
No log 2.5652 118 0.9216 0.4171 0.9216 0.9600
No log 2.6087 120 0.8830 0.4034 0.8830 0.9397
No log 2.6522 122 0.8182 0.4471 0.8182 0.9046
No log 2.6957 124 0.8057 0.4802 0.8057 0.8976
No log 2.7391 126 0.8109 0.3860 0.8109 0.9005
No log 2.7826 128 0.9758 0.3480 0.9758 0.9878
No log 2.8261 130 1.0573 0.3395 1.0573 1.0282
No log 2.8696 132 1.0346 0.3270 1.0346 1.0172
No log 2.9130 134 1.0548 0.4 1.0548 1.0271
No log 2.9565 136 0.9700 0.4320 0.9700 0.9849
No log 3.0 138 0.9235 0.5778 0.9235 0.9610
No log 3.0435 140 0.9135 0.5696 0.9135 0.9558
No log 3.0870 142 0.9774 0.5352 0.9774 0.9887
No log 3.1304 144 0.9622 0.5273 0.9622 0.9809
No log 3.1739 146 0.8328 0.5317 0.8328 0.9126
No log 3.2174 148 0.7664 0.5291 0.7664 0.8755
No log 3.2609 150 0.7656 0.4476 0.7656 0.8750
No log 3.3043 152 0.7765 0.4220 0.7765 0.8812
No log 3.3478 154 0.7946 0.4075 0.7946 0.8914
No log 3.3913 156 0.7942 0.4075 0.7942 0.8912
No log 3.4348 158 0.8552 0.4318 0.8552 0.9247
No log 3.4783 160 0.8420 0.4318 0.8420 0.9176
No log 3.5217 162 0.7468 0.4742 0.7468 0.8642
No log 3.5652 164 0.7086 0.5591 0.7086 0.8418
No log 3.6087 166 0.7318 0.6403 0.7318 0.8554
No log 3.6522 168 0.7128 0.5984 0.7128 0.8443
No log 3.6957 170 0.7336 0.5618 0.7336 0.8565
No log 3.7391 172 0.7669 0.5160 0.7669 0.8757
No log 3.7826 174 0.8076 0.4695 0.8076 0.8987
No log 3.8261 176 0.7728 0.4599 0.7728 0.8791
No log 3.8696 178 0.7781 0.4244 0.7781 0.8821
No log 3.9130 180 0.7985 0.3725 0.7985 0.8936
No log 3.9565 182 0.8444 0.4812 0.8444 0.9189
No log 4.0 184 0.8665 0.4696 0.8665 0.9309
No log 4.0435 186 0.8360 0.4712 0.8360 0.9143
No log 4.0870 188 0.8721 0.4806 0.8721 0.9339
No log 4.1304 190 0.9350 0.4423 0.9350 0.9669
No log 4.1739 192 0.8898 0.4169 0.8898 0.9433
No log 4.2174 194 0.7870 0.4824 0.7870 0.8871
No log 4.2609 196 0.7596 0.5152 0.7596 0.8716
No log 4.3043 198 0.7493 0.5184 0.7493 0.8656
No log 4.3478 200 0.7265 0.5483 0.7265 0.8523
No log 4.3913 202 0.7464 0.5425 0.7464 0.8639
No log 4.4348 204 0.7488 0.5540 0.7488 0.8653
No log 4.4783 206 0.6779 0.5700 0.6779 0.8233
No log 4.5217 208 0.6921 0.5684 0.6921 0.8319
No log 4.5652 210 0.7350 0.5366 0.7350 0.8573
No log 4.6087 212 0.6886 0.5357 0.6886 0.8298
No log 4.6522 214 0.7527 0.4610 0.7527 0.8676
No log 4.6957 216 0.8487 0.4417 0.8487 0.9212
No log 4.7391 218 0.8201 0.4565 0.8201 0.9056
No log 4.7826 220 0.7235 0.5572 0.7235 0.8506
No log 4.8261 222 0.6686 0.6537 0.6686 0.8177
No log 4.8696 224 0.6704 0.6207 0.6704 0.8188
No log 4.9130 226 0.6788 0.6066 0.6788 0.8239
No log 4.9565 228 0.8042 0.5631 0.8042 0.8968
No log 5.0 230 0.8719 0.5510 0.8719 0.9337
No log 5.0435 232 0.9192 0.5090 0.9192 0.9587
No log 5.0870 234 0.7679 0.5292 0.7679 0.8763
No log 5.1304 236 0.6985 0.4676 0.6985 0.8358
No log 5.1739 238 0.7304 0.5425 0.7304 0.8546
No log 5.2174 240 0.7291 0.5066 0.7291 0.8538
No log 5.2609 242 0.7417 0.4473 0.7417 0.8612
No log 5.3043 244 0.8019 0.4428 0.8019 0.8955
No log 5.3478 246 0.8708 0.4428 0.8708 0.9332
No log 5.3913 248 0.8724 0.4428 0.8724 0.9340
No log 5.4348 250 0.7952 0.4461 0.7952 0.8917
No log 5.4783 252 0.7736 0.4613 0.7736 0.8796
No log 5.5217 254 0.7833 0.4869 0.7833 0.8850
No log 5.5652 256 0.8329 0.4728 0.8329 0.9126
No log 5.6087 258 0.8621 0.4071 0.8621 0.9285
No log 5.6522 260 0.8533 0.4071 0.8533 0.9237
No log 5.6957 262 0.8407 0.3503 0.8407 0.9169
No log 5.7391 264 0.8301 0.3643 0.8301 0.9111
No log 5.7826 266 0.8029 0.4327 0.8029 0.8961
No log 5.8261 268 0.7795 0.4893 0.7795 0.8829
No log 5.8696 270 0.7696 0.5023 0.7696 0.8773
No log 5.9130 272 0.7667 0.5127 0.7667 0.8756
No log 5.9565 274 0.7613 0.4278 0.7613 0.8725
No log 6.0 276 0.7587 0.5055 0.7587 0.8711
No log 6.0435 278 0.7973 0.4220 0.7973 0.8929
No log 6.0870 280 0.8669 0.4180 0.8669 0.9311
No log 6.1304 282 0.8315 0.3921 0.8315 0.9118
No log 6.1739 284 0.7541 0.4612 0.7541 0.8684
No log 6.2174 286 0.7427 0.4760 0.7427 0.8618
No log 6.2609 288 0.7458 0.4478 0.7458 0.8636
No log 6.3043 290 0.7993 0.4439 0.7993 0.8940
No log 6.3478 292 0.8660 0.4429 0.8660 0.9306
No log 6.3913 294 0.9374 0.4212 0.9374 0.9682
No log 6.4348 296 0.9130 0.4212 0.9130 0.9555
No log 6.4783 298 0.7994 0.4335 0.7994 0.8941
No log 6.5217 300 0.7352 0.5403 0.7352 0.8575
No log 6.5652 302 0.7434 0.4685 0.7434 0.8622
No log 6.6087 304 0.7269 0.5736 0.7269 0.8526
No log 6.6522 306 0.7602 0.4254 0.7602 0.8719
No log 6.6957 308 0.8362 0.4584 0.8362 0.9144
No log 6.7391 310 0.8248 0.4597 0.8248 0.9082
No log 6.7826 312 0.7598 0.4742 0.7598 0.8717
No log 6.8261 314 0.7478 0.4747 0.7478 0.8648
No log 6.8696 316 0.7795 0.5130 0.7795 0.8829
No log 6.9130 318 0.7573 0.5022 0.7573 0.8702
No log 6.9565 320 0.7094 0.5161 0.7094 0.8422
No log 7.0 322 0.7025 0.5921 0.7025 0.8381
No log 7.0435 324 0.7294 0.5697 0.7294 0.8540
No log 7.0870 326 0.7043 0.6352 0.7043 0.8392
No log 7.1304 328 0.6680 0.6278 0.6680 0.8173
No log 7.1739 330 0.6645 0.5680 0.6645 0.8151
No log 7.2174 332 0.6751 0.5235 0.6751 0.8216
No log 7.2609 334 0.6857 0.5966 0.6857 0.8281
No log 7.3043 336 0.7141 0.5032 0.7141 0.8450
No log 7.3478 338 0.7693 0.5103 0.7693 0.8771
No log 7.3913 340 0.8610 0.4041 0.8610 0.9279
No log 7.4348 342 0.9054 0.4428 0.9054 0.9515
No log 7.4783 344 0.8654 0.5070 0.8654 0.9303
No log 7.5217 346 0.8054 0.5197 0.8054 0.8974
No log 7.5652 348 0.8035 0.5413 0.8035 0.8964
No log 7.6087 350 0.8725 0.5147 0.8725 0.9341
No log 7.6522 352 0.9069 0.4558 0.9069 0.9523
No log 7.6957 354 0.8233 0.5279 0.8233 0.9074
No log 7.7391 356 0.7266 0.5477 0.7266 0.8524
No log 7.7826 358 0.7019 0.5017 0.7019 0.8378
No log 7.8261 360 0.7004 0.4947 0.7004 0.8369
No log 7.8696 362 0.7036 0.4804 0.7036 0.8388
No log 7.9130 364 0.7197 0.5359 0.7197 0.8484
No log 7.9565 366 0.7190 0.5359 0.7190 0.8480
No log 8.0 368 0.7007 0.4789 0.7007 0.8371
No log 8.0435 370 0.6985 0.4831 0.6985 0.8358
No log 8.0870 372 0.7004 0.5434 0.7004 0.8369
No log 8.1304 374 0.6816 0.5169 0.6816 0.8256
No log 8.1739 376 0.7046 0.5855 0.7046 0.8394
No log 8.2174 378 0.7253 0.6491 0.7253 0.8516
No log 8.2609 380 0.6949 0.5455 0.6949 0.8336
No log 8.3043 382 0.6885 0.5396 0.6885 0.8298
No log 8.3478 384 0.7283 0.5325 0.7283 0.8534
No log 8.3913 386 0.7299 0.5194 0.7299 0.8544
No log 8.4348 388 0.7254 0.5163 0.7254 0.8517
No log 8.4783 390 0.7384 0.5361 0.7384 0.8593
No log 8.5217 392 0.7939 0.4958 0.7939 0.8910
No log 8.5652 394 0.8001 0.4697 0.8001 0.8945
No log 8.6087 396 0.8049 0.4697 0.8049 0.8972
No log 8.6522 398 0.8032 0.4461 0.8032 0.8962
No log 8.6957 400 0.8239 0.5192 0.8239 0.9077
No log 8.7391 402 0.8953 0.4907 0.8953 0.9462
No log 8.7826 404 0.9954 0.4568 0.9954 0.9977
No log 8.8261 406 1.0070 0.4568 1.0070 1.0035
No log 8.8696 408 0.9037 0.4809 0.9037 0.9506
No log 8.9130 410 0.8482 0.4697 0.8482 0.9210
No log 8.9565 412 0.8131 0.4186 0.8131 0.9017
No log 9.0 414 0.8382 0.4180 0.8382 0.9155
No log 9.0435 416 0.8871 0.4428 0.8871 0.9418
No log 9.0870 418 0.9232 0.4555 0.9232 0.9608
No log 9.1304 420 0.8815 0.4681 0.8815 0.9389
No log 9.1739 422 0.8842 0.4912 0.8842 0.9403
No log 9.2174 424 0.8408 0.5157 0.8408 0.9170
No log 9.2609 426 0.7852 0.5413 0.7852 0.8861
No log 9.3043 428 0.7803 0.5195 0.7803 0.8834
No log 9.3478 430 0.7912 0.4727 0.7912 0.8895
No log 9.3913 432 0.8262 0.4327 0.8262 0.9090
No log 9.4348 434 0.8676 0.3921 0.8676 0.9315
No log 9.4783 436 0.8470 0.4203 0.8470 0.9203
No log 9.5217 438 0.7744 0.4727 0.7744 0.8800
No log 9.5652 440 0.7491 0.4371 0.7491 0.8655
No log 9.6087 442 0.7466 0.4762 0.7466 0.8640
No log 9.6522 444 0.7573 0.4368 0.7573 0.8702
No log 9.6957 446 0.7669 0.4615 0.7669 0.8757
No log 9.7391 448 0.7549 0.4981 0.7549 0.8689
No log 9.7826 450 0.7516 0.4903 0.7516 0.8669
No log 9.8261 452 0.7808 0.4386 0.7808 0.8836
No log 9.8696 454 0.7935 0.4737 0.7935 0.8908
No log 9.9130 456 0.8008 0.4477 0.8008 0.8949
No log 9.9565 458 0.7863 0.4405 0.7863 0.8868
No log 10.0 460 0.7772 0.4626 0.7772 0.8816
No log 10.0435 462 0.7698 0.4286 0.7698 0.8774
No log 10.0870 464 0.7674 0.4259 0.7674 0.8760
No log 10.1304 466 0.7611 0.4898 0.7611 0.8724
No log 10.1739 468 0.7597 0.4898 0.7597 0.8716
No log 10.2174 470 0.7587 0.4138 0.7587 0.8710
No log 10.2609 472 0.7555 0.4138 0.7555 0.8692
No log 10.3043 474 0.7772 0.4490 0.7772 0.8816
No log 10.3478 476 0.8012 0.4847 0.8012 0.8951
No log 10.3913 478 0.7927 0.4958 0.7927 0.8903
No log 10.4348 480 0.7870 0.5181 0.7870 0.8871
No log 10.4783 482 0.7617 0.4862 0.7617 0.8727
No log 10.5217 484 0.7591 0.4507 0.7591 0.8713
No log 10.5652 486 0.7625 0.4527 0.7625 0.8732
No log 10.6087 488 0.7748 0.4644 0.7748 0.8802
No log 10.6522 490 0.7765 0.4644 0.7765 0.8812
No log 10.6957 492 0.7664 0.5025 0.7664 0.8755
No log 10.7391 494 0.7564 0.4676 0.7564 0.8697
No log 10.7826 496 0.7435 0.4560 0.7435 0.8623
No log 10.8261 498 0.7455 0.5260 0.7455 0.8634
0.2663 10.8696 500 0.7429 0.5146 0.7429 0.8619
0.2663 10.9130 502 0.7342 0.4578 0.7342 0.8569
0.2663 10.9565 504 0.7356 0.4578 0.7356 0.8577
0.2663 11.0 506 0.7418 0.4428 0.7418 0.8613
0.2663 11.0435 508 0.7495 0.4544 0.7495 0.8657
0.2663 11.0870 510 0.7574 0.4774 0.7574 0.8703

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
10
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k18_task5_organization

Finetuned
(4222)
this model