ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k8_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0039
  • Qwk: 0.3923
  • Mse: 1.0039
  • Rmse: 1.0019

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0690 2 4.8026 0.0010 4.8026 2.1915
No log 0.1379 4 2.8954 -0.0191 2.8954 1.7016
No log 0.2069 6 2.0638 -0.0575 2.0638 1.4366
No log 0.2759 8 2.7687 -0.0903 2.7687 1.6639
No log 0.3448 10 2.7343 -0.1004 2.7343 1.6536
No log 0.4138 12 2.2382 -0.1191 2.2382 1.4961
No log 0.4828 14 1.5683 0.0477 1.5683 1.2523
No log 0.5517 16 1.4376 0.1659 1.4376 1.1990
No log 0.6207 18 1.4091 0.0987 1.4091 1.1871
No log 0.6897 20 1.4297 0.1282 1.4297 1.1957
No log 0.7586 22 1.7028 0.0372 1.7028 1.3049
No log 0.8276 24 1.9029 0.0504 1.9029 1.3794
No log 0.8966 26 2.1552 -0.0425 2.1552 1.4681
No log 0.9655 28 1.7814 -0.0080 1.7814 1.3347
No log 1.0345 30 1.1898 0.1715 1.1898 1.0908
No log 1.1034 32 1.1012 0.2886 1.1012 1.0494
No log 1.1724 34 1.1135 0.2836 1.1135 1.0552
No log 1.2414 36 1.1898 0.1622 1.1898 1.0908
No log 1.3103 38 1.4510 0.0403 1.4510 1.2046
No log 1.3793 40 1.6396 0.0084 1.6396 1.2805
No log 1.4483 42 1.8247 0.0369 1.8247 1.3508
No log 1.5172 44 1.7427 0.0812 1.7427 1.3201
No log 1.5862 46 1.4549 0.1565 1.4549 1.2062
No log 1.6552 48 1.4131 0.1403 1.4131 1.1887
No log 1.7241 50 1.7525 0.1154 1.7525 1.3238
No log 1.7931 52 1.9619 0.0474 1.9619 1.4007
No log 1.8621 54 1.6985 0.1273 1.6985 1.3033
No log 1.9310 56 1.4646 0.2227 1.4646 1.2102
No log 2.0 58 1.2687 0.1542 1.2687 1.1264
No log 2.0690 60 1.3552 0.2191 1.3552 1.1641
No log 2.1379 62 1.5384 0.1902 1.5384 1.2403
No log 2.2069 64 1.7362 0.1249 1.7362 1.3176
No log 2.2759 66 1.5147 0.0750 1.5147 1.2307
No log 2.3448 68 1.3504 0.0362 1.3504 1.1621
No log 2.4138 70 1.2090 0.1622 1.2090 1.0996
No log 2.4828 72 1.1720 0.1542 1.1720 1.0826
No log 2.5517 74 1.1748 0.2137 1.1748 1.0839
No log 2.6207 76 1.1546 0.2886 1.1546 1.0745
No log 2.6897 78 1.1963 0.2636 1.1963 1.0937
No log 2.7586 80 1.4567 0.1346 1.4567 1.2069
No log 2.8276 82 1.9059 -0.0098 1.9059 1.3805
No log 2.8966 84 1.9289 0.0399 1.9289 1.3889
No log 2.9655 86 1.6718 0.0707 1.6718 1.2930
No log 3.0345 88 1.3268 0.1495 1.3268 1.1519
No log 3.1034 90 1.1708 0.1565 1.1708 1.0820
No log 3.1724 92 1.1365 0.2191 1.1365 1.0661
No log 3.2414 94 1.1863 0.1715 1.1863 1.0892
No log 3.3103 96 1.3321 0.1827 1.3321 1.1542
No log 3.3793 98 1.5690 0.1351 1.5690 1.2526
No log 3.4483 100 1.7978 0.1397 1.7978 1.3408
No log 3.5172 102 1.7191 0.1530 1.7191 1.3112
No log 3.5862 104 1.5373 0.2406 1.5373 1.2399
No log 3.6552 106 1.6535 0.1735 1.6535 1.2859
No log 3.7241 108 1.9698 0.0660 1.9698 1.4035
No log 3.7931 110 1.9101 0.0564 1.9101 1.3820
No log 3.8621 112 1.4911 0.2201 1.4911 1.2211
No log 3.9310 114 1.4363 0.2292 1.4363 1.1984
No log 4.0 116 1.4751 0.1027 1.4751 1.2145
No log 4.0690 118 1.6144 0.1279 1.6144 1.2706
No log 4.1379 120 1.5285 0.1279 1.5285 1.2363
No log 4.2069 122 1.2644 0.2005 1.2644 1.1245
No log 4.2759 124 1.0775 0.2914 1.0775 1.0380
No log 4.3448 126 1.0635 0.2914 1.0635 1.0313
No log 4.4138 128 1.1655 0.2150 1.1655 1.0796
No log 4.4828 130 1.4015 0.1168 1.4015 1.1838
No log 4.5517 132 1.6588 0.0718 1.6588 1.2879
No log 4.6207 134 1.6725 0.0162 1.6725 1.2933
No log 4.6897 136 1.4430 0.2362 1.4430 1.2013
No log 4.7586 138 1.2620 0.2498 1.2620 1.1234
No log 4.8276 140 1.1573 0.3152 1.1573 1.0758
No log 4.8966 142 1.1871 0.2864 1.1871 1.0896
No log 4.9655 144 1.1638 0.2718 1.1638 1.0788
No log 5.0345 146 1.2325 0.2938 1.2325 1.1102
No log 5.1034 148 1.6091 0.1731 1.6091 1.2685
No log 5.1724 150 1.7418 0.1289 1.7418 1.3198
No log 5.2414 152 1.5637 0.1731 1.5637 1.2505
No log 5.3103 154 1.4197 0.2763 1.4197 1.1915
No log 5.3793 156 1.2476 0.2324 1.2476 1.1170
No log 5.4483 158 1.2636 0.2640 1.2636 1.1241
No log 5.5172 160 1.4660 0.1901 1.4660 1.2108
No log 5.5862 162 1.4838 0.1602 1.4838 1.2181
No log 5.6552 164 1.3581 0.1743 1.3581 1.1654
No log 5.7241 166 1.1977 0.3218 1.1977 1.0944
No log 5.7931 168 1.1491 0.3491 1.1491 1.0720
No log 5.8621 170 1.1900 0.3437 1.1900 1.0909
No log 5.9310 172 1.1824 0.3295 1.1824 1.0874
No log 6.0 174 1.3602 0.1417 1.3602 1.1663
No log 6.0690 176 1.5226 0.1715 1.5226 1.2339
No log 6.1379 178 1.4407 0.1277 1.4407 1.2003
No log 6.2069 180 1.1509 0.3675 1.1509 1.0728
No log 6.2759 182 0.9603 0.2784 0.9603 0.9799
No log 6.3448 184 0.9502 0.3115 0.9502 0.9748
No log 6.4138 186 1.0707 0.3790 1.0707 1.0347
No log 6.4828 188 1.4149 0.3333 1.4149 1.1895
No log 6.5517 190 1.6028 0.2015 1.6028 1.2660
No log 6.6207 192 1.4070 0.3333 1.4070 1.1862
No log 6.6897 194 1.1361 0.3601 1.1361 1.0659
No log 6.7586 196 1.0144 0.2665 1.0144 1.0072
No log 6.8276 198 1.0011 0.2562 1.0011 1.0006
No log 6.8966 200 1.0349 0.2815 1.0349 1.0173
No log 6.9655 202 1.2121 0.3739 1.2121 1.1009
No log 7.0345 204 1.3373 0.3738 1.3373 1.1564
No log 7.1034 206 1.2341 0.3868 1.2341 1.1109
No log 7.1724 208 1.0480 0.3532 1.0480 1.0237
No log 7.2414 210 1.0148 0.2596 1.0148 1.0074
No log 7.3103 212 1.0315 0.2596 1.0315 1.0156
No log 7.3793 214 1.0643 0.3218 1.0643 1.0316
No log 7.4483 216 1.1199 0.3078 1.1199 1.0583
No log 7.5172 218 1.1550 0.2900 1.1550 1.0747
No log 7.5862 220 1.1679 0.3497 1.1679 1.0807
No log 7.6552 222 1.3321 0.3470 1.3321 1.1542
No log 7.7241 224 1.3512 0.3470 1.3512 1.1624
No log 7.7931 226 1.2404 0.3005 1.2404 1.1137
No log 7.8621 228 1.2289 0.3307 1.2289 1.1086
No log 7.9310 230 1.2439 0.3307 1.2439 1.1153
No log 8.0 232 1.1831 0.3521 1.1831 1.0877
No log 8.0690 234 1.0608 0.3256 1.0608 1.0299
No log 8.1379 236 1.0414 0.3402 1.0414 1.0205
No log 8.2069 238 1.1081 0.3973 1.1081 1.0527
No log 8.2759 240 1.1211 0.3633 1.1211 1.0588
No log 8.3448 242 1.0427 0.3546 1.0427 1.0211
No log 8.4138 244 1.0327 0.4074 1.0327 1.0162
No log 8.4828 246 1.0503 0.4059 1.0503 1.0248
No log 8.5517 248 1.0557 0.4098 1.0557 1.0274
No log 8.6207 250 0.9891 0.3614 0.9891 0.9945
No log 8.6897 252 0.9672 0.3059 0.9672 0.9835
No log 8.7586 254 0.9797 0.3943 0.9797 0.9898
No log 8.8276 256 1.0018 0.3720 1.0018 1.0009
No log 8.8966 258 0.9992 0.3720 0.9992 0.9996
No log 8.9655 260 0.9707 0.2871 0.9707 0.9853
No log 9.0345 262 0.9443 0.3399 0.9443 0.9718
No log 9.1034 264 0.9528 0.3443 0.9528 0.9761
No log 9.1724 266 0.9629 0.3443 0.9629 0.9813
No log 9.2414 268 1.0022 0.3433 1.0022 1.0011
No log 9.3103 270 0.9719 0.3154 0.9719 0.9859
No log 9.3793 272 0.9299 0.3738 0.9299 0.9643
No log 9.4483 274 0.9388 0.3692 0.9388 0.9689
No log 9.5172 276 0.9621 0.3463 0.9621 0.9809
No log 9.5862 278 0.9972 0.3556 0.9972 0.9986
No log 9.6552 280 1.0336 0.3243 1.0336 1.0167
No log 9.7241 282 1.0169 0.3675 1.0169 1.0084
No log 9.7931 284 1.0020 0.3338 1.0020 1.0010
No log 9.8621 286 1.0088 0.3433 1.0088 1.0044
No log 9.9310 288 1.0109 0.3433 1.0109 1.0055
No log 10.0 290 0.9798 0.3328 0.9798 0.9898
No log 10.0690 292 0.9633 0.4019 0.9633 0.9815
No log 10.1379 294 0.9978 0.4197 0.9978 0.9989
No log 10.2069 296 1.0437 0.3657 1.0437 1.0216
No log 10.2759 298 1.0307 0.3765 1.0307 1.0152
No log 10.3448 300 0.9918 0.3723 0.9918 0.9959
No log 10.4138 302 0.9577 0.3879 0.9577 0.9786
No log 10.4828 304 0.9551 0.4019 0.9551 0.9773
No log 10.5517 306 0.9948 0.3390 0.9948 0.9974
No log 10.6207 308 1.0650 0.3976 1.0650 1.0320
No log 10.6897 310 1.0178 0.3584 1.0178 1.0089
No log 10.7586 312 0.9986 0.3529 0.9986 0.9993
No log 10.8276 314 0.9926 0.3486 0.9926 0.9963
No log 10.8966 316 0.9879 0.3486 0.9879 0.9939
No log 10.9655 318 0.9663 0.3346 0.9663 0.9830
No log 11.0345 320 0.9669 0.3218 0.9669 0.9833
No log 11.1034 322 0.9941 0.4155 0.9941 0.9970
No log 11.1724 324 0.9735 0.3750 0.9735 0.9867
No log 11.2414 326 0.9227 0.3738 0.9227 0.9605
No log 11.3103 328 0.9262 0.3738 0.9262 0.9624
No log 11.3793 330 0.9411 0.3681 0.9411 0.9701
No log 11.4483 332 1.0288 0.4155 1.0288 1.0143
No log 11.5172 334 1.0996 0.3207 1.0996 1.0486
No log 11.5862 336 1.0848 0.3521 1.0848 1.0415
No log 11.6552 338 1.0013 0.3699 1.0013 1.0007
No log 11.7241 340 0.9345 0.4019 0.9345 0.9667
No log 11.7931 342 0.9097 0.3541 0.9097 0.9538
No log 11.8621 344 0.8875 0.3639 0.8875 0.9421
No log 11.9310 346 0.8983 0.3913 0.8983 0.9478
No log 12.0 348 0.9133 0.4045 0.9133 0.9556
No log 12.0690 350 0.9228 0.3813 0.9228 0.9606
No log 12.1379 352 0.9489 0.3942 0.9489 0.9741
No log 12.2069 354 0.9793 0.4283 0.9794 0.9896
No log 12.2759 356 0.9702 0.4023 0.9702 0.9850
No log 12.3448 358 0.9377 0.4082 0.9377 0.9684
No log 12.4138 360 0.9296 0.3938 0.9296 0.9641
No log 12.4828 362 0.9313 0.3237 0.9313 0.9651
No log 12.5517 364 0.9417 0.3869 0.9417 0.9704
No log 12.6207 366 0.9557 0.3448 0.9557 0.9776
No log 12.6897 368 0.9660 0.3852 0.9660 0.9829
No log 12.7586 370 0.9589 0.3806 0.9589 0.9792
No log 12.8276 372 0.9400 0.4098 0.9400 0.9695
No log 12.8966 374 0.9270 0.3636 0.9270 0.9628
No log 12.9655 376 0.9039 0.3636 0.9039 0.9507
No log 13.0345 378 0.9395 0.3939 0.9395 0.9693
No log 13.1034 380 1.0285 0.4383 1.0285 1.0142
No log 13.1724 382 1.0085 0.4136 1.0085 1.0042
No log 13.2414 384 0.9542 0.3939 0.9542 0.9768
No log 13.3103 386 0.9232 0.3862 0.9232 0.9608
No log 13.3793 388 0.9511 0.3891 0.9511 0.9752
No log 13.4483 390 1.0128 0.3930 1.0128 1.0064
No log 13.5172 392 1.0994 0.3523 1.0994 1.0485
No log 13.5862 394 1.0877 0.3561 1.0877 1.0429
No log 13.6552 396 0.9954 0.4250 0.9954 0.9977
No log 13.7241 398 0.8977 0.3855 0.8977 0.9475
No log 13.7931 400 0.8671 0.3821 0.8671 0.9312
No log 13.8621 402 0.8467 0.4197 0.8467 0.9201
No log 13.9310 404 0.8699 0.4549 0.8699 0.9327
No log 14.0 406 0.9548 0.4668 0.9548 0.9771
No log 14.0690 408 1.0019 0.4833 1.0019 1.0010
No log 14.1379 410 0.9540 0.4645 0.9540 0.9768
No log 14.2069 412 0.8928 0.4341 0.8928 0.9449
No log 14.2759 414 0.8590 0.3541 0.8590 0.9268
No log 14.3448 416 0.8630 0.4023 0.8630 0.9290
No log 14.4138 418 0.8665 0.4120 0.8665 0.9309
No log 14.4828 420 0.8858 0.4175 0.8858 0.9412
No log 14.5517 422 0.9172 0.4341 0.9172 0.9577
No log 14.6207 424 0.9616 0.4155 0.9616 0.9806
No log 14.6897 426 0.9911 0.4227 0.9911 0.9955
No log 14.7586 428 0.9497 0.4410 0.9497 0.9745
No log 14.8276 430 0.9490 0.4283 0.9490 0.9741
No log 14.8966 432 0.9241 0.4469 0.9241 0.9613
No log 14.9655 434 0.9097 0.4469 0.9097 0.9538
No log 15.0345 436 0.9227 0.4469 0.9227 0.9606
No log 15.1034 438 0.9422 0.4723 0.9422 0.9707
No log 15.1724 440 0.9158 0.4439 0.9158 0.9570
No log 15.2414 442 0.8860 0.3443 0.8860 0.9413
No log 15.3103 444 0.8909 0.3539 0.8909 0.9439
No log 15.3793 446 0.8968 0.3354 0.8968 0.9470
No log 15.4483 448 0.9000 0.3458 0.9000 0.9487
No log 15.5172 450 0.9316 0.4774 0.9316 0.9652
No log 15.5862 452 0.9552 0.4372 0.9552 0.9774
No log 15.6552 454 1.0092 0.4225 1.0092 1.0046
No log 15.7241 456 0.9859 0.4406 0.9859 0.9929
No log 15.7931 458 0.9153 0.3637 0.9153 0.9567
No log 15.8621 460 0.9020 0.3838 0.9020 0.9498
No log 15.9310 462 0.8998 0.4646 0.8998 0.9486
No log 16.0 464 0.9057 0.3891 0.9057 0.9517
No log 16.0690 466 0.9536 0.3544 0.9536 0.9765
No log 16.1379 468 1.0314 0.4050 1.0314 1.0156
No log 16.2069 470 0.9973 0.3719 0.9973 0.9987
No log 16.2759 472 0.9330 0.3993 0.9330 0.9659
No log 16.3448 474 0.9225 0.4007 0.9225 0.9605
No log 16.4138 476 0.9214 0.3875 0.9214 0.9599
No log 16.4828 478 0.9494 0.3813 0.9494 0.9743
No log 16.5517 480 0.9860 0.3480 0.9860 0.9930
No log 16.6207 482 1.0043 0.3790 1.0043 1.0021
No log 16.6897 484 0.9579 0.3806 0.9579 0.9787
No log 16.7586 486 0.9147 0.3437 0.9147 0.9564
No log 16.8276 488 0.9284 0.3720 0.9284 0.9635
No log 16.8966 490 0.9297 0.3463 0.9297 0.9642
No log 16.9655 492 0.9230 0.3570 0.9230 0.9607
No log 17.0345 494 0.9338 0.3614 0.9338 0.9664
No log 17.1034 496 0.9762 0.4155 0.9762 0.9880
No log 17.1724 498 0.9983 0.4318 0.9983 0.9991
0.3653 17.2414 500 0.9696 0.4318 0.9696 0.9847
0.3653 17.3103 502 0.9514 0.4063 0.9514 0.9754
0.3653 17.3793 504 0.9092 0.3154 0.9092 0.9535
0.3653 17.4483 506 0.9021 0.3205 0.9021 0.9498
0.3653 17.5172 508 0.9216 0.3412 0.9216 0.9600
0.3653 17.5862 510 1.0039 0.3923 1.0039 1.0019

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
184
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k8_task2_organization

Finetuned
(4222)
this model