ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k8_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9661
  • Qwk: 0.6533
  • Mse: 0.9661
  • Rmse: 0.9829

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0526 2 7.6389 -0.0211 7.6389 2.7639
No log 0.1053 4 4.8657 0.0818 4.8657 2.2058
No log 0.1579 6 3.2396 0.0791 3.2396 1.7999
No log 0.2105 8 2.6831 0.0649 2.6831 1.6380
No log 0.2632 10 2.6391 0.0851 2.6391 1.6245
No log 0.3158 12 1.8489 0.1754 1.8489 1.3597
No log 0.3684 14 1.7921 0.2321 1.7921 1.3387
No log 0.4211 16 1.9278 0.3051 1.9278 1.3885
No log 0.4737 18 2.2755 0.0851 2.2755 1.5085
No log 0.5263 20 2.3344 0.0972 2.3344 1.5279
No log 0.5789 22 2.3314 0.0966 2.3314 1.5269
No log 0.6316 24 2.1191 0.2014 2.1191 1.4557
No log 0.6842 26 2.0751 0.2797 2.0751 1.4405
No log 0.7368 28 1.7096 0.375 1.7096 1.3075
No log 0.7895 30 1.4355 0.4194 1.4355 1.1981
No log 0.8421 32 1.4234 0.4427 1.4234 1.1930
No log 0.8947 34 1.5939 0.3857 1.5939 1.2625
No log 0.9474 36 1.7861 0.3862 1.7861 1.3365
No log 1.0 38 1.6012 0.4143 1.6012 1.2654
No log 1.0526 40 1.2904 0.4889 1.2904 1.1360
No log 1.1053 42 1.2816 0.5612 1.2816 1.1321
No log 1.1579 44 1.5523 0.5098 1.5523 1.2459
No log 1.2105 46 1.3343 0.5655 1.3343 1.1551
No log 1.2632 48 1.0564 0.6345 1.0564 1.0278
No log 1.3158 50 1.0882 0.6275 1.0882 1.0432
No log 1.3684 52 1.3237 0.5181 1.3237 1.1505
No log 1.4211 54 1.1042 0.6429 1.1042 1.0508
No log 1.4737 56 0.8221 0.7273 0.8221 0.9067
No log 1.5263 58 0.8633 0.6711 0.8633 0.9292
No log 1.5789 60 0.8166 0.7205 0.8166 0.9037
No log 1.6316 62 0.8382 0.7362 0.8382 0.9156
No log 1.6842 64 0.9301 0.7081 0.9301 0.9644
No log 1.7368 66 1.1942 0.6228 1.1942 1.0928
No log 1.7895 68 1.1054 0.6415 1.1054 1.0514
No log 1.8421 70 0.7795 0.7211 0.7795 0.8829
No log 1.8947 72 0.9029 0.5957 0.9029 0.9502
No log 1.9474 74 0.9433 0.6143 0.9433 0.9712
No log 2.0 76 0.8929 0.6286 0.8929 0.9449
No log 2.0526 78 0.8549 0.775 0.8549 0.9246
No log 2.1053 80 0.8763 0.7547 0.8763 0.9361
No log 2.1579 82 1.0655 0.5694 1.0655 1.0322
No log 2.2105 84 1.9231 0.2029 1.9231 1.3868
No log 2.2632 86 2.1823 0.1000 2.1823 1.4773
No log 2.3158 88 1.7709 0.2206 1.7709 1.3308
No log 2.3684 90 1.0660 0.6056 1.0660 1.0325
No log 2.4211 92 0.8600 0.7160 0.8600 0.9274
No log 2.4737 94 0.8229 0.7020 0.8229 0.9071
No log 2.5263 96 0.8679 0.625 0.8679 0.9316
No log 2.5789 98 0.8836 0.5755 0.8836 0.9400
No log 2.6316 100 0.7671 0.7310 0.7671 0.8758
No log 2.6842 102 0.9109 0.6575 0.9109 0.9544
No log 2.7368 104 0.9860 0.5972 0.9860 0.9930
No log 2.7895 106 1.0828 0.5652 1.0828 1.0406
No log 2.8421 108 0.9009 0.6620 0.9009 0.9492
No log 2.8947 110 0.7725 0.7347 0.7725 0.8789
No log 2.9474 112 0.8007 0.7273 0.8007 0.8948
No log 3.0 114 1.0029 0.6946 1.0029 1.0014
No log 3.0526 116 1.3191 0.6316 1.3191 1.1485
No log 3.1053 118 1.1493 0.6328 1.1493 1.0720
No log 3.1579 120 0.9186 0.6923 0.9186 0.9584
No log 3.2105 122 0.8804 0.6857 0.8804 0.9383
No log 3.2632 124 0.8418 0.6986 0.8418 0.9175
No log 3.3158 126 0.9163 0.6538 0.9163 0.9572
No log 3.3684 128 1.0902 0.6211 1.0902 1.0442
No log 3.4211 130 0.9186 0.6234 0.9186 0.9584
No log 3.4737 132 0.7593 0.7123 0.7593 0.8714
No log 3.5263 134 0.8119 0.6809 0.8119 0.9011
No log 3.5789 136 0.8319 0.6475 0.8319 0.9121
No log 3.6316 138 0.8612 0.6277 0.8612 0.9280
No log 3.6842 140 0.8573 0.6187 0.8573 0.9259
No log 3.7368 142 1.0395 0.6111 1.0395 1.0196
No log 3.7895 144 1.1056 0.6301 1.1056 1.0515
No log 3.8421 146 0.9700 0.5857 0.9700 0.9849
No log 3.8947 148 0.8527 0.6853 0.8527 0.9234
No log 3.9474 150 0.8570 0.6713 0.8570 0.9257
No log 4.0 152 0.8303 0.6573 0.8303 0.9112
No log 4.0526 154 0.8659 0.6755 0.8659 0.9305
No log 4.1053 156 0.9780 0.6708 0.9780 0.9889
No log 4.1579 158 0.9356 0.6875 0.9356 0.9673
No log 4.2105 160 0.9523 0.7089 0.9523 0.9759
No log 4.2632 162 1.1579 0.6420 1.1579 1.0761
No log 4.3158 164 1.3176 0.6512 1.3176 1.1479
No log 4.3684 166 1.0502 0.6667 1.0502 1.0248
No log 4.4211 168 0.8500 0.6842 0.8500 0.9220
No log 4.4737 170 0.7833 0.7484 0.7833 0.8850
No log 4.5263 172 0.7379 0.6897 0.7379 0.8590
No log 4.5789 174 0.7630 0.6897 0.7630 0.8735
No log 4.6316 176 0.8016 0.6809 0.8016 0.8953
No log 4.6842 178 0.9337 0.64 0.9337 0.9663
No log 4.7368 180 0.9434 0.6623 0.9434 0.9713
No log 4.7895 182 0.8336 0.7564 0.8336 0.9130
No log 4.8421 184 0.8470 0.6667 0.8470 0.9203
No log 4.8947 186 0.8243 0.6901 0.8243 0.9079
No log 4.9474 188 0.8392 0.6806 0.8392 0.9161
No log 5.0 190 0.8863 0.6974 0.8863 0.9414
No log 5.0526 192 0.8756 0.7190 0.8756 0.9357
No log 5.1053 194 1.0258 0.6667 1.0258 1.0128
No log 5.1579 196 1.0584 0.6667 1.0584 1.0288
No log 5.2105 198 0.8801 0.6795 0.8801 0.9381
No log 5.2632 200 0.7223 0.7248 0.7223 0.8499
No log 5.3158 202 0.7696 0.7397 0.7696 0.8773
No log 5.3684 204 0.7817 0.6986 0.7817 0.8841
No log 5.4211 206 1.0869 0.6826 1.0869 1.0425
No log 5.4737 208 1.6065 0.5397 1.6065 1.2675
No log 5.5263 210 1.9285 0.4600 1.9285 1.3887
No log 5.5789 212 1.4636 0.5946 1.4636 1.2098
No log 5.6316 214 0.8008 0.7162 0.8008 0.8949
No log 5.6842 216 0.7553 0.7042 0.7553 0.8691
No log 5.7368 218 0.7599 0.7143 0.7599 0.8717
No log 5.7895 220 0.8472 0.6370 0.8472 0.9205
No log 5.8421 222 0.9274 0.6107 0.9274 0.9630
No log 5.8947 224 0.8506 0.6107 0.8506 0.9223
No log 5.9474 226 0.7744 0.6569 0.7744 0.8800
No log 6.0 228 0.6893 0.7083 0.6893 0.8302
No log 6.0526 230 0.6666 0.7222 0.6666 0.8165
No log 6.1053 232 0.6949 0.6944 0.6949 0.8336
No log 6.1579 234 0.7023 0.7183 0.7023 0.8380
No log 6.2105 236 0.7352 0.6950 0.7352 0.8574
No log 6.2632 238 0.8826 0.6176 0.8826 0.9395
No log 6.3158 240 1.0331 0.5333 1.0331 1.0164
No log 6.3684 242 0.9794 0.5926 0.9794 0.9896
No log 6.4211 244 0.8436 0.6074 0.8436 0.9185
No log 6.4737 246 0.7558 0.7190 0.7558 0.8694
No log 6.5263 248 0.7276 0.7389 0.7276 0.8530
No log 6.5789 250 0.8921 0.7152 0.8921 0.9445
No log 6.6316 252 1.2593 0.6557 1.2593 1.1222
No log 6.6842 254 1.3688 0.6455 1.3688 1.1700
No log 6.7368 256 1.0076 0.6851 1.0076 1.0038
No log 6.7895 258 0.6899 0.7831 0.6899 0.8306
No log 6.8421 260 0.6405 0.7285 0.6405 0.8003
No log 6.8947 262 0.6900 0.7162 0.6900 0.8307
No log 6.9474 264 0.8192 0.7190 0.8192 0.9051
No log 7.0 266 1.0228 0.6538 1.0228 1.0113
No log 7.0526 268 1.1707 0.6584 1.1707 1.0820
No log 7.1053 270 1.0783 0.6144 1.0783 1.0384
No log 7.1579 272 0.9233 0.5152 0.9233 0.9609
No log 7.2105 274 0.8882 0.6269 0.8882 0.9425
No log 7.2632 276 0.8130 0.6906 0.8130 0.9016
No log 7.3158 278 0.7124 0.7260 0.7124 0.8440
No log 7.3684 280 0.7290 0.7347 0.7290 0.8538
No log 7.4211 282 0.8039 0.6667 0.8039 0.8966
No log 7.4737 284 0.8975 0.6581 0.8975 0.9474
No log 7.5263 286 0.8974 0.6709 0.8974 0.9473
No log 7.5789 288 0.8420 0.6667 0.8420 0.9176
No log 7.6316 290 0.8084 0.6667 0.8084 0.8991
No log 7.6842 292 0.7627 0.7075 0.7627 0.8733
No log 7.7368 294 0.7159 0.7383 0.7159 0.8461
No log 7.7895 296 0.7915 0.6883 0.7915 0.8897
No log 7.8421 298 0.8638 0.6709 0.8638 0.9294
No log 7.8947 300 0.9147 0.6709 0.9147 0.9564
No log 7.9474 302 0.9067 0.6497 0.9067 0.9522
No log 8.0 304 0.7818 0.7237 0.7818 0.8842
No log 8.0526 306 0.7688 0.72 0.7688 0.8768
No log 8.1053 308 0.7636 0.6980 0.7636 0.8739
No log 8.1579 310 0.8453 0.6928 0.8453 0.9194
No log 8.2105 312 1.0453 0.6626 1.0453 1.0224
No log 8.2632 314 1.0787 0.6626 1.0787 1.0386
No log 8.3158 316 0.9227 0.6623 0.9227 0.9606
No log 8.3684 318 0.7483 0.6809 0.7483 0.8650
No log 8.4211 320 0.7215 0.7092 0.7215 0.8494
No log 8.4737 322 0.7252 0.6857 0.7252 0.8516
No log 8.5263 324 0.7762 0.6809 0.7762 0.8810
No log 8.5789 326 1.0727 0.6536 1.0727 1.0357
No log 8.6316 328 1.2539 0.6391 1.2539 1.1198
No log 8.6842 330 1.0768 0.6543 1.0768 1.0377
No log 8.7368 332 0.7971 0.6974 0.7971 0.8928
No log 8.7895 334 0.6205 0.7448 0.6205 0.7877
No log 8.8421 336 0.7630 0.7361 0.7630 0.8735
No log 8.8947 338 0.8539 0.7133 0.8539 0.9240
No log 8.9474 340 0.7358 0.7552 0.7358 0.8578
No log 9.0 342 0.6685 0.7297 0.6685 0.8176
No log 9.0526 344 0.7867 0.6887 0.7867 0.8870
No log 9.1053 346 0.8353 0.6575 0.8353 0.9139
No log 9.1579 348 0.7967 0.6331 0.7967 0.8926
No log 9.2105 350 0.6959 0.7083 0.6959 0.8342
No log 9.2632 352 0.6727 0.7465 0.6727 0.8202
No log 9.3158 354 0.6768 0.7465 0.6768 0.8227
No log 9.3684 356 0.6848 0.7234 0.6848 0.8275
No log 9.4211 358 0.7323 0.7222 0.7323 0.8558
No log 9.4737 360 0.8868 0.6087 0.8868 0.9417
No log 9.5263 362 1.0768 0.6174 1.0768 1.0377
No log 9.5789 364 1.1301 0.6174 1.1301 1.0630
No log 9.6316 366 1.0379 0.6099 1.0379 1.0188
No log 9.6842 368 0.9133 0.6232 0.9133 0.9557
No log 9.7368 370 0.8531 0.6475 0.8531 0.9237
No log 9.7895 372 0.8084 0.6761 0.8084 0.8991
No log 9.8421 374 0.8364 0.6345 0.8364 0.9146
No log 9.8947 376 0.8414 0.6711 0.8414 0.9173
No log 9.9474 378 0.8869 0.6709 0.8869 0.9417
No log 10.0 380 0.8906 0.6709 0.8906 0.9437
No log 10.0526 382 0.8149 0.6538 0.8149 0.9027
No log 10.1053 384 0.7975 0.6538 0.7975 0.8930
No log 10.1579 386 0.8452 0.6538 0.8452 0.9193
No log 10.2105 388 0.8952 0.6788 0.8952 0.9462
No log 10.2632 390 0.8600 0.6867 0.8600 0.9274
No log 10.3158 392 0.8030 0.7066 0.8030 0.8961
No log 10.3684 394 0.8405 0.6867 0.8405 0.9168
No log 10.4211 396 0.8132 0.6871 0.8132 0.9018
No log 10.4737 398 0.6951 0.7172 0.6951 0.8337
No log 10.5263 400 0.6448 0.7973 0.6448 0.8030
No log 10.5789 402 0.6877 0.7755 0.6877 0.8293
No log 10.6316 404 0.6964 0.7919 0.6964 0.8345
No log 10.6842 406 0.6665 0.7568 0.6665 0.8164
No log 10.7368 408 0.6537 0.7568 0.6537 0.8085
No log 10.7895 410 0.6600 0.7871 0.6600 0.8124
No log 10.8421 412 0.6712 0.7975 0.6712 0.8193
No log 10.8947 414 0.7108 0.7451 0.7108 0.8431
No log 10.9474 416 0.7142 0.7451 0.7142 0.8451
No log 11.0 418 0.6656 0.7763 0.6656 0.8158
No log 11.0526 420 0.6437 0.7763 0.6437 0.8023
No log 11.1053 422 0.6429 0.7742 0.6429 0.8018
No log 11.1579 424 0.6500 0.7785 0.6500 0.8062
No log 11.2105 426 0.6700 0.7703 0.6700 0.8185
No log 11.2632 428 0.7622 0.6667 0.7622 0.8731
No log 11.3158 430 0.8510 0.6197 0.8510 0.9225
No log 11.3684 432 0.8056 0.6294 0.8056 0.8975
No log 11.4211 434 0.6955 0.7172 0.6955 0.8340
No log 11.4737 436 0.6657 0.7619 0.6657 0.8159
No log 11.5263 438 0.6683 0.7397 0.6683 0.8175
No log 11.5789 440 0.6896 0.7310 0.6896 0.8305
No log 11.6316 442 0.7107 0.7310 0.7107 0.8430
No log 11.6842 444 0.7460 0.7310 0.7460 0.8637
No log 11.7368 446 0.7667 0.6993 0.7667 0.8756
No log 11.7895 448 0.7987 0.6383 0.7987 0.8937
No log 11.8421 450 0.8240 0.6753 0.8240 0.9077
No log 11.8947 452 0.8190 0.6962 0.8190 0.9050
No log 11.9474 454 0.7570 0.7654 0.7570 0.8701
No log 12.0 456 0.7621 0.7317 0.7621 0.8730
No log 12.0526 458 0.7260 0.7702 0.7260 0.8521
No log 12.1053 460 0.7029 0.7310 0.7029 0.8384
No log 12.1579 462 0.7464 0.7133 0.7464 0.8639
No log 12.2105 464 0.8004 0.6232 0.8004 0.8946
No log 12.2632 466 0.7673 0.6901 0.7673 0.8759
No log 12.3158 468 0.7428 0.6853 0.7428 0.8619
No log 12.3684 470 0.7535 0.7237 0.7535 0.8680
No log 12.4211 472 0.7387 0.7516 0.7387 0.8594
No log 12.4737 474 0.6761 0.8075 0.6761 0.8222
No log 12.5263 476 0.6569 0.8272 0.6569 0.8105
No log 12.5789 478 0.6779 0.7975 0.6779 0.8233
No log 12.6316 480 0.7758 0.75 0.7758 0.8808
No log 12.6842 482 0.8481 0.6710 0.8481 0.9209
No log 12.7368 484 0.8547 0.6301 0.8547 0.9245
No log 12.7895 486 0.8352 0.5735 0.8352 0.9139
No log 12.8421 488 0.8434 0.6074 0.8434 0.9184
No log 12.8947 490 0.8093 0.6423 0.8093 0.8996
No log 12.9474 492 0.7863 0.6619 0.7863 0.8867
No log 13.0 494 0.8107 0.6620 0.8107 0.9004
No log 13.0526 496 0.7981 0.6573 0.7981 0.8934
No log 13.1053 498 0.7446 0.7027 0.7446 0.8629
0.359 13.1579 500 0.7118 0.7547 0.7118 0.8437
0.359 13.2105 502 0.6742 0.8242 0.6742 0.8211
0.359 13.2632 504 0.6670 0.8171 0.6670 0.8167
0.359 13.3158 506 0.6701 0.7397 0.6701 0.8186
0.359 13.3684 508 0.7000 0.7172 0.7000 0.8367
0.359 13.4211 510 0.7466 0.6853 0.7466 0.8641
0.359 13.4737 512 0.8463 0.6099 0.8463 0.9199
0.359 13.5263 514 0.9661 0.6533 0.9661 0.9829

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
182
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k8_task1_organization

Finetuned
(4222)
this model