ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k9_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9683
  • Qwk: -0.0855
  • Mse: 0.9683
  • Rmse: 0.9840

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0769 2 3.5915 -0.0053 3.5915 1.8951
No log 0.1538 4 1.8903 0.0810 1.8903 1.3749
No log 0.2308 6 1.8124 0.0304 1.8124 1.3462
No log 0.3077 8 1.1491 -0.0658 1.1491 1.0720
No log 0.3846 10 1.1910 -0.0931 1.1910 1.0913
No log 0.4615 12 0.8592 0.0676 0.8592 0.9269
No log 0.5385 14 0.7360 0.0909 0.7360 0.8579
No log 0.6154 16 0.7420 0.0416 0.7420 0.8614
No log 0.6923 18 0.8631 0.0316 0.8631 0.9290
No log 0.7692 20 0.9048 0.0873 0.9048 0.9512
No log 0.8462 22 0.7461 0.0296 0.7461 0.8638
No log 0.9231 24 0.6739 -0.0035 0.6739 0.8209
No log 1.0 26 0.8077 0.1097 0.8077 0.8987
No log 1.0769 28 1.7620 -0.0497 1.7620 1.3274
No log 1.1538 30 1.2722 -0.0500 1.2722 1.1279
No log 1.2308 32 0.8123 -0.0240 0.8123 0.9013
No log 1.3077 34 0.7968 -0.0753 0.7968 0.8926
No log 1.3846 36 0.6846 0.0 0.6846 0.8274
No log 1.4615 38 0.6912 0.0 0.6912 0.8314
No log 1.5385 40 0.7697 -0.0131 0.7697 0.8773
No log 1.6154 42 0.8234 -0.1701 0.8234 0.9074
No log 1.6923 44 0.7873 -0.1239 0.7873 0.8873
No log 1.7692 46 0.7139 -0.0035 0.7139 0.8450
No log 1.8462 48 0.6807 0.0555 0.6807 0.8251
No log 1.9231 50 0.7423 -0.0131 0.7423 0.8616
No log 2.0 52 0.9671 -0.0236 0.9671 0.9834
No log 2.0769 54 1.3617 -0.0457 1.3617 1.1669
No log 2.1538 56 1.2444 -0.0457 1.2444 1.1155
No log 2.2308 58 0.8381 -0.1251 0.8381 0.9155
No log 2.3077 60 0.7028 0.0555 0.7028 0.8384
No log 2.3846 62 0.6856 0.0555 0.6856 0.8280
No log 2.4615 64 0.6851 0.0555 0.6851 0.8277
No log 2.5385 66 0.7256 -0.0035 0.7256 0.8518
No log 2.6154 68 0.9222 -0.0331 0.9222 0.9603
No log 2.6923 70 1.0254 -0.1019 1.0254 1.0126
No log 2.7692 72 0.9107 0.1114 0.9107 0.9543
No log 2.8462 74 0.7704 0.0296 0.7704 0.8777
No log 2.9231 76 0.7172 -0.0035 0.7172 0.8468
No log 3.0 78 0.7684 -0.1236 0.7684 0.8766
No log 3.0769 80 0.9354 -0.0253 0.9354 0.9672
No log 3.1538 82 0.8962 -0.0122 0.8962 0.9467
No log 3.2308 84 0.7064 -0.0035 0.7064 0.8405
No log 3.3077 86 0.7405 0.0555 0.7405 0.8605
No log 3.3846 88 0.8036 0.1694 0.8036 0.8965
No log 3.4615 90 0.7915 0.0 0.7915 0.8897
No log 3.5385 92 0.6897 0.0355 0.6897 0.8305
No log 3.6154 94 0.8380 0.1150 0.8380 0.9154
No log 3.6923 96 0.7355 0.1449 0.7355 0.8576
No log 3.7692 98 0.7026 -0.0626 0.7026 0.8382
No log 3.8462 100 0.7731 0.1143 0.7731 0.8792
No log 3.9231 102 0.7678 -0.0644 0.7678 0.8763
No log 4.0 104 0.7413 0.0555 0.7413 0.8610
No log 4.0769 106 0.7434 -0.0188 0.7434 0.8622
No log 4.1538 108 0.9720 -0.0163 0.9720 0.9859
No log 4.2308 110 0.8984 -0.0056 0.8984 0.9478
No log 4.3077 112 0.7737 0.0191 0.7737 0.8796
No log 4.3846 114 0.7644 -0.0101 0.7644 0.8743
No log 4.4615 116 0.7788 0.0 0.7788 0.8825
No log 4.5385 118 0.8152 -0.0406 0.8152 0.9029
No log 4.6154 120 0.7580 -0.1067 0.7580 0.8706
No log 4.6923 122 0.7034 0.1202 0.7034 0.8387
No log 4.7692 124 0.7122 0.1627 0.7122 0.8439
No log 4.8462 126 0.7009 0.0857 0.7009 0.8372
No log 4.9231 128 0.7488 -0.1067 0.7488 0.8653
No log 5.0 130 0.7902 0.0 0.7902 0.8889
No log 5.0769 132 0.7500 0.0 0.7500 0.8660
No log 5.1538 134 0.7251 0.0909 0.7251 0.8515
No log 5.2308 136 0.7198 -0.0152 0.7198 0.8484
No log 5.3077 138 0.7265 0.1146 0.7265 0.8523
No log 5.3846 140 0.7329 0.0110 0.7329 0.8561
No log 5.4615 142 0.8956 0.0433 0.8956 0.9464
No log 5.5385 144 0.8318 0.0392 0.8318 0.9120
No log 5.6154 146 0.7123 0.0814 0.7123 0.8440
No log 5.6923 148 0.9094 0.1273 0.9094 0.9536
No log 5.7692 150 0.8671 0.1360 0.8671 0.9312
No log 5.8462 152 0.7296 0.0981 0.7296 0.8542
No log 5.9231 154 0.8556 0.0008 0.8556 0.9250
No log 6.0 156 0.8100 0.0328 0.8100 0.9000
No log 6.0769 158 0.6967 0.0964 0.6967 0.8347
No log 6.1538 160 0.8893 0.1788 0.8893 0.9430
No log 6.2308 162 0.9598 0.0805 0.9598 0.9797
No log 6.3077 164 0.7592 0.1836 0.7592 0.8713
No log 6.3846 166 0.6565 0.0 0.6565 0.8103
No log 6.4615 168 0.6691 0.0 0.6691 0.8180
No log 6.5385 170 0.6382 0.0416 0.6382 0.7989
No log 6.6154 172 0.6693 0.1565 0.6693 0.8181
No log 6.6923 174 0.6736 0.1565 0.6736 0.8207
No log 6.7692 176 0.6942 -0.0406 0.6942 0.8332
No log 6.8462 178 0.7058 -0.0717 0.7058 0.8401
No log 6.9231 180 0.6532 0.0474 0.6532 0.8082
No log 7.0 182 0.6298 0.2070 0.6298 0.7936
No log 7.0769 184 0.6478 0.1387 0.6478 0.8049
No log 7.1538 186 0.6686 0.0495 0.6686 0.8177
No log 7.2308 188 0.6735 0.2150 0.6735 0.8207
No log 7.3077 190 0.6931 0.1298 0.6931 0.8325
No log 7.3846 192 0.7177 0.2591 0.7177 0.8472
No log 7.4615 194 0.8568 0.0470 0.8568 0.9256
No log 7.5385 196 0.8504 0.0815 0.8504 0.9222
No log 7.6154 198 0.7128 0.2315 0.7128 0.8443
No log 7.6923 200 0.6946 0.1553 0.6946 0.8334
No log 7.7692 202 0.6970 0.1878 0.6970 0.8349
No log 7.8462 204 0.8280 0.0008 0.8280 0.9099
No log 7.9231 206 0.8722 0.0092 0.8722 0.9339
No log 8.0 208 0.7644 0.0271 0.7644 0.8743
No log 8.0769 210 0.6963 0.1444 0.6963 0.8345
No log 8.1538 212 0.6932 0.1627 0.6932 0.8326
No log 8.2308 214 0.6912 0.1878 0.6912 0.8314
No log 8.3077 216 0.7740 -0.0887 0.7740 0.8798
No log 8.3846 218 0.7450 -0.0259 0.7450 0.8632
No log 8.4615 220 0.6956 0.2150 0.6956 0.8340
No log 8.5385 222 0.7063 0.2105 0.7063 0.8404
No log 8.6154 224 0.7097 0.0798 0.7097 0.8424
No log 8.6923 226 0.7127 0.2922 0.7127 0.8442
No log 8.7692 228 0.7510 0.2679 0.7510 0.8666
No log 8.8462 230 0.7211 0.2318 0.7211 0.8492
No log 8.9231 232 0.7085 0.1659 0.7085 0.8417
No log 9.0 234 0.7145 0.0460 0.7145 0.8453
No log 9.0769 236 0.7737 0.0710 0.7737 0.8796
No log 9.1538 238 0.7353 0.0595 0.7353 0.8575
No log 9.2308 240 0.7439 0.0187 0.7439 0.8625
No log 9.3077 242 0.7280 0.0157 0.7280 0.8532
No log 9.3846 244 0.7838 -0.0121 0.7838 0.8853
No log 9.4615 246 0.7880 -0.0591 0.7880 0.8877
No log 9.5385 248 0.7198 0.0513 0.7198 0.8484
No log 9.6154 250 0.7360 -0.0350 0.7360 0.8579
No log 9.6923 252 0.7455 -0.0762 0.7455 0.8634
No log 9.7692 254 0.7137 0.0776 0.7137 0.8448
No log 9.8462 256 0.7105 0.0412 0.7105 0.8429
No log 9.9231 258 0.7187 0.0501 0.7187 0.8478
No log 10.0 260 0.7235 0.1647 0.7235 0.8506
No log 10.0769 262 0.7381 0.1686 0.7381 0.8591
No log 10.1538 264 0.7358 0.2430 0.7358 0.8578
No log 10.2308 266 0.7275 0.1700 0.7275 0.8529
No log 10.3077 268 0.7497 0.1724 0.7497 0.8659
No log 10.3846 270 0.7147 0.1315 0.7147 0.8454
No log 10.4615 272 0.7294 0.0947 0.7294 0.8541
No log 10.5385 274 0.7267 0.0509 0.7267 0.8525
No log 10.6154 276 0.7113 0.0449 0.7113 0.8434
No log 10.6923 278 0.7527 0.0644 0.7527 0.8676
No log 10.7692 280 0.6943 0.1024 0.6943 0.8332
No log 10.8462 282 0.7754 0.0786 0.7754 0.8805
No log 10.9231 284 0.7188 0.0909 0.7188 0.8478
No log 11.0 286 0.6699 0.1512 0.6699 0.8185
No log 11.0769 288 0.7430 0.1494 0.7430 0.8619
No log 11.1538 290 0.7792 -0.0016 0.7792 0.8827
No log 11.2308 292 0.6962 0.1080 0.6962 0.8344
No log 11.3077 294 0.6682 0.1979 0.6682 0.8175
No log 11.3846 296 0.6727 0.2339 0.6727 0.8202
No log 11.4615 298 0.6983 0.0571 0.6983 0.8356
No log 11.5385 300 0.7257 0.0089 0.7257 0.8519
No log 11.6154 302 0.7579 -0.0750 0.7579 0.8706
No log 11.6923 304 0.7452 -0.0750 0.7452 0.8633
No log 11.7692 306 0.7083 0.0874 0.7083 0.8416
No log 11.8462 308 0.7349 -0.0329 0.7349 0.8573
No log 11.9231 310 0.7505 -0.0293 0.7505 0.8663
No log 12.0 312 0.7372 0.0081 0.7372 0.8586
No log 12.0769 314 0.7757 -0.0648 0.7757 0.8807
No log 12.1538 316 0.8181 -0.0492 0.8181 0.9045
No log 12.2308 318 0.7583 -0.0345 0.7583 0.8708
No log 12.3077 320 0.7241 0.0481 0.7241 0.8509
No log 12.3846 322 0.7201 0.0503 0.7201 0.8486
No log 12.4615 324 0.7646 -0.0259 0.7646 0.8744
No log 12.5385 326 0.7355 -0.0350 0.7355 0.8576
No log 12.6154 328 0.7147 0.1787 0.7147 0.8454
No log 12.6923 330 0.7245 0.1144 0.7245 0.8512
No log 12.7692 332 0.7494 0.1263 0.7494 0.8657
No log 12.8462 334 0.8892 0.0753 0.8892 0.9430
No log 12.9231 336 0.8809 0.0747 0.8809 0.9386
No log 13.0 338 0.7734 0.0993 0.7734 0.8794
No log 13.0769 340 0.7016 0.0783 0.7016 0.8376
No log 13.1538 342 0.7414 0.0588 0.7414 0.8610
No log 13.2308 344 0.7158 0.2150 0.7158 0.8461
No log 13.3077 346 0.7276 0.0410 0.7276 0.8530
No log 13.3846 348 0.8382 0.0050 0.8382 0.9155
No log 13.4615 350 0.8127 -0.0334 0.8127 0.9015
No log 13.5385 352 0.7086 0.1304 0.7086 0.8418
No log 13.6154 354 0.7200 0.2105 0.7200 0.8485
No log 13.6923 356 0.7133 0.2105 0.7133 0.8446
No log 13.7692 358 0.7115 0.2225 0.7115 0.8435
No log 13.8462 360 0.7226 0.0106 0.7226 0.8501
No log 13.9231 362 0.7123 0.0532 0.7123 0.8440
No log 14.0 364 0.6888 0.1304 0.6888 0.8299
No log 14.0769 366 0.6965 0.1965 0.6965 0.8346
No log 14.1538 368 0.7196 0.2208 0.7196 0.8483
No log 14.2308 370 0.6887 0.2078 0.6887 0.8299
No log 14.3077 372 0.7369 -0.0293 0.7369 0.8584
No log 14.3846 374 0.7245 0.0110 0.7245 0.8512
No log 14.4615 376 0.6922 0.1192 0.6922 0.8320
No log 14.5385 378 0.7014 0.1921 0.7014 0.8375
No log 14.6154 380 0.7003 0.1986 0.7003 0.8368
No log 14.6923 382 0.6994 0.1298 0.6994 0.8363
No log 14.7692 384 0.6926 0.0874 0.6926 0.8323
No log 14.8462 386 0.6731 0.2225 0.6731 0.8204
No log 14.9231 388 0.6799 0.0821 0.6799 0.8246
No log 15.0 390 0.7051 -0.0469 0.7051 0.8397
No log 15.0769 392 0.6995 0.1304 0.6995 0.8364
No log 15.1538 394 0.6981 0.1304 0.6981 0.8355
No log 15.2308 396 0.7268 0.1354 0.7268 0.8525
No log 15.3077 398 0.7383 0.0926 0.7383 0.8593
No log 15.3846 400 0.7518 0.0926 0.7518 0.8670
No log 15.4615 402 0.7460 0.0465 0.7460 0.8637
No log 15.5385 404 0.7391 0.0412 0.7391 0.8597
No log 15.6154 406 0.7467 0.0465 0.7467 0.8641
No log 15.6923 408 0.7420 0.0412 0.7420 0.8614
No log 15.7692 410 0.7492 0.0376 0.7492 0.8655
No log 15.8462 412 0.7587 0.0376 0.7587 0.8710
No log 15.9231 414 0.7602 0.0749 0.7602 0.8719
No log 16.0 416 0.7594 0.0376 0.7594 0.8714
No log 16.0769 418 0.7522 0.0791 0.7522 0.8673
No log 16.1538 420 0.7604 0.0026 0.7604 0.8720
No log 16.2308 422 0.8035 -0.1204 0.8035 0.8964
No log 16.3077 424 0.7525 0.0081 0.7525 0.8674
No log 16.3846 426 0.7201 0.2195 0.7201 0.8486
No log 16.4615 428 0.7785 0.2248 0.7785 0.8823
No log 16.5385 430 0.7529 0.2166 0.7529 0.8677
No log 16.6154 432 0.8081 -0.2338 0.8081 0.8990
No log 16.6923 434 0.8437 -0.2048 0.8437 0.9185
No log 16.7692 436 0.7816 -0.1606 0.7816 0.8841
No log 16.8462 438 0.7367 0.1927 0.7367 0.8583
No log 16.9231 440 0.7564 0.2009 0.7564 0.8697
No log 17.0 442 0.7779 0.1395 0.7779 0.8820
No log 17.0769 444 0.7794 0.2277 0.7794 0.8828
No log 17.1538 446 0.7389 0.1787 0.7389 0.8596
No log 17.2308 448 0.7948 -0.2071 0.7948 0.8915
No log 17.3077 450 0.9400 0.0164 0.9400 0.9695
No log 17.3846 452 0.9005 -0.1054 0.9005 0.9489
No log 17.4615 454 0.7622 0.0030 0.7622 0.8730
No log 17.5385 456 0.7551 0.2424 0.7551 0.8690
No log 17.6154 458 0.7439 0.2424 0.7439 0.8625
No log 17.6923 460 0.7297 0.1856 0.7297 0.8542
No log 17.7692 462 0.7514 -0.0831 0.7514 0.8668
No log 17.8462 464 0.7524 -0.0426 0.7524 0.8674
No log 17.9231 466 0.7809 -0.0357 0.7809 0.8837
No log 18.0 468 0.8093 0.0886 0.8093 0.8996
No log 18.0769 470 0.7976 0.0134 0.7976 0.8931
No log 18.1538 472 0.8231 0.0627 0.8231 0.9073
No log 18.2308 474 0.8965 0.0476 0.8965 0.9468
No log 18.3077 476 0.8791 0.0454 0.8791 0.9376
No log 18.3846 478 0.7841 -0.0585 0.7841 0.8855
No log 18.4615 480 0.7261 0.1249 0.7261 0.8521
No log 18.5385 482 0.7338 0.1096 0.7338 0.8566
No log 18.6154 484 0.7220 0.1612 0.7220 0.8497
No log 18.6923 486 0.7083 0.1254 0.7083 0.8416
No log 18.7692 488 0.7280 -0.0921 0.7280 0.8533
No log 18.8462 490 0.7507 -0.1601 0.7507 0.8664
No log 18.9231 492 0.7064 0.0970 0.7064 0.8405
No log 19.0 494 0.6799 0.1199 0.6799 0.8245
No log 19.0769 496 0.6805 0.1565 0.6805 0.8249
No log 19.1538 498 0.6732 0.1259 0.6732 0.8205
0.3172 19.2308 500 0.6734 0.1148 0.6734 0.8206
0.3172 19.3077 502 0.6720 0.1318 0.6720 0.8198
0.3172 19.3846 504 0.6797 0.1379 0.6797 0.8244
0.3172 19.4615 506 0.6819 0.1828 0.6819 0.8258
0.3172 19.5385 508 0.6834 0.2195 0.6834 0.8267
0.3172 19.6154 510 0.6887 0.1705 0.6887 0.8299
0.3172 19.6923 512 0.6979 0.1240 0.6979 0.8354
0.3172 19.7692 514 0.7128 0.0359 0.7128 0.8443
0.3172 19.8462 516 0.7508 0.0883 0.7508 0.8665
0.3172 19.9231 518 0.8153 0.1304 0.8153 0.9030
0.3172 20.0 520 0.8446 0.1717 0.8446 0.9190
0.3172 20.0769 522 0.7803 0.0897 0.7803 0.8834
0.3172 20.1538 524 0.7379 -0.0051 0.7379 0.8590
0.3172 20.2308 526 0.7165 0.2150 0.7165 0.8464
0.3172 20.3077 528 0.7181 0.1627 0.7181 0.8474
0.3172 20.3846 530 0.7168 0.1627 0.7168 0.8466
0.3172 20.4615 532 0.7010 0.2150 0.7010 0.8372
0.3172 20.5385 534 0.7514 0.0051 0.7514 0.8668
0.3172 20.6154 536 0.8141 0.0307 0.8141 0.9023
0.3172 20.6923 538 0.7931 0.0600 0.7931 0.8906
0.3172 20.7692 540 0.7355 0.0408 0.7355 0.8576
0.3172 20.8462 542 0.7072 0.1196 0.7072 0.8410
0.3172 20.9231 544 0.7313 0.2009 0.7313 0.8551
0.3172 21.0 546 0.7424 0.1817 0.7424 0.8616
0.3172 21.0769 548 0.7295 0.1495 0.7295 0.8541
0.3172 21.1538 550 0.7393 0.0407 0.7393 0.8598
0.3172 21.2308 552 0.8230 0.0897 0.8230 0.9072
0.3172 21.3077 554 0.9404 0.0764 0.9404 0.9697
0.3172 21.3846 556 0.9972 -0.0099 0.9972 0.9986
0.3172 21.4615 558 0.9683 -0.0855 0.9683 0.9840

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
8
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k9_task3_organization

Finetuned
(4222)
this model