ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k16_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.2248
  • Qwk: 0.2353
  • Mse: 1.2248
  • Rmse: 1.1067

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0345 2 4.8686 -0.0020 4.8686 2.2065
No log 0.0690 4 3.0498 -0.0321 3.0498 1.7464
No log 0.1034 6 2.6477 -0.0732 2.6477 1.6272
No log 0.1379 8 2.4772 -0.0732 2.4772 1.5739
No log 0.1724 10 2.2557 -0.0056 2.2557 1.5019
No log 0.2069 12 1.6088 0.0372 1.6088 1.2684
No log 0.2414 14 1.5309 0.0227 1.5309 1.2373
No log 0.2759 16 1.7065 0.0372 1.7065 1.3063
No log 0.3103 18 1.9574 0.0369 1.9574 1.3991
No log 0.3448 20 1.9432 0.0486 1.9432 1.3940
No log 0.3793 22 1.7236 0.0959 1.7236 1.3128
No log 0.4138 24 1.5300 0.0792 1.5300 1.2369
No log 0.4483 26 1.4329 0.0792 1.4329 1.1970
No log 0.4828 28 1.5123 -0.0064 1.5123 1.2297
No log 0.5172 30 1.7816 0.0971 1.7816 1.3348
No log 0.5517 32 1.8312 0.1366 1.8312 1.3532
No log 0.5862 34 1.6016 0.0827 1.6016 1.2655
No log 0.6207 36 1.4659 0.0572 1.4659 1.2108
No log 0.6552 38 1.3233 0.0999 1.3233 1.1503
No log 0.6897 40 1.2169 0.2254 1.2169 1.1031
No log 0.7241 42 1.2640 0.0750 1.2640 1.1243
No log 0.7586 44 1.3733 0.0310 1.3733 1.1719
No log 0.7931 46 1.5647 0.0372 1.5647 1.2509
No log 0.8276 48 1.7370 0.0971 1.7370 1.3180
No log 0.8621 50 1.8570 0.1266 1.8570 1.3627
No log 0.8966 52 1.9064 0.1485 1.9064 1.3807
No log 0.9310 54 1.5401 0.2801 1.5401 1.2410
No log 0.9655 56 1.2549 0.2439 1.2549 1.1202
No log 1.0 58 1.1128 0.3848 1.1128 1.0549
No log 1.0345 60 1.1208 0.3704 1.1208 1.0587
No log 1.0690 62 1.1554 0.3348 1.1554 1.0749
No log 1.1034 64 1.2821 0.2112 1.2821 1.1323
No log 1.1379 66 1.4029 0.1943 1.4029 1.1845
No log 1.1724 68 1.4858 0.1032 1.4858 1.2189
No log 1.2069 70 1.4951 0.0733 1.4951 1.2228
No log 1.2414 72 1.5251 0.0756 1.5251 1.2350
No log 1.2759 74 1.6724 0.1625 1.6724 1.2932
No log 1.3103 76 1.7450 0.1269 1.7450 1.3210
No log 1.3448 78 1.8644 0.1290 1.8644 1.3654
No log 1.3793 80 2.1672 0.1794 2.1672 1.4721
No log 1.4138 82 2.9343 0.0493 2.9343 1.7130
No log 1.4483 84 3.4562 0.0605 3.4562 1.8591
No log 1.4828 86 2.7949 0.0516 2.7949 1.6718
No log 1.5172 88 1.6745 0.1629 1.6745 1.2940
No log 1.5517 90 1.3549 0.1848 1.3549 1.1640
No log 1.5862 92 1.2246 0.2532 1.2246 1.1066
No log 1.6207 94 1.1946 0.3078 1.1946 1.0930
No log 1.6552 96 1.2547 0.3218 1.2547 1.1201
No log 1.6897 98 1.3591 0.3005 1.3591 1.1658
No log 1.7241 100 1.3445 0.3449 1.3445 1.1595
No log 1.7586 102 1.4563 0.3184 1.4563 1.2068
No log 1.7931 104 1.6056 0.1498 1.6056 1.2671
No log 1.8276 106 1.5081 0.2338 1.5081 1.2280
No log 1.8621 108 1.1913 0.2919 1.1913 1.0915
No log 1.8966 110 1.0889 0.3283 1.0889 1.0435
No log 1.9310 112 1.3032 0.2863 1.3032 1.1416
No log 1.9655 114 1.6148 0.2252 1.6148 1.2708
No log 2.0 116 1.6644 0.1748 1.6644 1.2901
No log 2.0345 118 1.4274 0.2690 1.4274 1.1948
No log 2.0690 120 1.1659 0.2851 1.1659 1.0798
No log 2.1034 122 1.0705 0.2802 1.0705 1.0346
No log 2.1379 124 1.1111 0.3282 1.1111 1.0541
No log 2.1724 126 1.1450 0.2583 1.1450 1.0701
No log 2.2069 128 1.1598 0.2009 1.1598 1.0769
No log 2.2414 130 1.2489 0.2060 1.2489 1.1176
No log 2.2759 132 1.2722 0.1833 1.2722 1.1279
No log 2.3103 134 1.3659 0.2501 1.3659 1.1687
No log 2.3448 136 1.4982 0.2688 1.4982 1.2240
No log 2.3793 138 1.6330 0.2949 1.6330 1.2779
No log 2.4138 140 1.7332 0.2330 1.7332 1.3165
No log 2.4483 142 1.5181 0.3149 1.5181 1.2321
No log 2.4828 144 1.2711 0.2119 1.2711 1.1274
No log 2.5172 146 1.2315 0.2863 1.2315 1.1097
No log 2.5517 148 1.2763 0.2418 1.2763 1.1297
No log 2.5862 150 1.4232 0.3040 1.4232 1.1930
No log 2.6207 152 1.4091 0.2479 1.4091 1.1871
No log 2.6552 154 1.2457 0.3795 1.2457 1.1161
No log 2.6897 156 1.2459 0.3677 1.2459 1.1162
No log 2.7241 158 1.4097 0.2709 1.4097 1.1873
No log 2.7586 160 1.5372 0.3294 1.5372 1.2398
No log 2.7931 162 1.4083 0.2873 1.4083 1.1867
No log 2.8276 164 1.3105 0.2611 1.3105 1.1448
No log 2.8621 166 1.1509 0.2539 1.1509 1.0728
No log 2.8966 168 1.0755 0.2775 1.0755 1.0371
No log 2.9310 170 1.0950 0.2995 1.0950 1.0464
No log 2.9655 172 1.1305 0.3602 1.1305 1.0632
No log 3.0 174 1.1070 0.3190 1.1070 1.0521
No log 3.0345 176 1.1888 0.2495 1.1888 1.0903
No log 3.0690 178 1.2212 0.2871 1.2212 1.1051
No log 3.1034 180 1.2633 0.2524 1.2633 1.1240
No log 3.1379 182 1.1988 0.3407 1.1988 1.0949
No log 3.1724 184 1.1372 0.3254 1.1372 1.0664
No log 3.2069 186 1.1348 0.3517 1.1348 1.0653
No log 3.2414 188 1.0759 0.3886 1.0759 1.0372
No log 3.2759 190 1.0559 0.3811 1.0559 1.0275
No log 3.3103 192 1.0497 0.4373 1.0497 1.0246
No log 3.3448 194 1.0350 0.3042 1.0350 1.0173
No log 3.3793 196 1.0578 0.2697 1.0578 1.0285
No log 3.4138 198 1.1310 0.2963 1.1310 1.0635
No log 3.4483 200 1.0921 0.2843 1.0921 1.0451
No log 3.4828 202 1.0590 0.2898 1.0590 1.0291
No log 3.5172 204 1.0731 0.2474 1.0731 1.0359
No log 3.5517 206 1.0998 0.2679 1.0998 1.0487
No log 3.5862 208 1.1097 0.2474 1.1097 1.0534
No log 3.6207 210 1.1100 0.2474 1.1100 1.0535
No log 3.6552 212 1.1285 0.2823 1.1285 1.0623
No log 3.6897 214 1.1244 0.2241 1.1244 1.0604
No log 3.7241 216 1.1552 0.2218 1.1552 1.0748
No log 3.7586 218 1.2063 0.2606 1.2063 1.0983
No log 3.7931 220 1.1587 0.3042 1.1587 1.0765
No log 3.8276 222 1.2537 0.1142 1.2537 1.1197
No log 3.8621 224 1.4885 0.1779 1.4885 1.2200
No log 3.8966 226 1.4856 0.0253 1.4856 1.2189
No log 3.9310 228 1.3892 0.0839 1.3892 1.1787
No log 3.9655 230 1.1920 0.0931 1.1920 1.0918
No log 4.0 232 1.1594 0.2092 1.1594 1.0768
No log 4.0345 234 1.1387 0.2512 1.1387 1.0671
No log 4.0690 236 1.1386 0.1663 1.1386 1.0670
No log 4.1034 238 1.2587 0.1896 1.2587 1.1219
No log 4.1379 240 1.4155 0.2053 1.4155 1.1897
No log 4.1724 242 1.4327 0.2044 1.4327 1.1970
No log 4.2069 244 1.3242 0.1522 1.3242 1.1507
No log 4.2414 246 1.3034 0.0603 1.3034 1.1417
No log 4.2759 248 1.3354 0.1323 1.3354 1.1556
No log 4.3103 250 1.4208 0.2143 1.4208 1.1920
No log 4.3448 252 1.4794 0.2069 1.4794 1.2163
No log 4.3793 254 1.4600 0.2405 1.4600 1.2083
No log 4.4138 256 1.2661 0.3480 1.2661 1.1252
No log 4.4483 258 1.1204 0.2280 1.1204 1.0585
No log 4.4828 260 1.1325 0.2973 1.1325 1.0642
No log 4.5172 262 1.1170 0.2973 1.1170 1.0569
No log 4.5517 264 1.1069 0.1606 1.1069 1.0521
No log 4.5862 266 1.2398 0.2140 1.2398 1.1135
No log 4.6207 268 1.3618 0.2780 1.3618 1.1670
No log 4.6552 270 1.2649 0.2424 1.2649 1.1247
No log 4.6897 272 1.1138 0.1513 1.1138 1.0553
No log 4.7241 274 1.1358 0.3617 1.1358 1.0657
No log 4.7586 276 1.1233 0.3343 1.1233 1.0599
No log 4.7931 278 1.0793 0.1548 1.0793 1.0389
No log 4.8276 280 1.1218 0.2989 1.1218 1.0592
No log 4.8621 282 1.2309 0.3005 1.2309 1.1094
No log 4.8966 284 1.1893 0.2939 1.1893 1.0906
No log 4.9310 286 1.0817 0.3218 1.0817 1.0401
No log 4.9655 288 1.0450 0.2322 1.0450 1.0222
No log 5.0 290 1.0815 0.2352 1.0815 1.0399
No log 5.0345 292 1.0753 0.2029 1.0753 1.0369
No log 5.0690 294 1.0906 0.2572 1.0906 1.0443
No log 5.1034 296 1.1607 0.3373 1.1607 1.0773
No log 5.1379 298 1.1652 0.2984 1.1652 1.0795
No log 5.1724 300 1.1274 0.2221 1.1274 1.0618
No log 5.2069 302 1.1084 0.1517 1.1084 1.0528
No log 5.2414 304 1.1479 0.2221 1.1479 1.0714
No log 5.2759 306 1.1632 0.1296 1.1632 1.0785
No log 5.3103 308 1.1558 0.1049 1.1558 1.0751
No log 5.3448 310 1.1464 0.1548 1.1464 1.0707
No log 5.3793 312 1.1730 0.2352 1.1730 1.0830
No log 5.4138 314 1.1569 0.2453 1.1569 1.0756
No log 5.4483 316 1.1644 0.1612 1.1644 1.0791
No log 5.4828 318 1.3076 0.2780 1.3076 1.1435
No log 5.5172 320 1.3382 0.2601 1.3382 1.1568
No log 5.5517 322 1.2459 0.2005 1.2459 1.1162
No log 5.5862 324 1.1709 0.2265 1.1709 1.0821
No log 5.6207 326 1.1548 0.2156 1.1548 1.0746
No log 5.6552 328 1.1614 0.2156 1.1614 1.0777
No log 5.6897 330 1.2285 0.1982 1.2285 1.1084
No log 5.7241 332 1.3824 0.1284 1.3824 1.1758
No log 5.7586 334 1.4426 0.1942 1.4426 1.2011
No log 5.7931 336 1.3791 0.2105 1.3791 1.1743
No log 5.8276 338 1.2281 0.1935 1.2281 1.1082
No log 5.8621 340 1.1468 0.2353 1.1468 1.0709
No log 5.8966 342 1.1301 0.2358 1.1301 1.0631
No log 5.9310 344 1.1661 0.2824 1.1661 1.0798
No log 5.9655 346 1.2234 0.2624 1.2234 1.1061
No log 6.0 348 1.2023 0.2714 1.2023 1.0965
No log 6.0345 350 1.1420 0.3154 1.1420 1.0687
No log 6.0690 352 1.0904 0.2499 1.0904 1.0442
No log 6.1034 354 1.0904 0.2647 1.0904 1.0442
No log 6.1379 356 1.1481 0.2539 1.1481 1.0715
No log 6.1724 358 1.2272 0.1862 1.2272 1.1078
No log 6.2069 360 1.2222 0.1898 1.2222 1.1055
No log 6.2414 362 1.1255 0.2539 1.1255 1.0609
No log 6.2759 364 1.0722 0.2679 1.0722 1.0355
No log 6.3103 366 1.0749 0.2919 1.0749 1.0368
No log 6.3448 368 1.1192 0.2918 1.1192 1.0579
No log 6.3793 370 1.3162 0.2096 1.3162 1.1473
No log 6.4138 372 1.4350 0.2354 1.4350 1.1979
No log 6.4483 374 1.3248 0.1785 1.3248 1.1510
No log 6.4828 376 1.1742 0.1943 1.1742 1.0836
No log 6.5172 378 1.1413 0.2505 1.1413 1.0683
No log 6.5517 380 1.1453 0.1634 1.1453 1.0702
No log 6.5862 382 1.1668 0.1975 1.1668 1.0802
No log 6.6207 384 1.1549 0.1824 1.1549 1.0747
No log 6.6552 386 1.1372 0.2016 1.1372 1.0664
No log 6.6897 388 1.1119 0.1903 1.1119 1.0545
No log 6.7241 390 1.1210 0.1709 1.1210 1.0588
No log 6.7586 392 1.2060 0.1505 1.2060 1.0982
No log 6.7931 394 1.2557 0.2140 1.2557 1.1206
No log 6.8276 396 1.1977 0.1598 1.1977 1.0944
No log 6.8621 398 1.1630 0.2372 1.1630 1.0784
No log 6.8966 400 1.1228 0.2211 1.1228 1.0596
No log 6.9310 402 1.0965 0.2100 1.0965 1.0471
No log 6.9655 404 1.0959 0.1903 1.0959 1.0469
No log 7.0 406 1.1510 0.2128 1.1510 1.0728
No log 7.0345 408 1.1789 0.1950 1.1789 1.0858
No log 7.0690 410 1.1304 0.2544 1.1304 1.0632
No log 7.1034 412 1.0837 0.3015 1.0837 1.0410
No log 7.1379 414 1.0835 0.2728 1.0835 1.0409
No log 7.1724 416 1.1215 0.2697 1.1215 1.0590
No log 7.2069 418 1.2241 0.2832 1.2241 1.1064
No log 7.2414 420 1.2968 0.2604 1.2968 1.1388
No log 7.2759 422 1.2137 0.2524 1.2137 1.1017
No log 7.3103 424 1.1015 0.2823 1.1015 1.0495
No log 7.3448 426 1.0805 0.3042 1.0805 1.0395
No log 7.3793 428 1.0895 0.2658 1.0895 1.0438
No log 7.4138 430 1.0914 0.2492 1.0914 1.0447
No log 7.4483 432 1.1234 0.2007 1.1234 1.0599
No log 7.4828 434 1.1359 0.2438 1.1359 1.0658
No log 7.5172 436 1.1220 0.2577 1.1220 1.0593
No log 7.5517 438 1.0872 0.3458 1.0872 1.0427
No log 7.5862 440 1.0739 0.3430 1.0739 1.0363
No log 7.6207 442 1.0849 0.3618 1.0849 1.0416
No log 7.6552 444 1.1495 0.2742 1.1495 1.0721
No log 7.6897 446 1.3413 0.2520 1.3413 1.1581
No log 7.7241 448 1.5769 0.1840 1.5769 1.2557
No log 7.7586 450 1.5701 0.1869 1.5701 1.2530
No log 7.7931 452 1.3589 0.1929 1.3589 1.1657
No log 7.8276 454 1.1470 0.2584 1.1470 1.0710
No log 7.8621 456 1.1144 0.2380 1.1144 1.0557
No log 7.8966 458 1.1183 0.2676 1.1183 1.0575
No log 7.9310 460 1.1471 0.2728 1.1471 1.0710
No log 7.9655 462 1.2419 0.1865 1.2419 1.1144
No log 8.0 464 1.2035 0.2170 1.2035 1.0970
No log 8.0345 466 1.1010 0.2447 1.1010 1.0493
No log 8.0690 468 1.0770 0.3117 1.0770 1.0378
No log 8.1034 470 1.0742 0.2813 1.0742 1.0364
No log 8.1379 472 1.0817 0.2647 1.0817 1.0400
No log 8.1724 474 1.1088 0.2844 1.1088 1.0530
No log 8.2069 476 1.0917 0.2844 1.0917 1.0448
No log 8.2414 478 1.0460 0.3354 1.0460 1.0227
No log 8.2759 480 1.0444 0.3354 1.0444 1.0220
No log 8.3103 482 1.0639 0.2647 1.0639 1.0315
No log 8.3448 484 1.0736 0.3013 1.0736 1.0362
No log 8.3793 486 1.0540 0.2794 1.0540 1.0266
No log 8.4138 488 1.0380 0.3354 1.0380 1.0188
No log 8.4483 490 1.0334 0.3164 1.0334 1.0166
No log 8.4828 492 1.0433 0.2477 1.0433 1.0214
No log 8.5172 494 1.0435 0.2946 1.0435 1.0215
No log 8.5517 496 1.0612 0.2431 1.0612 1.0301
No log 8.5862 498 1.1024 0.2608 1.1024 1.0500
0.3401 8.6207 500 1.1078 0.2608 1.1078 1.0525
0.3401 8.6552 502 1.0800 0.2108 1.0800 1.0392
0.3401 8.6897 504 1.0609 0.2679 1.0609 1.0300
0.3401 8.7241 506 1.0601 0.2296 1.0601 1.0296
0.3401 8.7586 508 1.0848 0.3486 1.0848 1.0416
0.3401 8.7931 510 1.1342 0.2832 1.1342 1.0650
0.3401 8.8276 512 1.1645 0.2744 1.1645 1.0791
0.3401 8.8621 514 1.1761 0.2744 1.1761 1.0845
0.3401 8.8966 516 1.1758 0.2744 1.1758 1.0843
0.3401 8.9310 518 1.1432 0.2804 1.1432 1.0692
0.3401 8.9655 520 1.1527 0.2395 1.1527 1.0737
0.3401 9.0 522 1.2053 0.1768 1.2053 1.0979
0.3401 9.0345 524 1.2041 0.1980 1.2041 1.0973
0.3401 9.0690 526 1.1917 0.2210 1.1917 1.0917
0.3401 9.1034 528 1.2248 0.2353 1.2248 1.1067

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
135M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k16_task2_organization

Finetuned
(4005)
this model