ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k1_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4079
  • Qwk: 0.6747
  • Mse: 0.4079
  • Rmse: 0.6387

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.4 2 2.4831 -0.0449 2.4831 1.5758
No log 0.8 4 1.2421 0.0997 1.2421 1.1145
No log 1.2 6 0.7761 0.1372 0.7761 0.8810
No log 1.6 8 0.9516 0.0327 0.9516 0.9755
No log 2.0 10 1.0156 0.1709 1.0156 1.0078
No log 2.4 12 1.0486 0.2903 1.0486 1.0240
No log 2.8 14 0.8951 0.2094 0.8951 0.9461
No log 3.2 16 0.7865 0.0608 0.7865 0.8868
No log 3.6 18 0.7732 0.1232 0.7732 0.8793
No log 4.0 20 0.7328 0.1646 0.7328 0.8560
No log 4.4 22 0.6772 0.1498 0.6772 0.8229
No log 4.8 24 0.6729 0.4243 0.6729 0.8203
No log 5.2 26 0.6281 0.4414 0.6281 0.7925
No log 5.6 28 0.5447 0.4314 0.5447 0.7380
No log 6.0 30 0.5724 0.5301 0.5724 0.7566
No log 6.4 32 0.7240 0.3783 0.7240 0.8509
No log 6.8 34 0.6116 0.4652 0.6116 0.7820
No log 7.2 36 0.4918 0.6039 0.4918 0.7013
No log 7.6 38 0.4871 0.6087 0.4871 0.6979
No log 8.0 40 0.5781 0.4892 0.5781 0.7603
No log 8.4 42 0.6469 0.4377 0.6469 0.8043
No log 8.8 44 0.5355 0.5538 0.5355 0.7318
No log 9.2 46 0.5506 0.6519 0.5506 0.7420
No log 9.6 48 0.5990 0.6439 0.5990 0.7739
No log 10.0 50 0.4921 0.6730 0.4921 0.7015
No log 10.4 52 0.6389 0.4881 0.6389 0.7993
No log 10.8 54 0.5715 0.5124 0.5715 0.7560
No log 11.2 56 0.4566 0.6676 0.4566 0.6757
No log 11.6 58 0.4451 0.6932 0.4451 0.6671
No log 12.0 60 0.4772 0.6389 0.4772 0.6908
No log 12.4 62 0.4194 0.7119 0.4194 0.6476
No log 12.8 64 0.4590 0.6080 0.4590 0.6775
No log 13.2 66 0.4159 0.6334 0.4159 0.6449
No log 13.6 68 0.4343 0.6716 0.4343 0.6590
No log 14.0 70 0.7542 0.3961 0.7542 0.8684
No log 14.4 72 0.8285 0.3961 0.8285 0.9102
No log 14.8 74 0.5589 0.5489 0.5589 0.7476
No log 15.2 76 0.4470 0.6115 0.4470 0.6686
No log 15.6 78 0.6694 0.6021 0.6694 0.8181
No log 16.0 80 0.5837 0.5463 0.5837 0.7640
No log 16.4 82 0.4361 0.7004 0.4361 0.6604
No log 16.8 84 0.5784 0.5887 0.5784 0.7605
No log 17.2 86 0.5749 0.5887 0.5749 0.7582
No log 17.6 88 0.4430 0.6855 0.4430 0.6656
No log 18.0 90 0.5173 0.5695 0.5173 0.7192
No log 18.4 92 0.6279 0.5677 0.6279 0.7924
No log 18.8 94 0.5099 0.5748 0.5099 0.7141
No log 19.2 96 0.4249 0.6575 0.4249 0.6519
No log 19.6 98 0.4997 0.5429 0.4997 0.7069
No log 20.0 100 0.4833 0.5779 0.4833 0.6952
No log 20.4 102 0.4276 0.6467 0.4276 0.6539
No log 20.8 104 0.4874 0.6267 0.4874 0.6981
No log 21.2 106 0.4674 0.6356 0.4674 0.6837
No log 21.6 108 0.4254 0.6577 0.4254 0.6522
No log 22.0 110 0.4193 0.6870 0.4193 0.6475
No log 22.4 112 0.4127 0.6870 0.4127 0.6424
No log 22.8 114 0.4061 0.6957 0.4061 0.6372
No log 23.2 116 0.4116 0.6705 0.4116 0.6416
No log 23.6 118 0.4243 0.6793 0.4243 0.6514
No log 24.0 120 0.4216 0.6884 0.4216 0.6493
No log 24.4 122 0.4122 0.6884 0.4122 0.6421
No log 24.8 124 0.4169 0.6701 0.4169 0.6457
No log 25.2 126 0.3839 0.7032 0.3839 0.6196
No log 25.6 128 0.3949 0.6572 0.3949 0.6284
No log 26.0 130 0.4068 0.6900 0.4068 0.6378
No log 26.4 132 0.4124 0.6805 0.4124 0.6422
No log 26.8 134 0.4075 0.6496 0.4075 0.6383
No log 27.2 136 0.4115 0.6852 0.4115 0.6415
No log 27.6 138 0.4076 0.6161 0.4076 0.6384
No log 28.0 140 0.4018 0.6092 0.4018 0.6339
No log 28.4 142 0.4190 0.6601 0.4190 0.6473
No log 28.8 144 0.4581 0.6537 0.4581 0.6769
No log 29.2 146 0.4479 0.6342 0.4479 0.6693
No log 29.6 148 0.4045 0.6289 0.4045 0.6360
No log 30.0 150 0.4168 0.6472 0.4168 0.6456
No log 30.4 152 0.4197 0.6662 0.4197 0.6479
No log 30.8 154 0.4241 0.6337 0.4241 0.6512
No log 31.2 156 0.4369 0.6526 0.4369 0.6610
No log 31.6 158 0.4398 0.6526 0.4398 0.6631
No log 32.0 160 0.4225 0.6377 0.4225 0.6500
No log 32.4 162 0.4417 0.6228 0.4417 0.6646
No log 32.8 164 0.4787 0.5668 0.4787 0.6919
No log 33.2 166 0.4782 0.5701 0.4782 0.6915
No log 33.6 168 0.4402 0.6739 0.4402 0.6635
No log 34.0 170 0.4453 0.6491 0.4453 0.6673
No log 34.4 172 0.4401 0.6491 0.4401 0.6634
No log 34.8 174 0.4374 0.6667 0.4374 0.6614
No log 35.2 176 0.4714 0.6693 0.4714 0.6866
No log 35.6 178 0.4698 0.6944 0.4698 0.6854
No log 36.0 180 0.4550 0.6656 0.4550 0.6745
No log 36.4 182 0.4608 0.6572 0.4608 0.6788
No log 36.8 184 0.4634 0.6100 0.4634 0.6807
No log 37.2 186 0.4762 0.6259 0.4762 0.6901
No log 37.6 188 0.4609 0.6066 0.4609 0.6789
No log 38.0 190 0.4454 0.6739 0.4454 0.6674
No log 38.4 192 0.4601 0.5845 0.4601 0.6783
No log 38.8 194 0.4530 0.5816 0.4530 0.6730
No log 39.2 196 0.4426 0.6747 0.4426 0.6653
No log 39.6 198 0.4426 0.6747 0.4426 0.6653
No log 40.0 200 0.4423 0.5816 0.4423 0.6651
No log 40.4 202 0.4385 0.6655 0.4385 0.6622
No log 40.8 204 0.4623 0.5212 0.4623 0.6800
No log 41.2 206 0.4670 0.5283 0.4670 0.6834
No log 41.6 208 0.4370 0.5265 0.4370 0.6610
No log 42.0 210 0.4334 0.6505 0.4334 0.6583
No log 42.4 212 0.4560 0.6505 0.4560 0.6753
No log 42.8 214 0.4447 0.6505 0.4447 0.6668
No log 43.2 216 0.4195 0.6843 0.4195 0.6477
No log 43.6 218 0.4407 0.6223 0.4407 0.6638
No log 44.0 220 0.4784 0.5794 0.4784 0.6917
No log 44.4 222 0.4857 0.5794 0.4857 0.6969
No log 44.8 224 0.4548 0.5512 0.4548 0.6744
No log 45.2 226 0.4261 0.6464 0.4261 0.6527
No log 45.6 228 0.4481 0.6228 0.4481 0.6694
No log 46.0 230 0.4889 0.6687 0.4889 0.6992
No log 46.4 232 0.4764 0.6406 0.4764 0.6902
No log 46.8 234 0.4430 0.6408 0.4430 0.6656
No log 47.2 236 0.4287 0.6125 0.4287 0.6548
No log 47.6 238 0.4287 0.6330 0.4287 0.6547
No log 48.0 240 0.4345 0.6317 0.4345 0.6592
No log 48.4 242 0.4475 0.6518 0.4475 0.6690
No log 48.8 244 0.4303 0.6339 0.4303 0.6559
No log 49.2 246 0.4267 0.6351 0.4267 0.6532
No log 49.6 248 0.4334 0.5890 0.4334 0.6583
No log 50.0 250 0.4368 0.5319 0.4368 0.6609
No log 50.4 252 0.4275 0.6053 0.4275 0.6539
No log 50.8 254 0.4226 0.6344 0.4226 0.6501
No log 51.2 256 0.4291 0.6129 0.4291 0.6550
No log 51.6 258 0.4271 0.6129 0.4271 0.6535
No log 52.0 260 0.4295 0.6129 0.4295 0.6553
No log 52.4 262 0.4310 0.6129 0.4310 0.6565
No log 52.8 264 0.4313 0.6554 0.4313 0.6567
No log 53.2 266 0.4327 0.6383 0.4327 0.6578
No log 53.6 268 0.4340 0.6301 0.4340 0.6588
No log 54.0 270 0.4374 0.6301 0.4374 0.6614
No log 54.4 272 0.4406 0.6301 0.4406 0.6638
No log 54.8 274 0.4422 0.6301 0.4422 0.6650
No log 55.2 276 0.4374 0.6383 0.4374 0.6613
No log 55.6 278 0.4405 0.6241 0.4405 0.6637
No log 56.0 280 0.4351 0.6024 0.4351 0.6596
No log 56.4 282 0.4339 0.6255 0.4339 0.6587
No log 56.8 284 0.4345 0.6024 0.4345 0.6591
No log 57.2 286 0.4449 0.6228 0.4449 0.6670
No log 57.6 288 0.4422 0.6326 0.4422 0.6650
No log 58.0 290 0.4353 0.6317 0.4353 0.6597
No log 58.4 292 0.4295 0.6317 0.4295 0.6554
No log 58.8 294 0.4251 0.5899 0.4251 0.6520
No log 59.2 296 0.4248 0.6053 0.4248 0.6517
No log 59.6 298 0.4320 0.5248 0.4320 0.6572
No log 60.0 300 0.4257 0.6269 0.4257 0.6525
No log 60.4 302 0.4199 0.6269 0.4199 0.6480
No log 60.8 304 0.4135 0.6269 0.4135 0.6430
No log 61.2 306 0.4090 0.6269 0.4090 0.6395
No log 61.6 308 0.4069 0.6269 0.4069 0.6379
No log 62.0 310 0.4052 0.6269 0.4052 0.6366
No log 62.4 312 0.4000 0.6555 0.4000 0.6325
No log 62.8 314 0.4036 0.6935 0.4036 0.6353
No log 63.2 316 0.4026 0.6747 0.4026 0.6345
No log 63.6 318 0.4008 0.6555 0.4008 0.6331
No log 64.0 320 0.3995 0.6555 0.3995 0.6320
No log 64.4 322 0.4044 0.6269 0.4044 0.6359
No log 64.8 324 0.3995 0.6464 0.3995 0.6321
No log 65.2 326 0.3991 0.6919 0.3991 0.6317
No log 65.6 328 0.4092 0.7053 0.4092 0.6397
No log 66.0 330 0.4094 0.6968 0.4094 0.6399
No log 66.4 332 0.4025 0.6828 0.4025 0.6344
No log 66.8 334 0.3987 0.6650 0.3987 0.6314
No log 67.2 336 0.3971 0.6464 0.3971 0.6301
No log 67.6 338 0.4003 0.6464 0.4003 0.6327
No log 68.0 340 0.3993 0.6556 0.3993 0.6319
No log 68.4 342 0.3996 0.6555 0.3996 0.6321
No log 68.8 344 0.4027 0.6439 0.4027 0.6346
No log 69.2 346 0.4082 0.6426 0.4082 0.6389
No log 69.6 348 0.4068 0.6627 0.4068 0.6378
No log 70.0 350 0.3972 0.6747 0.3972 0.6302
No log 70.4 352 0.4005 0.6464 0.4005 0.6328
No log 70.8 354 0.4225 0.6034 0.4225 0.6500
No log 71.2 356 0.4465 0.5706 0.4465 0.6682
No log 71.6 358 0.4450 0.5706 0.4450 0.6671
No log 72.0 360 0.4248 0.5554 0.4248 0.6517
No log 72.4 362 0.4073 0.6467 0.4073 0.6382
No log 72.8 364 0.4083 0.6739 0.4083 0.6390
No log 73.2 366 0.4163 0.6431 0.4163 0.6452
No log 73.6 368 0.4200 0.6431 0.4200 0.6481
No log 74.0 370 0.4237 0.6330 0.4237 0.6509
No log 74.4 372 0.4237 0.6125 0.4237 0.6509
No log 74.8 374 0.4212 0.6555 0.4212 0.6490
No log 75.2 376 0.4222 0.6555 0.4222 0.6498
No log 75.6 378 0.4225 0.6555 0.4225 0.6500
No log 76.0 380 0.4225 0.6555 0.4225 0.6500
No log 76.4 382 0.4236 0.6555 0.4236 0.6509
No log 76.8 384 0.4293 0.6053 0.4293 0.6552
No log 77.2 386 0.4354 0.5846 0.4354 0.6599
No log 77.6 388 0.4349 0.5846 0.4349 0.6595
No log 78.0 390 0.4314 0.6554 0.4314 0.6568
No log 78.4 392 0.4346 0.6739 0.4346 0.6592
No log 78.8 394 0.4403 0.6634 0.4403 0.6635
No log 79.2 396 0.4431 0.6702 0.4431 0.6656
No log 79.6 398 0.4395 0.6880 0.4395 0.6630
No log 80.0 400 0.4288 0.6807 0.4288 0.6549
No log 80.4 402 0.4176 0.6639 0.4176 0.6462
No log 80.8 404 0.4103 0.6747 0.4103 0.6405
No log 81.2 406 0.4072 0.6747 0.4072 0.6381
No log 81.6 408 0.4076 0.6555 0.4076 0.6384
No log 82.0 410 0.4116 0.6377 0.4116 0.6415
No log 82.4 412 0.4122 0.6577 0.4122 0.6420
No log 82.8 414 0.4095 0.6555 0.4095 0.6399
No log 83.2 416 0.4078 0.6739 0.4078 0.6386
No log 83.6 418 0.4126 0.6730 0.4126 0.6423
No log 84.0 420 0.4177 0.6634 0.4177 0.6463
No log 84.4 422 0.4181 0.6634 0.4181 0.6466
No log 84.8 424 0.4153 0.6634 0.4153 0.6445
No log 85.2 426 0.4138 0.6730 0.4138 0.6433
No log 85.6 428 0.4119 0.6730 0.4119 0.6418
No log 86.0 430 0.4096 0.6739 0.4096 0.6400
No log 86.4 432 0.4085 0.6554 0.4085 0.6392
No log 86.8 434 0.4090 0.6554 0.4090 0.6396
No log 87.2 436 0.4085 0.6555 0.4085 0.6391
No log 87.6 438 0.4072 0.6555 0.4072 0.6381
No log 88.0 440 0.4064 0.6555 0.4064 0.6375
No log 88.4 442 0.4061 0.6739 0.4061 0.6372
No log 88.8 444 0.4060 0.6739 0.4060 0.6372
No log 89.2 446 0.4049 0.6739 0.4049 0.6363
No log 89.6 448 0.4038 0.6747 0.4038 0.6354
No log 90.0 450 0.4033 0.6747 0.4033 0.6351
No log 90.4 452 0.4041 0.6747 0.4041 0.6357
No log 90.8 454 0.4045 0.6747 0.4045 0.6360
No log 91.2 456 0.4046 0.6747 0.4046 0.6360
No log 91.6 458 0.4053 0.6747 0.4053 0.6366
No log 92.0 460 0.4065 0.6643 0.4065 0.6376
No log 92.4 462 0.4066 0.6747 0.4066 0.6377
No log 92.8 464 0.4058 0.6747 0.4058 0.6370
No log 93.2 466 0.4054 0.6747 0.4054 0.6367
No log 93.6 468 0.4053 0.6747 0.4053 0.6366
No log 94.0 470 0.4054 0.6747 0.4054 0.6367
No log 94.4 472 0.4050 0.6747 0.4050 0.6364
No log 94.8 474 0.4055 0.6747 0.4055 0.6368
No log 95.2 476 0.4059 0.6747 0.4059 0.6371
No log 95.6 478 0.4061 0.6747 0.4061 0.6373
No log 96.0 480 0.4065 0.6747 0.4065 0.6375
No log 96.4 482 0.4069 0.6747 0.4069 0.6379
No log 96.8 484 0.4071 0.6747 0.4071 0.6380
No log 97.2 486 0.4074 0.6747 0.4074 0.6383
No log 97.6 488 0.4076 0.6747 0.4076 0.6384
No log 98.0 490 0.4077 0.6747 0.4077 0.6385
No log 98.4 492 0.4078 0.6747 0.4078 0.6386
No log 98.8 494 0.4078 0.6747 0.4078 0.6386
No log 99.2 496 0.4079 0.6747 0.4079 0.6387
No log 99.6 498 0.4079 0.6747 0.4079 0.6387
0.1889 100.0 500 0.4079 0.6747 0.4079 0.6387

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
7
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k1_task7_organization

Finetuned
(4204)
this model