ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k2_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4182
  • Qwk: 0.6001
  • Mse: 0.4182
  • Rmse: 0.6467

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.25 2 2.4023 0.0052 2.4023 1.5499
No log 0.5 4 1.1655 0.1259 1.1655 1.0796
No log 0.75 6 0.7541 0.1372 0.7541 0.8684
No log 1.0 8 0.7285 0.2319 0.7285 0.8535
No log 1.25 10 0.6985 0.2783 0.6985 0.8357
No log 1.5 12 0.8415 0.2375 0.8415 0.9173
No log 1.75 14 0.6713 0.3029 0.6713 0.8194
No log 2.0 16 0.6546 0.3690 0.6546 0.8091
No log 2.25 18 0.7346 0.4074 0.7346 0.8571
No log 2.5 20 0.6949 0.4172 0.6949 0.8336
No log 2.75 22 0.5933 0.4569 0.5933 0.7703
No log 3.0 24 0.5082 0.5046 0.5082 0.7129
No log 3.25 26 0.4662 0.4656 0.4662 0.6828
No log 3.5 28 0.4559 0.5488 0.4559 0.6752
No log 3.75 30 0.4866 0.6052 0.4866 0.6976
No log 4.0 32 0.5134 0.5950 0.5134 0.7165
No log 4.25 34 0.4695 0.4471 0.4695 0.6852
No log 4.5 36 0.6076 0.5112 0.6076 0.7795
No log 4.75 38 0.4910 0.5373 0.4910 0.7007
No log 5.0 40 0.4398 0.5939 0.4398 0.6632
No log 5.25 42 0.4167 0.6517 0.4167 0.6456
No log 5.5 44 0.4540 0.5779 0.4540 0.6738
No log 5.75 46 0.5788 0.5587 0.5788 0.7608
No log 6.0 48 0.8069 0.4464 0.8069 0.8983
No log 6.25 50 0.7189 0.4633 0.7189 0.8479
No log 6.5 52 0.4613 0.6017 0.4613 0.6792
No log 6.75 54 0.3903 0.6929 0.3903 0.6247
No log 7.0 56 0.4025 0.6598 0.4025 0.6344
No log 7.25 58 0.4032 0.6339 0.4032 0.6349
No log 7.5 60 0.4184 0.6946 0.4184 0.6468
No log 7.75 62 0.4944 0.6047 0.4944 0.7032
No log 8.0 64 0.4832 0.6114 0.4832 0.6951
No log 8.25 66 0.4557 0.6010 0.4557 0.6750
No log 8.5 68 0.4601 0.6068 0.4601 0.6783
No log 8.75 70 0.4618 0.5985 0.4618 0.6796
No log 9.0 72 0.4758 0.6514 0.4758 0.6898
No log 9.25 74 0.4553 0.6828 0.4553 0.6748
No log 9.5 76 0.4966 0.5514 0.4966 0.7047
No log 9.75 78 0.4454 0.6241 0.4454 0.6674
No log 10.0 80 0.4359 0.7012 0.4359 0.6602
No log 10.25 82 0.4307 0.7012 0.4307 0.6563
No log 10.5 84 0.4373 0.6346 0.4373 0.6613
No log 10.75 86 0.4976 0.5514 0.4976 0.7054
No log 11.0 88 0.4672 0.5723 0.4672 0.6835
No log 11.25 90 0.4436 0.6395 0.4436 0.6661
No log 11.5 92 0.4511 0.5782 0.4511 0.6716
No log 11.75 94 0.4826 0.5632 0.4826 0.6947
No log 12.0 96 0.4862 0.5560 0.4862 0.6973
No log 12.25 98 0.4290 0.6210 0.4290 0.6550
No log 12.5 100 0.4252 0.6694 0.4252 0.6521
No log 12.75 102 0.4218 0.6326 0.4218 0.6495
No log 13.0 104 0.4345 0.5961 0.4345 0.6592
No log 13.25 106 0.4152 0.6763 0.4152 0.6443
No log 13.5 108 0.5691 0.5657 0.5691 0.7544
No log 13.75 110 0.6412 0.4815 0.6412 0.8008
No log 14.0 112 0.5233 0.6568 0.5233 0.7234
No log 14.25 114 0.4051 0.6928 0.4051 0.6365
No log 14.5 116 0.4153 0.6552 0.4153 0.6444
No log 14.75 118 0.4181 0.6857 0.4181 0.6466
No log 15.0 120 0.4163 0.7302 0.4163 0.6452
No log 15.25 122 0.4130 0.6750 0.4130 0.6427
No log 15.5 124 0.4414 0.6171 0.4414 0.6644
No log 15.75 126 0.4170 0.6503 0.4170 0.6457
No log 16.0 128 0.4081 0.6832 0.4081 0.6388
No log 16.25 130 0.4136 0.6832 0.4136 0.6431
No log 16.5 132 0.4221 0.5930 0.4221 0.6497
No log 16.75 134 0.4370 0.6185 0.4370 0.6610
No log 17.0 136 0.4456 0.5640 0.4456 0.6675
No log 17.25 138 0.5101 0.5871 0.5101 0.7142
No log 17.5 140 0.6113 0.5857 0.6113 0.7818
No log 17.75 142 0.4805 0.5577 0.4805 0.6932
No log 18.0 144 0.4481 0.6115 0.4481 0.6694
No log 18.25 146 0.4500 0.6201 0.4500 0.6708
No log 18.5 148 0.4353 0.6542 0.4353 0.6598
No log 18.75 150 0.4359 0.6636 0.4359 0.6602
No log 19.0 152 0.4442 0.6293 0.4442 0.6665
No log 19.25 154 0.4435 0.5782 0.4435 0.6660
No log 19.5 156 0.4644 0.5189 0.4644 0.6815
No log 19.75 158 0.4754 0.5698 0.4754 0.6895
No log 20.0 160 0.4509 0.5714 0.4509 0.6715
No log 20.25 162 0.4323 0.6422 0.4323 0.6575
No log 20.5 164 0.4261 0.6422 0.4261 0.6528
No log 20.75 166 0.4836 0.5672 0.4836 0.6954
No log 21.0 168 0.8312 0.4490 0.8312 0.9117
No log 21.25 170 1.0020 0.3861 1.0020 1.0010
No log 21.5 172 0.7752 0.4830 0.7752 0.8805
No log 21.75 174 0.4486 0.6285 0.4486 0.6697
No log 22.0 176 0.4976 0.5736 0.4976 0.7054
No log 22.25 178 0.5704 0.4827 0.5704 0.7552
No log 22.5 180 0.4528 0.5736 0.4528 0.6729
No log 22.75 182 0.4399 0.6487 0.4399 0.6633
No log 23.0 184 0.5899 0.5019 0.5899 0.7681
No log 23.25 186 0.6176 0.4580 0.6176 0.7859
No log 23.5 188 0.4752 0.5908 0.4752 0.6893
No log 23.75 190 0.4128 0.6517 0.4128 0.6425
No log 24.0 192 0.4558 0.5308 0.4558 0.6751
No log 24.25 194 0.4158 0.6517 0.4158 0.6449
No log 24.5 196 0.4258 0.6873 0.4258 0.6526
No log 24.75 198 0.4246 0.6678 0.4246 0.6516
No log 25.0 200 0.4116 0.6503 0.4116 0.6416
No log 25.25 202 0.4441 0.6067 0.4441 0.6664
No log 25.5 204 0.4160 0.6503 0.4160 0.6450
No log 25.75 206 0.4072 0.6736 0.4072 0.6381
No log 26.0 208 0.4147 0.6620 0.4147 0.6440
No log 26.25 210 0.4290 0.6620 0.4290 0.6550
No log 26.5 212 0.4320 0.6702 0.4320 0.6572
No log 26.75 214 0.4203 0.6627 0.4203 0.6483
No log 27.0 216 0.4126 0.6435 0.4126 0.6424
No log 27.25 218 0.4116 0.6435 0.4116 0.6416
No log 27.5 220 0.4215 0.6611 0.4215 0.6493
No log 27.75 222 0.4108 0.6724 0.4108 0.6409
No log 28.0 224 0.4202 0.6143 0.4202 0.6482
No log 28.25 226 0.4436 0.6445 0.4436 0.6660
No log 28.5 228 0.3983 0.6407 0.3983 0.6311
No log 28.75 230 0.4082 0.7053 0.4082 0.6389
No log 29.0 232 0.4019 0.6919 0.4019 0.6340
No log 29.25 234 0.4088 0.6282 0.4088 0.6394
No log 29.5 236 0.4288 0.5702 0.4288 0.6548
No log 29.75 238 0.4198 0.6282 0.4198 0.6480
No log 30.0 240 0.4211 0.6503 0.4211 0.6489
No log 30.25 242 0.4395 0.6694 0.4395 0.6630
No log 30.5 244 0.4187 0.6887 0.4187 0.6470
No log 30.75 246 0.3989 0.6577 0.3989 0.6316
No log 31.0 248 0.4956 0.5650 0.4956 0.7040
No log 31.25 250 0.5640 0.4943 0.5640 0.7510
No log 31.5 252 0.4852 0.5460 0.4852 0.6966
No log 31.75 254 0.3910 0.6939 0.3910 0.6253
No log 32.0 256 0.4147 0.6505 0.4147 0.6440
No log 32.25 258 0.4173 0.6505 0.4173 0.6460
No log 32.5 260 0.3921 0.7033 0.3921 0.6262
No log 32.75 262 0.3977 0.6566 0.3977 0.6306
No log 33.0 264 0.4022 0.6377 0.4022 0.6342
No log 33.25 266 0.4228 0.6589 0.4228 0.6502
No log 33.5 268 0.4070 0.6364 0.4070 0.6380
No log 33.75 270 0.4097 0.6330 0.4097 0.6401
No log 34.0 272 0.4132 0.5883 0.4132 0.6428
No log 34.25 274 0.4208 0.5970 0.4208 0.6487
No log 34.5 276 0.4670 0.6381 0.4670 0.6834
No log 34.75 278 0.4765 0.6381 0.4765 0.6903
No log 35.0 280 0.4343 0.6060 0.4343 0.6591
No log 35.25 282 0.4197 0.5782 0.4197 0.6478
No log 35.5 284 0.4153 0.6243 0.4153 0.6444
No log 35.75 286 0.4099 0.6443 0.4099 0.6402
No log 36.0 288 0.4314 0.6691 0.4314 0.6568
No log 36.25 290 0.4329 0.6870 0.4329 0.6580
No log 36.5 292 0.4212 0.7044 0.4212 0.6490
No log 36.75 294 0.4090 0.6751 0.4090 0.6395
No log 37.0 296 0.4433 0.5855 0.4433 0.6658
No log 37.25 298 0.4477 0.5947 0.4477 0.6691
No log 37.5 300 0.4087 0.6197 0.4087 0.6393
No log 37.75 302 0.4013 0.6908 0.4013 0.6335
No log 38.0 304 0.4051 0.6908 0.4051 0.6365
No log 38.25 306 0.3993 0.6745 0.3993 0.6319
No log 38.5 308 0.4117 0.6357 0.4117 0.6416
No log 38.75 310 0.4175 0.5846 0.4175 0.6461
No log 39.0 312 0.4159 0.5915 0.4159 0.6449
No log 39.25 314 0.4161 0.6435 0.4161 0.6451
No log 39.5 316 0.4268 0.6053 0.4268 0.6533
No log 39.75 318 0.4537 0.5266 0.4537 0.6736
No log 40.0 320 0.4532 0.4918 0.4532 0.6732
No log 40.25 322 0.4442 0.5853 0.4442 0.6665
No log 40.5 324 0.4802 0.5467 0.4802 0.6929
No log 40.75 326 0.5420 0.5538 0.5420 0.7362
No log 41.0 328 0.5454 0.5538 0.5454 0.7385
No log 41.25 330 0.5047 0.5712 0.5047 0.7104
No log 41.5 332 0.4534 0.6303 0.4534 0.6734
No log 41.75 334 0.4347 0.6303 0.4347 0.6593
No log 42.0 336 0.4277 0.5899 0.4277 0.6540
No log 42.25 338 0.4268 0.6001 0.4268 0.6533
No log 42.5 340 0.4304 0.6293 0.4304 0.6561
No log 42.75 342 0.4318 0.6293 0.4318 0.6571
No log 43.0 344 0.4340 0.6293 0.4340 0.6588
No log 43.25 346 0.4226 0.6293 0.4226 0.6501
No log 43.5 348 0.4191 0.6001 0.4191 0.6473
No log 43.75 350 0.4388 0.5861 0.4388 0.6624
No log 44.0 352 0.4512 0.5356 0.4512 0.6717
No log 44.25 354 0.4455 0.5171 0.4455 0.6674
No log 44.5 356 0.4352 0.5227 0.4352 0.6597
No log 44.75 358 0.4332 0.5853 0.4332 0.6582
No log 45.0 360 0.4443 0.6293 0.4443 0.6666
No log 45.25 362 0.4515 0.6060 0.4515 0.6720
No log 45.5 364 0.4496 0.6293 0.4496 0.6705
No log 45.75 366 0.4313 0.6091 0.4313 0.6567
No log 46.0 368 0.4286 0.5915 0.4286 0.6546
No log 46.25 370 0.4254 0.5915 0.4254 0.6522
No log 46.5 372 0.4229 0.5915 0.4229 0.6503
No log 46.75 374 0.4220 0.5899 0.4220 0.6496
No log 47.0 376 0.4289 0.6517 0.4289 0.6549
No log 47.25 378 0.4437 0.6611 0.4437 0.6661
No log 47.5 380 0.4386 0.6293 0.4386 0.6622
No log 47.75 382 0.4265 0.5749 0.4265 0.6531
No log 48.0 384 0.4244 0.5379 0.4244 0.6514
No log 48.25 386 0.4320 0.5765 0.4320 0.6573
No log 48.5 388 0.4262 0.5379 0.4262 0.6528
No log 48.75 390 0.4229 0.5765 0.4229 0.6503
No log 49.0 392 0.4252 0.5853 0.4252 0.6521
No log 49.25 394 0.4342 0.6132 0.4342 0.6589
No log 49.5 396 0.4337 0.6132 0.4337 0.6586
No log 49.75 398 0.4236 0.5883 0.4236 0.6509
No log 50.0 400 0.4202 0.5208 0.4202 0.6483
No log 50.25 402 0.4172 0.5765 0.4172 0.6459
No log 50.5 404 0.4133 0.5765 0.4133 0.6429
No log 50.75 406 0.4100 0.5765 0.4100 0.6403
No log 51.0 408 0.4094 0.5765 0.4094 0.6398
No log 51.25 410 0.4086 0.5986 0.4086 0.6392
No log 51.5 412 0.4144 0.6503 0.4144 0.6438
No log 51.75 414 0.4162 0.6395 0.4162 0.6451
No log 52.0 416 0.4123 0.6503 0.4123 0.6421
No log 52.25 418 0.3993 0.6832 0.3993 0.6319
No log 52.5 420 0.4032 0.6269 0.4032 0.6350
No log 52.75 422 0.4179 0.6880 0.4179 0.6464
No log 53.0 424 0.4246 0.6431 0.4246 0.6517
No log 53.25 426 0.4119 0.6269 0.4119 0.6418
No log 53.5 428 0.4103 0.6001 0.4103 0.6406
No log 53.75 430 0.4176 0.6001 0.4176 0.6462
No log 54.0 432 0.4251 0.5457 0.4251 0.6520
No log 54.25 434 0.4306 0.5457 0.4306 0.6562
No log 54.5 436 0.4297 0.5457 0.4297 0.6555
No log 54.75 438 0.4260 0.5457 0.4260 0.6527
No log 55.0 440 0.4272 0.5457 0.4272 0.6536
No log 55.25 442 0.4353 0.6341 0.4353 0.6598
No log 55.5 444 0.4443 0.6776 0.4443 0.6666
No log 55.75 446 0.4441 0.6147 0.4441 0.6664
No log 56.0 448 0.4469 0.5784 0.4469 0.6685
No log 56.25 450 0.4475 0.5189 0.4475 0.6690
No log 56.5 452 0.4520 0.5286 0.4520 0.6723
No log 56.75 454 0.4532 0.5286 0.4532 0.6732
No log 57.0 456 0.4510 0.5286 0.4510 0.6715
No log 57.25 458 0.4487 0.5539 0.4487 0.6698
No log 57.5 460 0.4447 0.5539 0.4447 0.6669
No log 57.75 462 0.4433 0.5539 0.4433 0.6658
No log 58.0 464 0.4507 0.5286 0.4507 0.6714
No log 58.25 466 0.4554 0.5024 0.4554 0.6748
No log 58.5 468 0.4499 0.5286 0.4499 0.6707
No log 58.75 470 0.4458 0.5539 0.4458 0.6676
No log 59.0 472 0.4460 0.5641 0.4460 0.6678
No log 59.25 474 0.4428 0.5539 0.4428 0.6654
No log 59.5 476 0.4458 0.5286 0.4458 0.6677
No log 59.75 478 0.4512 0.5503 0.4512 0.6717
No log 60.0 480 0.4575 0.5732 0.4575 0.6764
No log 60.25 482 0.4565 0.5503 0.4565 0.6757
No log 60.5 484 0.4484 0.5503 0.4484 0.6696
No log 60.75 486 0.4419 0.5853 0.4419 0.6647
No log 61.0 488 0.4388 0.5853 0.4388 0.6624
No log 61.25 490 0.4300 0.5539 0.4300 0.6557
No log 61.5 492 0.4313 0.5457 0.4313 0.6567
No log 61.75 494 0.4263 0.5698 0.4263 0.6529
No log 62.0 496 0.4180 0.6001 0.4180 0.6465
No log 62.25 498 0.4135 0.6091 0.4135 0.6431
0.204 62.5 500 0.4229 0.6200 0.4229 0.6503
0.204 62.75 502 0.4566 0.6506 0.4566 0.6757
0.204 63.0 504 0.4909 0.5953 0.4909 0.7007
0.204 63.25 506 0.5118 0.5789 0.5118 0.7154
0.204 63.5 508 0.4912 0.5858 0.4912 0.7009
0.204 63.75 510 0.4511 0.6506 0.4511 0.6716
0.204 64.0 512 0.4237 0.6010 0.4237 0.6509
0.204 64.25 514 0.4207 0.6091 0.4207 0.6486
0.204 64.5 516 0.4203 0.6091 0.4203 0.6483
0.204 64.75 518 0.4184 0.6001 0.4184 0.6468
0.204 65.0 520 0.4182 0.6001 0.4182 0.6467

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
8
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k2_task7_organization

Finetuned
(4223)
this model