ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k4_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8016
  • Qwk: -0.0690
  • Mse: 0.8016
  • Rmse: 0.8953

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.2 2 4.5110 0.0163 4.5110 2.1239
No log 0.4 4 2.8700 0.0308 2.8700 1.6941
No log 0.6 6 1.5154 0.0444 1.5154 1.2310
No log 0.8 8 2.0564 0.0084 2.0564 1.4340
No log 1.0 10 2.5645 0.0115 2.5645 1.6014
No log 1.2 12 2.1755 -0.0092 2.1755 1.4749
No log 1.4 14 0.7674 -0.0287 0.7674 0.8760
No log 1.6 16 0.6576 0.0 0.6576 0.8109
No log 1.8 18 0.7380 0.0 0.7380 0.8591
No log 2.0 20 0.8484 0.0346 0.8484 0.9211
No log 2.2 22 1.0390 -0.0234 1.0390 1.0193
No log 2.4 24 1.1125 -0.0234 1.1125 1.0547
No log 2.6 26 0.9441 -0.0084 0.9441 0.9717
No log 2.8 28 0.7352 0.0460 0.7352 0.8575
No log 3.0 30 0.7047 0.0506 0.7047 0.8395
No log 3.2 32 0.7370 0.0460 0.7370 0.8585
No log 3.4 34 0.8144 -0.0264 0.8144 0.9024
No log 3.6 36 1.0106 -0.0435 1.0106 1.0053
No log 3.8 38 1.0994 -0.0207 1.0994 1.0485
No log 4.0 40 1.0389 -0.0411 1.0389 1.0193
No log 4.2 42 1.1896 -0.0435 1.1896 1.0907
No log 4.4 44 1.1784 -0.0997 1.1784 1.0855
No log 4.6 46 0.9737 0.0111 0.9737 0.9868
No log 4.8 48 1.0285 0.0046 1.0285 1.0141
No log 5.0 50 0.9702 0.0089 0.9702 0.9850
No log 5.2 52 0.8210 0.0442 0.8210 0.9061
No log 5.4 54 0.7553 0.1449 0.7553 0.8691
No log 5.6 56 0.9214 0.0545 0.9214 0.9599
No log 5.8 58 0.7915 0.1196 0.7915 0.8897
No log 6.0 60 0.6854 0.1379 0.6854 0.8279
No log 6.2 62 0.7735 0.1899 0.7735 0.8795
No log 6.4 64 1.1219 0.0753 1.1219 1.0592
No log 6.6 66 0.7652 0.1573 0.7652 0.8747
No log 6.8 68 0.8511 -0.0616 0.8511 0.9225
No log 7.0 70 0.8701 -0.0970 0.8701 0.9328
No log 7.2 72 0.8239 0.1415 0.8239 0.9077
No log 7.4 74 1.5174 0.0083 1.5174 1.2318
No log 7.6 76 1.4509 -0.0178 1.4509 1.2045
No log 7.8 78 0.9376 -0.0539 0.9376 0.9683
No log 8.0 80 0.7833 -0.0738 0.7833 0.8851
No log 8.2 82 0.7780 -0.0366 0.7780 0.8821
No log 8.4 84 0.8500 0.0062 0.8500 0.9220
No log 8.6 86 1.0576 -0.0169 1.0576 1.0284
No log 8.8 88 1.0546 0.0200 1.0546 1.0270
No log 9.0 90 0.9312 -0.0052 0.9312 0.9650
No log 9.2 92 0.7635 -0.0056 0.7635 0.8738
No log 9.4 94 0.7588 -0.0407 0.7588 0.8711
No log 9.6 96 0.7291 0.0414 0.7291 0.8539
No log 9.8 98 0.8106 0.1342 0.8106 0.9004
No log 10.0 100 0.8033 0.1342 0.8033 0.8963
No log 10.2 102 0.7053 0.1444 0.7053 0.8398
No log 10.4 104 0.7060 0.1021 0.7059 0.8402
No log 10.6 106 0.7709 0.1965 0.7709 0.8780
No log 10.8 108 0.8398 0.0786 0.8398 0.9164
No log 11.0 110 0.7318 0.0909 0.7318 0.8554
No log 11.2 112 0.7636 -0.1329 0.7636 0.8738
No log 11.4 114 0.7833 0.0053 0.7833 0.8850
No log 11.6 116 0.9985 -0.0441 0.9985 0.9992
No log 11.8 118 0.9323 -0.0391 0.9323 0.9656
No log 12.0 120 0.8697 0.0335 0.8697 0.9326
No log 12.2 122 0.9643 0.0164 0.9643 0.9820
No log 12.4 124 0.8485 -0.0015 0.8485 0.9212
No log 12.6 126 0.8690 -0.0008 0.8690 0.9322
No log 12.8 128 1.0881 -0.0586 1.0881 1.0431
No log 13.0 130 0.8712 0.0362 0.8712 0.9334
No log 13.2 132 0.7453 -0.0096 0.7453 0.8633
No log 13.4 134 0.8003 -0.1606 0.8003 0.8946
No log 13.6 136 0.7975 -0.0881 0.7975 0.8930
No log 13.8 138 0.7667 -0.1026 0.7667 0.8756
No log 14.0 140 0.7980 -0.0170 0.7980 0.8933
No log 14.2 142 0.8482 0.0152 0.8482 0.9210
No log 14.4 144 0.7704 0.0395 0.7704 0.8777
No log 14.6 146 0.7934 -0.0939 0.7934 0.8907
No log 14.8 148 0.8419 -0.1753 0.8419 0.9175
No log 15.0 150 0.7806 -0.1001 0.7806 0.8835
No log 15.2 152 0.7529 0.0225 0.7529 0.8677
No log 15.4 154 0.8726 0.0826 0.8726 0.9341
No log 15.6 156 0.8358 0.1342 0.8358 0.9142
No log 15.8 158 0.7564 0.0260 0.7564 0.8697
No log 16.0 160 0.7865 0.0031 0.7865 0.8868
No log 16.2 162 0.8127 -0.0366 0.8127 0.9015
No log 16.4 164 0.8071 -0.0595 0.8071 0.8984
No log 16.6 166 0.8370 0.1047 0.8370 0.9149
No log 16.8 168 0.9720 -0.0182 0.9720 0.9859
No log 17.0 170 0.9310 -0.0163 0.9310 0.9649
No log 17.2 172 0.7826 0.0282 0.7826 0.8846
No log 17.4 174 0.7614 -0.0062 0.7614 0.8726
No log 17.6 176 0.7589 0.0414 0.7589 0.8712
No log 17.8 178 0.7685 0.0814 0.7685 0.8767
No log 18.0 180 0.7925 0.1199 0.7925 0.8902
No log 18.2 182 0.7987 0.1254 0.7987 0.8937
No log 18.4 184 0.7881 0.0375 0.7881 0.8877
No log 18.6 186 0.7708 -0.0493 0.7708 0.8779
No log 18.8 188 0.7395 0.0061 0.7395 0.8600
No log 19.0 190 0.7288 0.0282 0.7288 0.8537
No log 19.2 192 0.7389 0.0214 0.7389 0.8596
No log 19.4 194 0.7512 0.0214 0.7512 0.8667
No log 19.6 196 0.6989 -0.0032 0.6989 0.8360
No log 19.8 198 0.7054 0.0479 0.7054 0.8399
No log 20.0 200 0.7791 0.0525 0.7791 0.8827
No log 20.2 202 0.8539 -0.0845 0.8539 0.9240
No log 20.4 204 0.7933 0.0956 0.7933 0.8907
No log 20.6 206 0.7146 -0.0033 0.7146 0.8454
No log 20.8 208 0.7222 -0.0451 0.7222 0.8498
No log 21.0 210 0.6956 0.1021 0.6956 0.8340
No log 21.2 212 0.8007 0.1291 0.8007 0.8948
No log 21.4 214 0.9794 -0.0923 0.9794 0.9896
No log 21.6 216 0.9072 -0.0122 0.9072 0.9525
No log 21.8 218 0.7914 0.0247 0.7914 0.8896
No log 22.0 220 0.7606 -0.0541 0.7606 0.8721
No log 22.2 222 0.7631 -0.0541 0.7631 0.8735
No log 22.4 224 0.7971 0.0869 0.7971 0.8928
No log 22.6 226 0.8021 0.0303 0.8021 0.8956
No log 22.8 228 0.7865 0.0869 0.7865 0.8869
No log 23.0 230 0.7566 0.0914 0.7566 0.8698
No log 23.2 232 0.7439 0.0436 0.7439 0.8625
No log 23.4 234 0.7324 0.0 0.7324 0.8558
No log 23.6 236 0.7397 0.0436 0.7397 0.8601
No log 23.8 238 0.8567 -0.1665 0.8567 0.9256
No log 24.0 240 0.9868 0.0353 0.9868 0.9934
No log 24.2 242 0.9174 -0.0916 0.9174 0.9578
No log 24.4 244 0.7618 0.1254 0.7618 0.8728
No log 24.6 246 0.7276 0.0436 0.7276 0.8530
No log 24.8 248 0.7248 0.0602 0.7248 0.8514
No log 25.0 250 0.7258 0.0964 0.7258 0.8519
No log 25.2 252 0.7890 -0.0823 0.7890 0.8882
No log 25.4 254 0.7725 -0.0351 0.7725 0.8789
No log 25.6 256 0.7144 0.0964 0.7144 0.8452
No log 25.8 258 0.7582 -0.0428 0.7582 0.8707
No log 26.0 260 0.8152 -0.1142 0.8152 0.9029
No log 26.2 262 0.8416 -0.0551 0.8416 0.9174
No log 26.4 264 0.7396 0.1137 0.7396 0.8600
No log 26.6 266 0.7103 -0.0096 0.7103 0.8428
No log 26.8 268 0.7425 0.0863 0.7425 0.8617
No log 27.0 270 0.7379 0.0375 0.7379 0.8590
No log 27.2 272 0.7481 0.0375 0.7481 0.8649
No log 27.4 274 0.7203 -0.0541 0.7203 0.8487
No log 27.6 276 0.7188 -0.0560 0.7188 0.8478
No log 27.8 278 0.7210 0.0436 0.7210 0.8491
No log 28.0 280 0.7184 0.0436 0.7184 0.8476
No log 28.2 282 0.7512 0.0282 0.7512 0.8667
No log 28.4 284 0.7490 0.0282 0.7490 0.8655
No log 28.6 286 0.7312 0.0914 0.7312 0.8551
No log 28.8 288 0.7490 0.0282 0.7490 0.8654
No log 29.0 290 0.7769 0.0714 0.7769 0.8814
No log 29.2 292 0.7482 0.0318 0.7482 0.8650
No log 29.4 294 0.7522 0.0759 0.7522 0.8673
No log 29.6 296 0.7437 0.0807 0.7437 0.8624
No log 29.8 298 0.7255 0.0964 0.7255 0.8518
No log 30.0 300 0.7136 0.0964 0.7136 0.8447
No log 30.2 302 0.7162 0.0964 0.7162 0.8463
No log 30.4 304 0.7122 0.0436 0.7122 0.8439
No log 30.6 306 0.7174 0.0436 0.7174 0.8470
No log 30.8 308 0.7281 0.0869 0.7281 0.8533
No log 31.0 310 0.7584 0.0289 0.7584 0.8709
No log 31.2 312 0.7929 -0.0262 0.7929 0.8905
No log 31.4 314 0.8013 -0.0262 0.8013 0.8952
No log 31.6 316 0.7596 0.0741 0.7596 0.8716
No log 31.8 318 0.7473 0.0269 0.7473 0.8645
No log 32.0 320 0.7204 0.0355 0.7204 0.8488
No log 32.2 322 0.7068 0.1023 0.7068 0.8407
No log 32.4 324 0.7154 0.1023 0.7154 0.8458
No log 32.6 326 0.7203 0.0454 0.7203 0.8487
No log 32.8 328 0.7539 -0.0252 0.7539 0.8683
No log 33.0 330 0.7729 0.0628 0.7729 0.8791
No log 33.2 332 0.7709 0.0628 0.7709 0.8780
No log 33.4 334 0.7809 0.0183 0.7809 0.8837
No log 33.6 336 0.8551 -0.1221 0.8551 0.9247
No log 33.8 338 0.9302 -0.0558 0.9302 0.9645
No log 34.0 340 0.8815 -0.0163 0.8815 0.9389
No log 34.2 342 0.7879 -0.0252 0.7879 0.8876
No log 34.4 344 0.7720 -0.0473 0.7720 0.8787
No log 34.6 346 0.8249 -0.1146 0.8249 0.9083
No log 34.8 348 0.8470 -0.0551 0.8470 0.9203
No log 35.0 350 0.8203 -0.1413 0.8203 0.9057
No log 35.2 352 0.7538 0.0628 0.7538 0.8682
No log 35.4 354 0.7302 0.1021 0.7302 0.8545
No log 35.6 356 0.8375 -0.1273 0.8375 0.9152
No log 35.8 358 0.9298 0.0378 0.9298 0.9643
No log 36.0 360 0.9172 0.0067 0.9172 0.9577
No log 36.2 362 0.8124 -0.0490 0.8124 0.9013
No log 36.4 364 0.7511 0.0999 0.7511 0.8667
No log 36.6 366 0.7360 -0.0032 0.7360 0.8579
No log 36.8 368 0.7486 0.0033 0.7486 0.8652
No log 37.0 370 0.7476 -0.0541 0.7476 0.8646
No log 37.2 372 0.7491 0.0414 0.7491 0.8655
No log 37.4 374 0.7612 0.0282 0.7612 0.8725
No log 37.6 376 0.7758 0.0214 0.7758 0.8808
No log 37.8 378 0.8226 -0.0442 0.8226 0.9070
No log 38.0 380 0.8868 -0.1274 0.8868 0.9417
No log 38.2 382 0.9473 -0.0331 0.9473 0.9733
No log 38.4 384 0.8907 -0.0916 0.8907 0.9438
No log 38.6 386 0.7970 0.0017 0.7970 0.8927
No log 38.8 388 0.7358 0.0355 0.7358 0.8578
No log 39.0 390 0.7256 0.0541 0.7256 0.8518
No log 39.2 392 0.7272 0.1023 0.7272 0.8527
No log 39.4 394 0.7496 -0.0228 0.7496 0.8658
No log 39.6 396 0.8048 0.0043 0.8048 0.8971
No log 39.8 398 0.8155 -0.0442 0.8155 0.9030
No log 40.0 400 0.8105 0.0043 0.8105 0.9003
No log 40.2 402 0.7864 0.0159 0.7864 0.8868
No log 40.4 404 0.7591 -0.0118 0.7591 0.8713
No log 40.6 406 0.7582 0.0454 0.7582 0.8708
No log 40.8 408 0.7824 -0.0228 0.7824 0.8845
No log 41.0 410 0.7983 0.0225 0.7983 0.8935
No log 41.2 412 0.8583 0.0017 0.8583 0.9264
No log 41.4 414 0.9477 -0.0930 0.9477 0.9735
No log 41.6 416 0.9561 -0.0912 0.9561 0.9778
No log 41.8 418 0.9013 -0.0138 0.9013 0.9494
No log 42.0 420 0.8214 0.0183 0.8214 0.9063
No log 42.2 422 0.7641 0.0914 0.7641 0.8741
No log 42.4 424 0.7546 0.0914 0.7546 0.8687
No log 42.6 426 0.7699 0.0914 0.7699 0.8774
No log 42.8 428 0.7575 0.0318 0.7575 0.8703
No log 43.0 430 0.7803 0.0095 0.7803 0.8833
No log 43.2 432 0.8714 -0.0518 0.8714 0.9335
No log 43.4 434 1.0199 0.0260 1.0199 1.0099
No log 43.6 436 1.0640 -0.0398 1.0640 1.0315
No log 43.8 438 0.9873 -0.0013 0.9873 0.9936
No log 44.0 440 0.8210 -0.0408 0.8210 0.9061
No log 44.2 442 0.7673 0.0205 0.7673 0.8759
No log 44.4 444 0.7636 0.0205 0.7636 0.8739
No log 44.6 446 0.7825 0.0146 0.7825 0.8846
No log 44.8 448 0.8527 -0.0456 0.8527 0.9234
No log 45.0 450 0.9534 -0.0320 0.9535 0.9764
No log 45.2 452 1.0374 0.0196 1.0374 1.0185
No log 45.4 454 1.0153 0.0217 1.0153 1.0076
No log 45.6 456 0.9421 -0.0301 0.9421 0.9706
No log 45.8 458 0.8409 -0.0474 0.8409 0.9170
No log 46.0 460 0.7876 -0.0373 0.7876 0.8875
No log 46.2 462 0.7454 0.0768 0.7454 0.8633
No log 46.4 464 0.7414 0.0318 0.7414 0.8611
No log 46.6 466 0.7491 0.0247 0.7491 0.8655
No log 46.8 468 0.7499 0.0732 0.7499 0.8660
No log 47.0 470 0.7459 0.0783 0.7459 0.8637
No log 47.2 472 0.7447 0.0783 0.7447 0.8630
No log 47.4 474 0.7497 0.0432 0.7497 0.8658
No log 47.6 476 0.7591 0.0357 0.7591 0.8712
No log 47.8 478 0.7990 0.0129 0.7990 0.8939
No log 48.0 480 0.8565 -0.1601 0.8565 0.9255
No log 48.2 482 0.8792 -0.1219 0.8792 0.9377
No log 48.4 484 0.8646 -0.0809 0.8646 0.9299
No log 48.6 486 0.8459 -0.0778 0.8459 0.9197
No log 48.8 488 0.7942 -0.0204 0.7942 0.8912
No log 49.0 490 0.7714 -0.0125 0.7714 0.8783
No log 49.2 492 0.7782 -0.0690 0.7782 0.8822
No log 49.4 494 0.8154 0.0071 0.8154 0.9030
No log 49.6 496 0.9166 -0.0236 0.9166 0.9574
No log 49.8 498 1.0327 -0.0101 1.0327 1.0162
0.2401 50.0 500 1.0581 -0.0101 1.0581 1.0286
0.2401 50.2 502 0.9721 0.0282 0.9721 0.9860
0.2401 50.4 504 0.8666 -0.0425 0.8666 0.9309
0.2401 50.6 506 0.8175 -0.0355 0.8175 0.9042
0.2401 50.8 508 0.8337 -0.0788 0.8337 0.9131
0.2401 51.0 510 0.8290 -0.0743 0.8290 0.9105
0.2401 51.2 512 0.8016 -0.0690 0.8016 0.8953

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
30
Safetensors
Model size
135M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k4_task3_organization

Finetuned
(3994)
this model