ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k16_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8375
  • Qwk: -0.1860
  • Mse: 0.8375
  • Rmse: 0.9151

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.05 2 3.3980 0.0062 3.3980 1.8434
No log 0.1 4 1.9708 0.0325 1.9708 1.4038
No log 0.15 6 2.1479 0.0247 2.1479 1.4656
No log 0.2 8 1.5226 0.0014 1.5226 1.2340
No log 0.25 10 0.8722 0.0714 0.8722 0.9339
No log 0.3 12 0.8362 -0.0264 0.8362 0.9144
No log 0.35 14 1.2668 -0.0744 1.2668 1.1255
No log 0.4 16 2.4290 -0.0305 2.4290 1.5585
No log 0.45 18 2.2510 -0.0299 2.2510 1.5003
No log 0.5 20 1.2395 -0.0500 1.2395 1.1133
No log 0.55 22 1.0327 0.0735 1.0327 1.0162
No log 0.6 24 0.9211 0.1374 0.9211 0.9597
No log 0.65 26 0.9035 0.0651 0.9035 0.9505
No log 0.7 28 0.9468 0.0196 0.9468 0.9730
No log 0.75 30 1.0873 -0.0193 1.0873 1.0427
No log 0.8 32 1.0973 -0.0490 1.0973 1.0475
No log 0.85 34 1.1334 -0.0490 1.1334 1.0646
No log 0.9 36 1.1097 -0.0736 1.1097 1.0534
No log 0.95 38 1.1091 -0.0736 1.1091 1.0531
No log 1.0 40 0.9710 -0.1289 0.9710 0.9854
No log 1.05 42 0.8849 -0.0518 0.8849 0.9407
No log 1.1 44 1.0090 -0.0398 1.0090 1.0045
No log 1.15 46 1.0081 -0.0695 1.0081 1.0040
No log 1.2 48 1.0240 -0.0149 1.0240 1.0119
No log 1.25 50 1.1165 -0.0457 1.1165 1.0566
No log 1.3 52 1.7257 -0.0234 1.7257 1.3136
No log 1.35 54 2.0120 0.0213 2.0120 1.4185
No log 1.4 56 1.5820 -0.0234 1.5820 1.2578
No log 1.45 58 0.9980 0.0176 0.9980 0.9990
No log 1.5 60 0.7930 -0.1236 0.7930 0.8905
No log 1.55 62 0.7940 -0.1738 0.7940 0.8911
No log 1.6 64 0.8188 -0.1251 0.8188 0.9049
No log 1.65 66 0.8699 -0.0218 0.8699 0.9327
No log 1.7 68 1.0542 -0.0468 1.0542 1.0267
No log 1.75 70 1.3178 -0.0234 1.3178 1.1480
No log 1.8 72 1.3516 -0.0234 1.3516 1.1626
No log 1.85 74 1.2642 -0.0234 1.2642 1.1244
No log 1.9 76 1.1988 -0.0207 1.1988 1.0949
No log 1.95 78 1.4454 -0.0234 1.4454 1.2022
No log 2.0 80 1.5705 -0.0234 1.5705 1.2532
No log 2.05 82 1.7435 0.0224 1.7435 1.3204
No log 2.1 84 1.8182 0.1052 1.8182 1.3484
No log 2.15 86 1.4009 0.0522 1.4009 1.1836
No log 2.2 88 0.8701 -0.0200 0.8701 0.9328
No log 2.25 90 0.8598 0.1107 0.8598 0.9273
No log 2.3 92 1.1318 0.0552 1.1318 1.0639
No log 2.35 94 1.4298 -0.0169 1.4298 1.1958
No log 2.4 96 1.7719 -0.0035 1.7719 1.3311
No log 2.45 98 1.6684 -0.0197 1.6684 1.2916
No log 2.5 100 1.0103 0.1846 1.0103 1.0051
No log 2.55 102 0.8119 0.1440 0.8119 0.9011
No log 2.6 104 0.9415 0.1896 0.9415 0.9703
No log 2.65 106 1.8861 0.0144 1.8861 1.3734
No log 2.7 108 1.9385 0.0100 1.9385 1.3923
No log 2.75 110 1.4843 -0.0465 1.4843 1.2183
No log 2.8 112 0.9572 0.0498 0.9572 0.9783
No log 2.85 114 0.9575 0.0515 0.9575 0.9785
No log 2.9 116 1.1685 -0.0345 1.1685 1.0810
No log 2.95 118 1.0505 0.0046 1.0505 1.0249
No log 3.0 120 0.8404 0.0442 0.8404 0.9167
No log 3.05 122 0.9039 -0.0101 0.9039 0.9508
No log 3.1 124 1.0932 0.0061 1.0932 1.0456
No log 3.15 126 0.9439 0.0576 0.9439 0.9716
No log 3.2 128 0.7306 0.0414 0.7306 0.8547
No log 3.25 130 0.7530 -0.0062 0.7530 0.8677
No log 3.3 132 1.2147 0.0059 1.2147 1.1021
No log 3.35 134 1.4358 -0.0028 1.4358 1.1983
No log 3.4 136 0.9147 0.1190 0.9147 0.9564
No log 3.45 138 0.8571 0.1004 0.8571 0.9258
No log 3.5 140 0.7939 0.0660 0.7939 0.8910
No log 3.55 142 0.9006 0.1147 0.9006 0.9490
No log 3.6 144 1.0425 0.0492 1.0425 1.0210
No log 3.65 146 0.7761 0.0732 0.7761 0.8810
No log 3.7 148 0.7586 0.1199 0.7586 0.8710
No log 3.75 150 0.8089 0.1047 0.8089 0.8994
No log 3.8 152 0.8439 0.1387 0.8439 0.9187
No log 3.85 154 0.8281 0.1336 0.8281 0.9100
No log 3.9 156 0.7730 0.1199 0.7730 0.8792
No log 3.95 158 0.7686 0.0269 0.7686 0.8767
No log 4.0 160 0.7838 0.0289 0.7838 0.8853
No log 4.05 162 0.8709 0.0741 0.8709 0.9332
No log 4.1 164 0.9673 0.0885 0.9673 0.9835
No log 4.15 166 0.8567 0.0529 0.8567 0.9256
No log 4.2 168 0.7672 0.0978 0.7672 0.8759
No log 4.25 170 0.7568 0.0543 0.7568 0.8699
No log 4.3 172 0.7545 0.1030 0.7545 0.8686
No log 4.35 174 0.7599 0.1189 0.7599 0.8717
No log 4.4 176 0.7794 0.0611 0.7794 0.8828
No log 4.45 178 0.7299 0.0543 0.7299 0.8544
No log 4.5 180 0.7442 0.0089 0.7442 0.8626
No log 4.55 182 0.7792 0.0338 0.7792 0.8827
No log 4.6 184 0.8163 0.0269 0.8163 0.9035
No log 4.65 186 0.8130 0.0269 0.8130 0.9017
No log 4.7 188 0.7804 -0.0912 0.7804 0.8834
No log 4.75 190 0.7917 0.0 0.7917 0.8898
No log 4.8 192 0.8124 -0.0533 0.8124 0.9013
No log 4.85 194 0.8226 -0.0549 0.8226 0.9070
No log 4.9 196 0.8317 -0.0387 0.8317 0.9120
No log 4.95 198 0.8409 0.0056 0.8409 0.9170
No log 5.0 200 0.9602 0.0081 0.9602 0.9799
No log 5.05 202 1.5578 -0.0927 1.5578 1.2481
No log 5.1 204 1.4374 -0.0366 1.4374 1.1989
No log 5.15 206 0.9327 0.0062 0.9327 0.9658
No log 5.2 208 0.8426 0.1537 0.8426 0.9179
No log 5.25 210 0.9593 0.0465 0.9593 0.9794
No log 5.3 212 0.8509 0.1466 0.8509 0.9225
No log 5.35 214 0.7798 -0.1116 0.7798 0.8831
No log 5.4 216 0.8301 0.0043 0.8301 0.9111
No log 5.45 218 0.7434 -0.0612 0.7434 0.8622
No log 5.5 220 0.7901 0.0116 0.7901 0.8889
No log 5.55 222 0.8026 -0.0798 0.8026 0.8959
No log 5.6 224 0.7562 0.0528 0.7562 0.8696
No log 5.65 226 0.8290 -0.0283 0.8290 0.9105
No log 5.7 228 0.9703 -0.0128 0.9703 0.9850
No log 5.75 230 0.8411 -0.0303 0.8411 0.9171
No log 5.8 232 0.7918 0.0116 0.7918 0.8898
No log 5.85 234 0.8623 -0.0061 0.8623 0.9286
No log 5.9 236 0.8824 -0.0061 0.8824 0.9393
No log 5.95 238 0.8482 -0.1045 0.8482 0.9210
No log 6.0 240 0.7918 -0.0387 0.7918 0.8898
No log 6.05 242 0.9314 0.0336 0.9314 0.9651
No log 6.1 244 1.2443 0.0666 1.2443 1.1155
No log 6.15 246 1.1552 0.1043 1.1552 1.0748
No log 6.2 248 0.8977 -0.0341 0.8977 0.9475
No log 6.25 250 0.7961 0.0116 0.7961 0.8922
No log 6.3 252 0.7840 0.0116 0.7840 0.8854
No log 6.35 254 0.8039 -0.1459 0.8039 0.8966
No log 6.4 256 0.9199 -0.0359 0.9199 0.9591
No log 6.45 258 1.0193 0.0627 1.0193 1.0096
No log 6.5 260 0.9275 0.0062 0.9275 0.9630
No log 6.55 262 0.8275 0.0135 0.8275 0.9097
No log 6.6 264 0.8362 0.0268 0.8362 0.9145
No log 6.65 266 0.8892 -0.0363 0.8892 0.9430
No log 6.7 268 0.8392 -0.0528 0.8392 0.9161
No log 6.75 270 0.8092 0.0030 0.8092 0.8996
No log 6.8 272 0.8861 -0.1557 0.8861 0.9413
No log 6.85 274 0.9530 -0.0854 0.9530 0.9762
No log 6.9 276 0.8476 -0.0303 0.8476 0.9207
No log 6.95 278 0.7574 -0.0030 0.7574 0.8703
No log 7.0 280 0.7765 -0.1397 0.7765 0.8812
No log 7.05 282 0.8307 -0.1045 0.8307 0.9114
No log 7.1 284 0.8350 -0.0268 0.8350 0.9138
No log 7.15 286 1.0571 -0.0504 1.0571 1.0281
No log 7.2 288 1.1950 0.0006 1.1950 1.0932
No log 7.25 290 0.9602 -0.0118 0.9602 0.9799
No log 7.3 292 0.8245 0.0570 0.8245 0.9080
No log 7.35 294 1.0252 0.0175 1.0252 1.0125
No log 7.4 296 1.0258 0.0476 1.0258 1.0128
No log 7.45 298 0.8855 -0.0941 0.8855 0.9410
No log 7.5 300 0.8090 0.0834 0.8090 0.8994
No log 7.55 302 0.9905 -0.0440 0.9905 0.9952
No log 7.6 304 1.0509 0.0443 1.0509 1.0251
No log 7.65 306 0.9313 -0.0787 0.9313 0.9650
No log 7.7 308 0.8171 -0.1994 0.8171 0.9039
No log 7.75 310 0.7983 -0.0532 0.7983 0.8935
No log 7.8 312 0.8142 -0.2008 0.8142 0.9023
No log 7.85 314 0.8196 0.0061 0.8196 0.9053
No log 7.9 316 0.8456 0.0030 0.8456 0.9196
No log 7.95 318 0.8878 -0.1515 0.8878 0.9423
No log 8.0 320 0.9298 -0.0303 0.9298 0.9643
No log 8.05 322 0.9261 -0.0359 0.9261 0.9624
No log 8.1 324 0.8904 -0.0955 0.8904 0.9436
No log 8.15 326 0.8764 -0.1277 0.8764 0.9362
No log 8.2 328 0.9198 -0.0939 0.9198 0.9591
No log 8.25 330 0.9723 -0.0690 0.9723 0.9861
No log 8.3 332 0.9964 0.0309 0.9964 0.9982
No log 8.35 334 0.9461 0.0421 0.9461 0.9727
No log 8.4 336 0.9206 0.0087 0.9206 0.9595
No log 8.45 338 0.9655 0.1147 0.9655 0.9826
No log 8.5 340 0.9827 0.0986 0.9827 0.9913
No log 8.55 342 0.8744 0.0956 0.8744 0.9351
No log 8.6 344 0.7491 0.0471 0.7491 0.8655
No log 8.65 346 0.7317 0.0061 0.7317 0.8554
No log 8.7 348 0.7638 -0.0488 0.7638 0.8740
No log 8.75 350 0.8191 -0.0163 0.8191 0.9050
No log 8.8 352 0.8401 -0.0163 0.8401 0.9166
No log 8.85 354 0.8436 -0.0138 0.8436 0.9185
No log 8.9 356 0.8878 0.0196 0.8878 0.9423
No log 8.95 358 1.0547 -0.0521 1.0547 1.0270
No log 9.0 360 1.1025 -0.0291 1.1025 1.0500
No log 9.05 362 1.0152 -0.0211 1.0152 1.0076
No log 9.1 364 0.8608 0.0065 0.8608 0.9278
No log 9.15 366 0.7887 0.0323 0.7887 0.8881
No log 9.2 368 0.8048 -0.0170 0.8048 0.8971
No log 9.25 370 0.8571 0.1239 0.8571 0.9258
No log 9.3 372 1.0086 -0.0892 1.0086 1.0043
No log 9.35 374 1.0425 -0.0586 1.0425 1.0210
No log 9.4 376 0.9674 -0.0885 0.9674 0.9836
No log 9.45 378 0.9273 -0.0076 0.9273 0.9630
No log 9.5 380 0.8926 -0.0262 0.8926 0.9448
No log 9.55 382 0.9879 -0.0526 0.9879 0.9940
No log 9.6 384 1.1174 -0.0320 1.1174 1.0571
No log 9.65 386 1.1052 -0.0320 1.1052 1.0513
No log 9.7 388 0.9658 -0.1221 0.9658 0.9827
No log 9.75 390 0.8575 -0.0870 0.8575 0.9260
No log 9.8 392 0.8695 -0.0661 0.8695 0.9324
No log 9.85 394 0.8474 -0.1268 0.8474 0.9206
No log 9.9 396 0.8693 0.0118 0.8693 0.9324
No log 9.95 398 1.0278 0.0404 1.0278 1.0138
No log 10.0 400 1.0613 0.0046 1.0613 1.0302
No log 10.05 402 0.9209 -0.0837 0.9209 0.9597
No log 10.1 404 0.8518 0.0028 0.8518 0.9229
No log 10.15 406 0.8689 0.0151 0.8689 0.9322
No log 10.2 408 0.8612 0.0441 0.8612 0.9280
No log 10.25 410 0.8805 0.0196 0.8805 0.9384
No log 10.3 412 0.9635 -0.1219 0.9635 0.9816
No log 10.35 414 1.0272 -0.0513 1.0272 1.0135
No log 10.4 416 0.9633 -0.1217 0.9633 0.9815
No log 10.45 418 0.8620 0.0393 0.8620 0.9284
No log 10.5 420 0.8430 0.0930 0.8430 0.9182
No log 10.55 422 0.8388 0.0700 0.8388 0.9159
No log 10.6 424 0.8456 0.0586 0.8456 0.9196
No log 10.65 426 0.8527 0.0091 0.8527 0.9234
No log 10.7 428 0.8836 0.0304 0.8836 0.9400
No log 10.75 430 0.8878 0.0676 0.8878 0.9422
No log 10.8 432 0.8878 0.0642 0.8878 0.9423
No log 10.85 434 0.8606 0.0642 0.8606 0.9277
No log 10.9 436 0.7905 0.1529 0.7905 0.8891
No log 10.95 438 0.8069 -0.0061 0.8069 0.8983
No log 11.0 440 0.8349 -0.0037 0.8349 0.9138
No log 11.05 442 0.8353 0.0879 0.8353 0.9139
No log 11.1 444 0.9289 -0.0076 0.9289 0.9638
No log 11.15 446 1.0491 -0.0236 1.0491 1.0243
No log 11.2 448 1.0305 -0.0200 1.0305 1.0151
No log 11.25 450 0.9091 -0.0097 0.9091 0.9534
No log 11.3 452 0.8672 0.0424 0.8672 0.9312
No log 11.35 454 0.8392 0.0690 0.8392 0.9161
No log 11.4 456 0.8383 0.0323 0.8383 0.9156
No log 11.45 458 0.8447 0.0289 0.8447 0.9191
No log 11.5 460 0.8464 0.0146 0.8464 0.9200
No log 11.55 462 0.8248 -0.0274 0.8248 0.9082
No log 11.6 464 0.8170 0.0999 0.8170 0.9039
No log 11.65 466 0.7918 -0.1106 0.7918 0.8898
No log 11.7 468 0.7661 -0.0541 0.7661 0.8753
No log 11.75 470 0.7695 -0.0541 0.7695 0.8772
No log 11.8 472 0.7919 -0.0628 0.7919 0.8899
No log 11.85 474 0.9172 -0.0031 0.9172 0.9577
No log 11.9 476 1.0509 0.0026 1.0509 1.0252
No log 11.95 478 1.0307 0.0086 1.0307 1.0152
No log 12.0 480 0.9044 0.0407 0.9044 0.9510
No log 12.05 482 0.8653 0.0139 0.8653 0.9302
No log 12.1 484 0.9013 0.0007 0.9013 0.9494
No log 12.15 486 0.8741 -0.0037 0.8741 0.9349
No log 12.2 488 0.8161 0.0893 0.8161 0.9034
No log 12.25 490 0.8368 0.0146 0.8368 0.9148
No log 12.3 492 0.9056 -0.0486 0.9056 0.9516
No log 12.35 494 0.8654 0.0424 0.8654 0.9303
No log 12.4 496 0.8424 0.0913 0.8424 0.9178
No log 12.45 498 0.7820 -0.0170 0.7820 0.8843
0.3498 12.5 500 0.7561 -0.0118 0.7561 0.8695
0.3498 12.55 502 0.7464 -0.0062 0.7464 0.8640
0.3498 12.6 504 0.7905 -0.0406 0.7905 0.8891
0.3498 12.65 506 0.8108 -0.0385 0.8108 0.9005
0.3498 12.7 508 0.7808 -0.0406 0.7808 0.8836
0.3498 12.75 510 0.7581 0.0414 0.7581 0.8707
0.3498 12.8 512 0.9350 0.1193 0.9350 0.9669
0.3498 12.85 514 1.1432 -0.0586 1.1432 1.0692
0.3498 12.9 516 1.0655 -0.0539 1.0655 1.0323
0.3498 12.95 518 0.8714 -0.0283 0.8714 0.9335
0.3498 13.0 520 0.8390 0.0145 0.8390 0.9160
0.3498 13.05 522 0.8596 0.0123 0.8596 0.9271
0.3498 13.1 524 0.8658 -0.0711 0.8658 0.9305
0.3498 13.15 526 0.8738 -0.2236 0.8738 0.9347
0.3498 13.2 528 0.9604 0.0250 0.9604 0.9800
0.3498 13.25 530 1.0659 -0.0597 1.0659 1.0324
0.3498 13.3 532 1.0364 -0.0228 1.0364 1.0180
0.3498 13.35 534 0.9549 0.0618 0.9549 0.9772
0.3498 13.4 536 0.8319 -0.1126 0.8319 0.9121
0.3498 13.45 538 0.8076 0.0926 0.8076 0.8986
0.3498 13.5 540 0.8095 -0.0506 0.8095 0.8997
0.3498 13.55 542 0.8087 -0.1610 0.8087 0.8993
0.3498 13.6 544 0.8168 -0.1580 0.8168 0.9038
0.3498 13.65 546 0.8663 0.0118 0.8663 0.9308
0.3498 13.7 548 0.9022 0.0040 0.9022 0.9498
0.3498 13.75 550 0.9149 0.0040 0.9149 0.9565
0.3498 13.8 552 0.8510 0.0118 0.8510 0.9225
0.3498 13.85 554 0.7843 0.0922 0.7843 0.8856
0.3498 13.9 556 0.7707 0.0978 0.7707 0.8779
0.3498 13.95 558 0.7391 0.0414 0.7391 0.8597
0.3498 14.0 560 0.7452 0.0338 0.7452 0.8633
0.3498 14.05 562 0.7955 0.0611 0.7955 0.8919
0.3498 14.1 564 0.8945 0.0407 0.8945 0.9458
0.3498 14.15 566 0.9827 -0.0118 0.9827 0.9913
0.3498 14.2 568 0.9865 -0.1230 0.9865 0.9932
0.3498 14.25 570 0.9572 -0.2310 0.9572 0.9784
0.3498 14.3 572 0.9056 -0.0743 0.9056 0.9516
0.3498 14.35 574 0.8606 -0.0170 0.8606 0.9277
0.3498 14.4 576 0.8375 -0.1860 0.8375 0.9151

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
37
Safetensors
Model size
135M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k16_task3_organization

Finetuned
(3994)
this model