ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k18_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9906
  • Qwk: -0.1257
  • Mse: 0.9906
  • Rmse: 0.9953

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0444 2 3.6132 -0.0252 3.6132 1.9008
No log 0.0889 4 2.3154 0.0431 2.3154 1.5217
No log 0.1333 6 2.3802 0.0304 2.3802 1.5428
No log 0.1778 8 1.5836 -0.0015 1.5836 1.2584
No log 0.2222 10 1.0974 0.0423 1.0974 1.0476
No log 0.2667 12 1.1507 -0.0178 1.1507 1.0727
No log 0.3111 14 1.0164 -0.0345 1.0164 1.0081
No log 0.3556 16 0.7562 -0.0188 0.7562 0.8696
No log 0.4 18 0.7388 -0.1223 0.7388 0.8596
No log 0.4444 20 0.7972 -0.0188 0.7972 0.8929
No log 0.4889 22 1.0593 0.0446 1.0593 1.0292
No log 0.5333 24 1.5707 0.0 1.5707 1.2533
No log 0.5778 26 1.7024 0.0 1.7024 1.3048
No log 0.6222 28 1.7264 0.0 1.7264 1.3139
No log 0.6667 30 1.5152 0.0 1.5152 1.2310
No log 0.7111 32 1.3409 0.0 1.3409 1.1580
No log 0.7556 34 1.3414 0.0 1.3414 1.1582
No log 0.8 36 1.3420 0.0 1.3420 1.1585
No log 0.8444 38 1.3156 0.0 1.3156 1.1470
No log 0.8889 40 1.1599 0.0 1.1599 1.0770
No log 0.9333 42 0.9175 0.0006 0.9175 0.9578
No log 0.9778 44 0.8567 -0.0861 0.8567 0.9256
No log 1.0222 46 0.8364 -0.0766 0.8364 0.9145
No log 1.0667 48 0.9196 0.0404 0.9196 0.9590
No log 1.1111 50 1.2287 -0.1023 1.2287 1.1084
No log 1.1556 52 1.4689 -0.1023 1.4689 1.2120
No log 1.2 54 1.4188 -0.1292 1.4188 1.1911
No log 1.2444 56 1.3135 -0.0372 1.3135 1.1461
No log 1.2889 58 1.3043 -0.0992 1.3043 1.1420
No log 1.3333 60 1.3114 -0.1019 1.3114 1.1452
No log 1.3778 62 1.1451 -0.0500 1.1451 1.0701
No log 1.4222 64 1.0647 0.0 1.0647 1.0318
No log 1.4667 66 0.9364 0.0156 0.9364 0.9677
No log 1.5111 68 0.8601 0.0486 0.8601 0.9274
No log 1.5556 70 0.8935 -0.0013 0.8935 0.9453
No log 1.6 72 0.9734 -0.0385 0.9734 0.9866
No log 1.6444 74 1.0560 -0.0704 1.0560 1.0276
No log 1.6889 76 1.2599 -0.0234 1.2599 1.1225
No log 1.7333 78 1.4230 -0.0234 1.4230 1.1929
No log 1.7778 80 1.3554 0.0 1.3554 1.1642
No log 1.8222 82 1.3528 0.0 1.3528 1.1631
No log 1.8667 84 1.2125 0.0 1.2125 1.1011
No log 1.9111 86 1.1367 0.0 1.1367 1.0661
No log 1.9556 88 1.1698 0.0016 1.1698 1.0816
No log 2.0 90 1.2174 0.0298 1.2174 1.1034
No log 2.0444 92 1.1943 0.0065 1.1943 1.0928
No log 2.0889 94 1.3236 0.0298 1.3236 1.1505
No log 2.1333 96 1.4891 0.0279 1.4891 1.2203
No log 2.1778 98 1.4976 0.0279 1.4976 1.2238
No log 2.2222 100 1.2778 0.0298 1.2778 1.1304
No log 2.2667 102 1.1577 0.0610 1.1577 1.0760
No log 2.3111 104 1.1609 0.0937 1.1609 1.0774
No log 2.3556 106 1.3082 0.0610 1.3082 1.1438
No log 2.4 108 1.1645 -0.0101 1.1645 1.0791
No log 2.4444 110 0.9901 -0.0200 0.9901 0.9951
No log 2.4889 112 1.0211 0.0089 1.0211 1.0105
No log 2.5333 114 1.1197 -0.0638 1.1197 1.0582
No log 2.5778 116 1.2421 -0.0067 1.2421 1.1145
No log 2.6222 118 1.3572 -0.0399 1.3572 1.1650
No log 2.6667 120 1.1483 -0.0331 1.1483 1.0716
No log 2.7111 122 0.9710 0.0377 0.9710 0.9854
No log 2.7556 124 1.0237 0.0068 1.0237 1.0118
No log 2.8 126 1.2900 -0.0610 1.2900 1.1358
No log 2.8444 128 1.4083 -0.1211 1.4083 1.1867
No log 2.8889 130 1.1138 -0.0526 1.1138 1.0554
No log 2.9333 132 0.9133 -0.1747 0.9133 0.9557
No log 2.9778 134 0.9660 -0.0766 0.9660 0.9828
No log 3.0222 136 1.3835 -0.1207 1.3835 1.1762
No log 3.0667 138 1.6503 -0.0367 1.6503 1.2846
No log 3.1111 140 1.3662 -0.1205 1.3662 1.1688
No log 3.1556 142 1.0934 -0.0799 1.0934 1.0457
No log 3.2 144 0.9984 -0.1394 0.9984 0.9992
No log 3.2444 146 0.9683 -0.1994 0.9683 0.9840
No log 3.2889 148 1.0200 0.0831 1.0200 1.0099
No log 3.3333 150 1.3870 -0.0620 1.3870 1.1777
No log 3.3778 152 1.4992 -0.0655 1.4992 1.2244
No log 3.4222 154 1.1368 -0.0855 1.1368 1.0662
No log 3.4667 156 0.8828 -0.0287 0.8828 0.9396
No log 3.5111 158 0.8365 -0.2278 0.8365 0.9146
No log 3.5556 160 0.8967 0.0129 0.8967 0.9469
No log 3.6 162 1.0099 0.0260 1.0099 1.0049
No log 3.6444 164 1.0427 -0.0138 1.0427 1.0211
No log 3.6889 166 1.0854 -0.0409 1.0854 1.0418
No log 3.7333 168 1.2383 -0.1148 1.2383 1.1128
No log 3.7778 170 1.2352 -0.1148 1.2352 1.1114
No log 3.8222 172 1.0719 -0.1605 1.0719 1.0353
No log 3.8667 174 1.0411 -0.1255 1.0411 1.0203
No log 3.9111 176 1.1836 -0.1184 1.1836 1.0879
No log 3.9556 178 1.4637 -0.0319 1.4637 1.2098
No log 4.0 180 1.2762 -0.1196 1.2762 1.1297
No log 4.0444 182 0.9826 -0.0390 0.9826 0.9913
No log 4.0889 184 0.9664 -0.0801 0.9664 0.9830
No log 4.1333 186 1.0640 -0.0459 1.0640 1.0315
No log 4.1778 188 0.9963 -0.1259 0.9963 0.9982
No log 4.2222 190 0.9392 -0.1668 0.9392 0.9691
No log 4.2667 192 0.9225 -0.1690 0.9225 0.9605
No log 4.3111 194 0.9080 -0.1168 0.9080 0.9529
No log 4.3556 196 0.9827 -0.2511 0.9827 0.9913
No log 4.4 198 1.0800 -0.0870 1.0800 1.0392
No log 4.4444 200 1.0205 -0.1688 1.0205 1.0102
No log 4.4889 202 0.9202 -0.2219 0.9202 0.9593
No log 4.5333 204 0.8946 -0.1675 0.8946 0.9458
No log 4.5778 206 0.9463 -0.1730 0.9463 0.9728
No log 4.6222 208 1.1469 0.0537 1.1469 1.0709
No log 4.6667 210 1.3076 -0.0319 1.3076 1.1435
No log 4.7111 212 1.0951 0.0152 1.0951 1.0465
No log 4.7556 214 0.8965 -0.0228 0.8965 0.9468
No log 4.8 216 0.9327 -0.1588 0.9327 0.9657
No log 4.8444 218 0.9538 -0.1163 0.9538 0.9766
No log 4.8889 220 0.9817 -0.1140 0.9817 0.9908
No log 4.9333 222 1.1630 0.0508 1.1630 1.0784
No log 4.9778 224 1.3508 -0.0647 1.3508 1.1622
No log 5.0222 226 1.3493 -0.0647 1.3493 1.1616
No log 5.0667 228 1.1394 -0.0211 1.1394 1.0674
No log 5.1111 230 0.9145 -0.0686 0.9145 0.9563
No log 5.1556 232 0.8803 -0.0062 0.8803 0.9382
No log 5.2 234 0.8531 -0.0091 0.8531 0.9236
No log 5.2444 236 0.8978 -0.1701 0.8978 0.9475
No log 5.2889 238 0.9696 -0.1261 0.9696 0.9847
No log 5.3333 240 0.9939 -0.1676 0.9939 0.9970
No log 5.3778 242 0.9983 -0.1501 0.9983 0.9991
No log 5.4222 244 1.0132 -0.1219 1.0132 1.0066
No log 5.4667 246 1.0260 -0.2346 1.0260 1.0129
No log 5.5111 248 0.9430 -0.1939 0.9430 0.9711
No log 5.5556 250 0.9127 -0.1172 0.9127 0.9554
No log 5.6 252 0.9655 -0.0790 0.9655 0.9826
No log 5.6444 254 1.0244 -0.0852 1.0244 1.0121
No log 5.6889 256 0.9986 -0.1257 0.9986 0.9993
No log 5.7333 258 0.9975 -0.1257 0.9975 0.9987
No log 5.7778 260 0.9204 -0.0753 0.9204 0.9594
No log 5.8222 262 0.9295 -0.1233 0.9295 0.9641
No log 5.8667 264 0.9913 -0.1180 0.9913 0.9956
No log 5.9111 266 1.0763 -0.1623 1.0763 1.0374
No log 5.9556 268 1.0454 -0.1708 1.0454 1.0224
No log 6.0 270 1.0701 -0.0425 1.0701 1.0344
No log 6.0444 272 0.9978 0.0017 0.9978 0.9989
No log 6.0889 274 0.9105 -0.0766 0.9105 0.9542
No log 6.1333 276 0.9083 -0.0252 0.9083 0.9531
No log 6.1778 278 1.1073 -0.0441 1.1073 1.0523
No log 6.2222 280 1.1891 -0.0847 1.1891 1.0905
No log 6.2667 282 1.1486 -0.0496 1.1486 1.0717
No log 6.3111 284 0.9643 0.0095 0.9643 0.9820
No log 6.3556 286 0.8858 -0.0252 0.8858 0.9412
No log 6.4 288 0.9058 -0.0309 0.9058 0.9518
No log 6.4444 290 1.0925 -0.0877 1.0925 1.0452
No log 6.4889 292 1.1998 -0.0586 1.1998 1.0953
No log 6.5333 294 1.0940 -0.0532 1.0940 1.0459
No log 6.5778 296 0.8902 -0.0309 0.8902 0.9435
No log 6.6222 298 0.8425 -0.0766 0.8425 0.9179
No log 6.6667 300 0.9056 -0.0309 0.9056 0.9516
No log 6.7111 302 1.0663 -0.0854 1.0663 1.0326
No log 6.7556 304 1.0535 -0.0828 1.0535 1.0264
No log 6.8 306 0.9602 0.0071 0.9602 0.9799
No log 6.8444 308 0.8448 -0.0766 0.8448 0.9191
No log 6.8889 310 0.8589 -0.0766 0.8589 0.9268
No log 6.9333 312 1.0113 -0.0442 1.0113 1.0056
No log 6.9778 314 1.2976 -0.0291 1.2976 1.1391
No log 7.0222 316 1.2749 -0.0586 1.2749 1.1291
No log 7.0667 318 1.0615 -0.0008 1.0615 1.0303
No log 7.1111 320 0.9441 -0.0495 0.9441 0.9717
No log 7.1556 322 0.9605 -0.1032 0.9605 0.9800
No log 7.2 324 0.9526 -0.0425 0.9526 0.9760
No log 7.2444 326 1.0046 0.0837 1.0046 1.0023
No log 7.2889 328 1.1180 -0.0877 1.1180 1.0574
No log 7.3333 330 1.0447 0.0287 1.0447 1.0221
No log 7.3778 332 0.8653 -0.0331 0.8653 0.9302
No log 7.4222 334 0.8072 -0.0264 0.8072 0.8984
No log 7.4667 336 0.8090 -0.0264 0.8090 0.8994
No log 7.5111 338 0.7826 -0.0711 0.7826 0.8847
No log 7.5556 340 0.7864 -0.0215 0.7864 0.8868
No log 7.6 342 0.9044 0.0099 0.9044 0.9510
No log 7.6444 344 1.0060 -0.0828 1.0060 1.0030
No log 7.6889 346 0.9865 -0.1212 0.9865 0.9932
No log 7.7333 348 0.9497 -0.0842 0.9497 0.9745
No log 7.7778 350 0.9052 0.0043 0.9052 0.9514
No log 7.8222 352 0.8267 0.0680 0.8267 0.9092
No log 7.8667 354 0.8158 0.0318 0.8158 0.9032
No log 7.9111 356 0.8703 0.0680 0.8703 0.9329
No log 7.9556 358 0.9129 0.0639 0.9129 0.9555
No log 8.0 360 0.9463 0.0152 0.9463 0.9728
No log 8.0444 362 1.0539 0.0304 1.0539 1.0266
No log 8.0889 364 1.0869 0.0277 1.0869 1.0426
No log 8.1333 366 1.1067 -0.0175 1.1067 1.0520
No log 8.1778 368 0.9972 -0.0474 0.9972 0.9986
No log 8.2222 370 0.8883 -0.0731 0.8883 0.9425
No log 8.2667 372 0.8701 0.0236 0.8701 0.9328
No log 8.3111 374 0.9005 -0.0656 0.9005 0.9490
No log 8.3556 376 0.9480 -0.0735 0.9480 0.9737
No log 8.4 378 0.9289 -0.0643 0.9289 0.9638
No log 8.4444 380 0.8984 0.0247 0.8984 0.9479
No log 8.4889 382 0.8739 0.0247 0.8739 0.9348
No log 8.5333 384 0.8852 -0.0766 0.8852 0.9409
No log 8.5778 386 0.9189 -0.1253 0.9189 0.9586
No log 8.6222 388 0.9035 -0.1253 0.9035 0.9505
No log 8.6667 390 0.9468 -0.0833 0.9468 0.9730
No log 8.7111 392 0.9282 -0.1253 0.9282 0.9634
No log 8.7556 394 0.8705 -0.0264 0.8705 0.9330
No log 8.8 396 0.8674 -0.0240 0.8674 0.9314
No log 8.8444 398 0.8857 -0.0264 0.8857 0.9411
No log 8.8889 400 0.9561 -0.1197 0.9561 0.9778
No log 8.9333 402 1.0582 -0.0097 1.0582 1.0287
No log 8.9778 404 1.0247 0.0333 1.0247 1.0123
No log 9.0222 406 0.9106 -0.0274 0.9106 0.9542
No log 9.0667 408 0.8893 -0.0252 0.8893 0.9430
No log 9.1111 410 0.8765 -0.0228 0.8765 0.9362
No log 9.1556 412 0.8625 0.0318 0.8625 0.9287
No log 9.2 414 0.8862 -0.0228 0.8862 0.9414
No log 9.2444 416 0.9215 -0.0743 0.9215 0.9600
No log 9.2889 418 0.9200 -0.1249 0.9200 0.9592
No log 9.3333 420 0.9029 -0.1249 0.9029 0.9502
No log 9.3778 422 0.8694 -0.0215 0.8694 0.9324
No log 9.4222 424 0.8952 -0.1461 0.8952 0.9461
No log 9.4667 426 0.9375 -0.0992 0.9375 0.9683
No log 9.5111 428 0.9991 -0.1246 0.9991 0.9995
No log 9.5556 430 1.0376 -0.1251 1.0376 1.0186
No log 9.6 432 1.1400 -0.0870 1.1400 1.0677
No log 9.6444 434 1.1570 -0.0504 1.1570 1.0756
No log 9.6889 436 1.0813 -0.0490 1.0813 1.0399
No log 9.7333 438 0.9370 -0.0823 0.9370 0.9680
No log 9.7778 440 0.8485 -0.0766 0.8485 0.9211
No log 9.8222 442 0.8304 -0.0711 0.8304 0.9112
No log 9.8667 444 0.8268 -0.0711 0.8268 0.9093
No log 9.9111 446 0.8516 -0.0331 0.8516 0.9228
No log 9.9556 448 0.9404 -0.0371 0.9404 0.9698
No log 10.0 450 0.9825 -0.0833 0.9825 0.9912
No log 10.0444 452 0.9825 -0.1633 0.9825 0.9912
No log 10.0889 454 0.9993 -0.1633 0.9993 0.9996
No log 10.1333 456 1.0058 -0.2059 1.0058 1.0029
No log 10.1778 458 0.9460 -0.1650 0.9460 0.9726
No log 10.2222 460 0.9402 -0.1650 0.9402 0.9696
No log 10.2667 462 0.9601 -0.2100 0.9601 0.9798
No log 10.3111 464 1.0057 -0.1638 1.0057 1.0028
No log 10.3556 466 1.0770 -0.1769 1.0770 1.0378
No log 10.4 468 1.1498 -0.0653 1.1498 1.0723
No log 10.4444 470 1.1355 -0.1002 1.1355 1.0656
No log 10.4889 472 1.0722 -0.0712 1.0722 1.0355
No log 10.5333 474 0.9790 -0.1935 0.9790 0.9894
No log 10.5778 476 0.9286 -0.1638 0.9286 0.9636
No log 10.6222 478 0.8871 -0.1722 0.8871 0.9418
No log 10.6667 480 0.8892 -0.1715 0.8892 0.9430
No log 10.7111 482 0.9641 -0.1257 0.9641 0.9819
No log 10.7556 484 1.0551 -0.1589 1.0551 1.0272
No log 10.8 486 1.1747 -0.0885 1.1747 1.0838
No log 10.8444 488 1.1715 -0.0870 1.1715 1.0823
No log 10.8889 490 1.0625 -0.1446 1.0625 1.0308
No log 10.9333 492 1.0564 -0.1281 1.0564 1.0278
No log 10.9778 494 1.0982 -0.1337 1.0982 1.0479
No log 11.0222 496 1.1883 -0.0742 1.1883 1.0901
No log 11.0667 498 1.1545 -0.0722 1.1545 1.0745
0.3414 11.1111 500 1.0871 -0.1060 1.0871 1.0427
0.3414 11.1556 502 1.0429 -0.1444 1.0429 1.0212
0.3414 11.2 504 1.0018 -0.0778 1.0018 1.0009
0.3414 11.2444 506 0.9621 -0.1253 0.9621 0.9809
0.3414 11.2889 508 0.9596 -0.1263 0.9596 0.9796
0.3414 11.3333 510 0.9553 -0.1253 0.9553 0.9774
0.3414 11.3778 512 0.9492 -0.0801 0.9492 0.9743
0.3414 11.4222 514 0.9436 -0.1512 0.9436 0.9714
0.3414 11.4667 516 0.9508 -0.1884 0.9508 0.9751
0.3414 11.5111 518 0.9930 -0.0711 0.9930 0.9965
0.3414 11.5556 520 1.1320 -0.0118 1.1320 1.0640
0.3414 11.6 522 1.1522 -0.0138 1.1522 1.0734
0.3414 11.6444 524 1.0492 -0.0355 1.0492 1.0243
0.3414 11.6889 526 0.9494 -0.1585 0.9494 0.9744
0.3414 11.7333 528 0.9122 -0.1599 0.9122 0.9551
0.3414 11.7778 530 0.9110 -0.0739 0.9110 0.9545
0.3414 11.8222 532 0.9879 -0.1701 0.9879 0.9939
0.3414 11.8667 534 1.0492 -0.0861 1.0492 1.0243
0.3414 11.9111 536 1.0034 -0.0842 1.0034 1.0017
0.3414 11.9556 538 1.0052 -0.1257 1.0052 1.0026
0.3414 12.0 540 1.0118 -0.1259 1.0118 1.0059
0.3414 12.0444 542 1.0147 -0.1259 1.0147 1.0073
0.3414 12.0889 544 1.0113 -0.1259 1.0113 1.0056
0.3414 12.1333 546 0.9906 -0.1257 0.9906 0.9953

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
8
Safetensors
Model size
135M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k18_task3_organization

Finetuned
(3994)
this model