ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k13_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8810
  • Qwk: -0.1033
  • Mse: 0.8810
  • Rmse: 0.9386

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0606 2 3.8797 0.0017 3.8797 1.9697
No log 0.1212 4 2.2173 0.0672 2.2173 1.4891
No log 0.1818 6 1.3323 0.0 1.3323 1.1543
No log 0.2424 8 1.8831 0.0425 1.8831 1.3723
No log 0.3030 10 1.7085 0.0 1.7085 1.3071
No log 0.3636 12 1.0686 -0.0457 1.0686 1.0337
No log 0.4242 14 0.7419 -0.1227 0.7419 0.8613
No log 0.4848 16 0.7259 -0.0069 0.7259 0.8520
No log 0.5455 18 0.7743 -0.1230 0.7743 0.8800
No log 0.6061 20 0.9668 0.0423 0.9668 0.9833
No log 0.6667 22 1.1082 -0.0247 1.1082 1.0527
No log 0.7273 24 1.2468 0.0 1.2468 1.1166
No log 0.7879 26 1.4488 0.0 1.4488 1.2036
No log 0.8485 28 1.4140 0.0 1.4140 1.1891
No log 0.9091 30 1.2783 0.0 1.2783 1.1306
No log 0.9697 32 1.0355 -0.0490 1.0355 1.0176
No log 1.0303 34 1.1671 -0.0490 1.1671 1.0803
No log 1.0909 36 1.2888 -0.0247 1.2888 1.1352
No log 1.1515 38 1.0918 -0.0728 1.0918 1.0449
No log 1.2121 40 1.0713 -0.0728 1.0713 1.0351
No log 1.2727 42 1.2809 -0.0751 1.2809 1.1318
No log 1.3333 44 1.4192 -0.0234 1.4192 1.1913
No log 1.3939 46 1.2224 -0.0728 1.2224 1.1056
No log 1.4545 48 0.9249 -0.0425 0.9249 0.9617
No log 1.5152 50 0.9596 -0.1579 0.9596 0.9796
No log 1.5758 52 0.9341 -0.1586 0.9341 0.9665
No log 1.6364 54 1.0390 -0.0948 1.0390 1.0193
No log 1.6970 56 1.3858 -0.0234 1.3858 1.1772
No log 1.7576 58 1.6151 0.0 1.6151 1.2709
No log 1.8182 60 1.3975 -0.0247 1.3975 1.1822
No log 1.8788 62 0.9126 -0.0456 0.9126 0.9553
No log 1.9394 64 0.7597 -0.0541 0.7597 0.8716
No log 2.0 66 0.7675 -0.0560 0.7675 0.8761
No log 2.0606 68 1.0588 0.0006 1.0588 1.0290
No log 2.1212 70 1.2690 -0.0692 1.2690 1.1265
No log 2.1818 72 1.1749 -0.0677 1.1749 1.0839
No log 2.2424 74 0.9064 -0.0054 0.9064 0.9520
No log 2.3030 76 0.8812 0.0287 0.8812 0.9387
No log 2.3636 78 0.9852 -0.0218 0.9852 0.9926
No log 2.4242 80 1.0274 -0.0236 1.0274 1.0136
No log 2.4848 82 0.8732 -0.0731 0.8732 0.9344
No log 2.5455 84 0.8264 -0.1463 0.8264 0.9091
No log 2.6061 86 0.8607 -0.0984 0.8607 0.9277
No log 2.6667 88 1.0071 -0.1263 1.0071 1.0036
No log 2.7273 90 0.9248 -0.1795 0.9248 0.9617
No log 2.7879 92 0.9311 -0.1466 0.9311 0.9649
No log 2.8485 94 1.0183 -0.1509 1.0183 1.0091
No log 2.9091 96 0.9124 -0.2036 0.9124 0.9552
No log 2.9697 98 0.8933 -0.0939 0.8933 0.9451
No log 3.0303 100 0.8449 -0.2051 0.8449 0.9192
No log 3.0909 102 1.0998 -0.0902 1.0998 1.0487
No log 3.1515 104 0.9966 -0.1253 0.9966 0.9983
No log 3.2121 106 0.8358 -0.1001 0.8358 0.9142
No log 3.2727 108 0.8829 -0.1399 0.8829 0.9396
No log 3.3333 110 0.8588 -0.1001 0.8588 0.9267
No log 3.3939 112 0.8501 -0.1106 0.8501 0.9220
No log 3.4545 114 1.4037 -0.0114 1.4037 1.1848
No log 3.5152 116 1.5863 -0.0445 1.5863 1.2595
No log 3.5758 118 1.0188 -0.1166 1.0188 1.0094
No log 3.6364 120 0.8649 -0.1033 0.8649 0.9300
No log 3.6970 122 0.8876 -0.2008 0.8876 0.9421
No log 3.7576 124 1.1339 -0.0500 1.1339 1.0648
No log 3.8182 126 1.2490 -0.0961 1.2490 1.1176
No log 3.8788 128 0.9312 -0.0336 0.9312 0.9650
No log 3.9394 130 0.9130 -0.1775 0.9130 0.9555
No log 4.0 132 0.9572 -0.0685 0.9572 0.9784
No log 4.0606 134 0.8055 -0.0902 0.8055 0.8975
No log 4.1212 136 1.0941 0.0353 1.0941 1.0460
No log 4.1818 138 1.4682 -0.0174 1.4682 1.2117
No log 4.2424 140 0.9630 0.0182 0.9630 0.9813
No log 4.3030 142 0.8254 -0.1331 0.8254 0.9085
No log 4.3636 144 0.9154 -0.0121 0.9154 0.9568
No log 4.4242 146 0.8256 -0.0449 0.8256 0.9086
No log 4.4848 148 1.0088 -0.1166 1.0088 1.0044
No log 4.5455 150 1.1728 0.0391 1.1728 1.0830
No log 4.6061 152 0.9447 -0.1162 0.9447 0.9720
No log 4.6667 154 0.7754 -0.1033 0.7754 0.8806
No log 4.7273 156 0.7767 -0.0449 0.7767 0.8813
No log 4.7879 158 0.7541 -0.0541 0.7541 0.8684
No log 4.8485 160 0.7398 0.0807 0.7398 0.8601
No log 4.9091 162 0.7776 0.0670 0.7776 0.8818
No log 4.9697 164 0.8339 -0.0833 0.8339 0.9132
No log 5.0303 166 0.8139 0.0449 0.8139 0.9021
No log 5.0909 168 0.8452 -0.0831 0.8452 0.9194
No log 5.1515 170 0.8229 -0.0138 0.8229 0.9071
No log 5.2121 172 0.7756 -0.1470 0.7756 0.8807
No log 5.2727 174 0.7408 -0.0131 0.7408 0.8607
No log 5.3333 176 0.7179 0.0460 0.7179 0.8473
No log 5.3939 178 0.7160 -0.0069 0.7160 0.8462
No log 5.4545 180 0.7588 -0.1018 0.7588 0.8711
No log 5.5152 182 0.8495 -0.0692 0.8495 0.9217
No log 5.5758 184 1.0640 0.0541 1.0640 1.0315
No log 5.6364 186 1.0626 0.0196 1.0626 1.0308
No log 5.6970 188 1.0551 -0.0099 1.0551 1.0272
No log 5.7576 190 0.9119 0.0406 0.9119 0.9549
No log 5.8182 192 0.7668 -0.1018 0.7668 0.8757
No log 5.8788 194 0.7590 -0.0499 0.7590 0.8712
No log 5.9394 196 0.7459 0.0033 0.7459 0.8637
No log 6.0 198 0.7879 -0.1001 0.7879 0.8877
No log 6.0606 200 0.9832 -0.0892 0.9832 0.9915
No log 6.1212 202 1.1024 0.0862 1.1024 1.0499
No log 6.1818 204 0.8525 0.1466 0.8525 0.9233
No log 6.2424 206 0.8620 0.0456 0.8620 0.9284
No log 6.3030 208 0.9246 -0.0490 0.9246 0.9616
No log 6.3636 210 1.0027 -0.0122 1.0027 1.0013
No log 6.4242 212 0.7898 0.0814 0.7898 0.8887
No log 6.4848 214 0.8631 -0.1833 0.8631 0.9290
No log 6.5455 216 0.8713 -0.0958 0.8713 0.9334
No log 6.6061 218 0.8871 -0.1459 0.8871 0.9419
No log 6.6667 220 1.1040 -0.0504 1.1040 1.0507
No log 6.7273 222 1.0428 -0.0056 1.0428 1.0212
No log 6.7879 224 0.7942 -0.0065 0.7942 0.8912
No log 6.8485 226 0.8016 -0.0499 0.8016 0.8953
No log 6.9091 228 0.8619 0.0181 0.8619 0.9284
No log 6.9697 230 0.7987 -0.0551 0.7987 0.8937
No log 7.0303 232 0.7493 -0.0551 0.7493 0.8656
No log 7.0909 234 0.8071 0.0129 0.8071 0.8984
No log 7.1515 236 0.9875 -0.0852 0.9875 0.9937
No log 7.2121 238 0.8318 -0.0351 0.8318 0.9120
No log 7.2727 240 0.8106 -0.1001 0.8106 0.9003
No log 7.3333 242 0.8659 -0.0406 0.8659 0.9306
No log 7.3939 244 0.8082 -0.1001 0.8082 0.8990
No log 7.4545 246 0.7896 -0.1163 0.7896 0.8886
No log 7.5152 248 1.0568 -0.0916 1.0568 1.0280
No log 7.5758 250 1.0659 -0.0936 1.0659 1.0324
No log 7.6364 252 0.8126 -0.0351 0.8126 0.9014
No log 7.6970 254 0.7883 -0.0499 0.7883 0.8879
No log 7.7576 256 0.9476 -0.1145 0.9476 0.9734
No log 7.8182 258 0.9892 -0.0187 0.9892 0.9946
No log 7.8788 260 0.8581 -0.0363 0.8581 0.9263
No log 7.9394 262 0.9386 -0.0788 0.9386 0.9688
No log 8.0 264 1.2834 -0.0658 1.2834 1.1329
No log 8.0606 266 1.2090 -0.0648 1.2090 1.0996
No log 8.1212 268 0.8880 -0.0408 0.8880 0.9424
No log 8.1818 270 0.7602 -0.0499 0.7602 0.8719
No log 8.2424 272 0.7572 -0.0499 0.7572 0.8701
No log 8.3030 274 0.7578 -0.0551 0.7578 0.8705
No log 8.3636 276 0.8076 -0.0331 0.8076 0.8987
No log 8.4242 278 0.7702 -0.0274 0.7702 0.8776
No log 8.4848 280 0.7722 0.0033 0.7722 0.8787
No log 8.5455 282 0.9438 -0.0377 0.9438 0.9715
No log 8.6061 284 0.9749 -0.0583 0.9749 0.9874
No log 8.6667 286 0.8582 -0.0786 0.8582 0.9264
No log 8.7273 288 0.8201 0.0236 0.8201 0.9056
No log 8.7879 290 0.8246 0.0236 0.8246 0.9081
No log 8.8485 292 0.7993 0.0236 0.7993 0.8940
No log 8.9091 294 0.8379 -0.1329 0.8379 0.9154
No log 8.9697 296 0.8588 -0.1263 0.8588 0.9267
No log 9.0303 298 0.8087 -0.1001 0.8087 0.8993
No log 9.0909 300 0.7546 -0.0499 0.7546 0.8687
No log 9.1515 302 0.7731 0.0714 0.7731 0.8793
No log 9.2121 304 0.7881 0.0159 0.7881 0.8878
No log 9.2727 306 0.7768 0.0814 0.7768 0.8814
No log 9.3333 308 0.7859 -0.0560 0.7859 0.8865
No log 9.3939 310 0.8287 -0.1268 0.8287 0.9103
No log 9.4545 312 0.8713 -0.0982 0.8713 0.9334
No log 9.5152 314 0.8315 -0.0334 0.8315 0.9118
No log 9.5758 316 0.8788 -0.0008 0.8788 0.9374
No log 9.6364 318 0.8616 -0.0408 0.8616 0.9282
No log 9.6970 320 0.8368 0.0749 0.8368 0.9147
No log 9.7576 322 0.8233 -0.1033 0.8233 0.9074
No log 9.8182 324 0.8242 0.0600 0.8242 0.9079
No log 9.8788 326 0.8057 0.0191 0.8057 0.8976
No log 9.9394 328 0.8293 -0.0390 0.8293 0.9107
No log 10.0 330 0.8720 -0.0008 0.8720 0.9338
No log 10.0606 332 0.8065 0.0723 0.8065 0.8981
No log 10.1212 334 0.8210 -0.0032 0.8210 0.9061
No log 10.1818 336 0.8498 -0.0892 0.8498 0.9219
No log 10.2424 338 0.9365 -0.0477 0.9365 0.9677
No log 10.3030 340 0.9513 -0.1298 0.9513 0.9753
No log 10.3636 342 0.8525 0.0116 0.8525 0.9233
No log 10.4242 344 0.8395 -0.0252 0.8395 0.9162
No log 10.4848 346 0.8430 0.0303 0.8430 0.9182
No log 10.5455 348 0.8428 0.0303 0.8428 0.9180
No log 10.6061 350 0.8522 -0.0449 0.8522 0.9231
No log 10.6667 352 0.8692 -0.0469 0.8692 0.9323
No log 10.7273 354 0.8917 -0.0614 0.8917 0.9443
No log 10.7879 356 0.9866 -0.0806 0.9866 0.9933
No log 10.8485 358 0.9795 -0.0806 0.9795 0.9897
No log 10.9091 360 0.8740 0.0574 0.8740 0.9349
No log 10.9697 362 0.8272 -0.0560 0.8272 0.9095
No log 11.0303 364 0.8183 0.0099 0.8183 0.9046
No log 11.0909 366 0.8223 0.0099 0.8223 0.9068
No log 11.1515 368 0.8324 -0.0578 0.8324 0.9124
No log 11.2121 370 0.8987 -0.1206 0.8987 0.9480
No log 11.2727 372 0.9144 -0.0711 0.9144 0.9563
No log 11.3333 374 0.9207 -0.1051 0.9207 0.9596
No log 11.3939 376 0.9159 -0.0738 0.9159 0.9570
No log 11.4545 378 0.9148 -0.1939 0.9148 0.9564
No log 11.5152 380 0.9381 -0.1509 0.9381 0.9686
No log 11.5758 382 0.9423 -0.1135 0.9423 0.9707
No log 11.6364 384 0.9397 -0.1568 0.9397 0.9694
No log 11.6970 386 0.9015 -0.2894 0.9015 0.9495
No log 11.7576 388 0.9428 -0.1529 0.9428 0.9710
No log 11.8182 390 0.9758 -0.1302 0.9758 0.9878
No log 11.8788 392 0.9276 -0.0907 0.9276 0.9631
No log 11.9394 394 0.9350 -0.0054 0.9350 0.9670
No log 12.0 396 0.9776 -0.0930 0.9776 0.9887
No log 12.0606 398 0.9103 -0.0878 0.9103 0.9541
No log 12.1212 400 0.8393 -0.1172 0.8393 0.9161
No log 12.1818 402 0.8741 -0.1201 0.8741 0.9350
No log 12.2424 404 0.9181 -0.0678 0.9181 0.9582
No log 12.3030 406 0.9043 -0.1263 0.9043 0.9510
No log 12.3636 408 0.8680 -0.0967 0.8680 0.9317
No log 12.4242 410 0.8984 0.0043 0.8984 0.9478
No log 12.4848 412 0.9303 -0.0474 0.9303 0.9645
No log 12.5455 414 0.8459 -0.0778 0.8459 0.9197
No log 12.6061 416 0.7993 0.0061 0.7993 0.8940
No log 12.6667 418 0.8095 -0.0798 0.8095 0.8997
No log 12.7273 420 0.7616 0.0524 0.7616 0.8727
No log 12.7879 422 0.7621 0.0714 0.7621 0.8730
No log 12.8485 424 0.7880 0.0512 0.7880 0.8877
No log 12.9091 426 0.8345 0.1291 0.8345 0.9135
No log 12.9697 428 0.8224 0.1141 0.8224 0.9069
No log 13.0303 430 0.8477 -0.0581 0.8477 0.9207
No log 13.0909 432 0.8562 -0.0543 0.8562 0.9253
No log 13.1515 434 0.8029 -0.1268 0.8029 0.8961
No log 13.2121 436 0.8073 -0.0849 0.8073 0.8985
No log 13.2727 438 0.7643 0.0973 0.7643 0.8743
No log 13.3333 440 0.7696 0.0874 0.7696 0.8773
No log 13.3939 442 0.8467 0.0285 0.8467 0.9202
No log 13.4545 444 0.9865 -0.0423 0.9865 0.9932
No log 13.5152 446 0.9653 0.0772 0.9653 0.9825
No log 13.5758 448 0.8518 -0.1333 0.8518 0.9229
No log 13.6364 450 0.8791 -0.0711 0.8791 0.9376
No log 13.6970 452 0.8785 -0.0303 0.8785 0.9373
No log 13.7576 454 0.8245 -0.0599 0.8245 0.9080
No log 13.8182 456 0.8267 -0.0599 0.8267 0.9092
No log 13.8788 458 0.8746 0.0913 0.8746 0.9352
No log 13.9394 460 0.8695 0.0586 0.8695 0.9325
No log 14.0 462 0.8639 -0.0647 0.8639 0.9295
No log 14.0606 464 0.8743 -0.0566 0.8743 0.9350
No log 14.1212 466 0.9133 -0.1422 0.9133 0.9557
No log 14.1818 468 0.9386 -0.1099 0.9386 0.9688
No log 14.2424 470 0.9170 -0.1099 0.9170 0.9576
No log 14.3030 472 0.9509 -0.1323 0.9509 0.9751
No log 14.3636 474 0.9205 -0.1054 0.9205 0.9594
No log 14.4242 476 0.9051 -0.0844 0.9051 0.9513
No log 14.4848 478 0.9704 -0.1107 0.9704 0.9851
No log 14.5455 480 1.0287 0.0587 1.0287 1.0143
No log 14.6061 482 0.9651 -0.0393 0.9651 0.9824
No log 14.6667 484 0.9007 -0.0251 0.9007 0.9491
No log 14.7273 486 0.8833 -0.0870 0.8833 0.9398
No log 14.7879 488 0.8989 -0.0844 0.8989 0.9481
No log 14.8485 490 0.9093 -0.0853 0.9093 0.9536
No log 14.9091 492 0.9225 0.0134 0.9225 0.9604
No log 14.9697 494 0.8956 -0.0007 0.8956 0.9464
No log 15.0303 496 0.8545 -0.0475 0.8545 0.9244
No log 15.0909 498 0.9503 -0.0722 0.9503 0.9748
0.306 15.1515 500 0.9903 -0.0317 0.9903 0.9951
0.306 15.2121 502 0.9486 -0.1709 0.9486 0.9739
0.306 15.2727 504 0.8837 -0.1142 0.8837 0.9401
0.306 15.3333 506 0.8138 -0.1397 0.8138 0.9021
0.306 15.3939 508 0.7846 0.0549 0.7846 0.8858
0.306 15.4545 510 0.8156 0.0476 0.8156 0.9031
0.306 15.5152 512 0.8369 -0.0251 0.8369 0.9148
0.306 15.5758 514 0.8780 0.0069 0.8780 0.9370
0.306 15.6364 516 0.9708 -0.0221 0.9708 0.9853
0.306 15.6970 518 0.9846 -0.0221 0.9846 0.9923
0.306 15.7576 520 0.9336 0.0549 0.9336 0.9662
0.306 15.8182 522 0.8707 -0.0441 0.8707 0.9331
0.306 15.8788 524 0.8495 0.0257 0.8495 0.9217
0.306 15.9394 526 0.8197 -0.0138 0.8197 0.9054
0.306 16.0 528 0.7951 0.0282 0.7951 0.8917
0.306 16.0606 530 0.7785 0.0355 0.7785 0.8823
0.306 16.1212 532 0.7932 -0.0541 0.7932 0.8906
0.306 16.1818 534 0.7954 -0.1088 0.7954 0.8918
0.306 16.2424 536 0.8050 -0.0967 0.8050 0.8972
0.306 16.3030 538 0.8470 -0.1142 0.8470 0.9203
0.306 16.3636 540 0.8810 -0.1033 0.8810 0.9386

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
46
Safetensors
Model size
135M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k13_task3_organization

Finetuned
(3994)
this model