ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k11_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8006
  • Qwk: -0.0992
  • Mse: 0.8006
  • Rmse: 0.8948

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0714 2 4.0078 -0.0086 4.0078 2.0019
No log 0.1429 4 2.3040 0.0050 2.3040 1.5179
No log 0.2143 6 1.7188 0.0213 1.7188 1.3110
No log 0.2857 8 2.5964 -0.0173 2.5964 1.6113
No log 0.3571 10 1.6501 0.0 1.6501 1.2845
No log 0.4286 12 0.9892 0.0338 0.9892 0.9946
No log 0.5 14 0.8504 -0.0008 0.8504 0.9222
No log 0.5714 16 0.8559 -0.0852 0.8559 0.9252
No log 0.6429 18 0.9105 -0.0345 0.9105 0.9542
No log 0.7143 20 0.8016 -0.0331 0.8016 0.8953
No log 0.7857 22 0.7142 0.0964 0.7142 0.8451
No log 0.8571 24 0.7123 -0.0101 0.7123 0.8440
No log 0.9286 26 0.7582 -0.0215 0.7582 0.8707
No log 1.0 28 0.7862 -0.0778 0.7862 0.8867
No log 1.0714 30 0.7845 -0.0778 0.7845 0.8857
No log 1.1429 32 0.7984 -0.0331 0.7984 0.8935
No log 1.2143 34 0.9085 -0.0200 0.9085 0.9531
No log 1.2857 36 1.0487 -0.0398 1.0487 1.0240
No log 1.3571 38 1.1888 -0.0247 1.1888 1.0903
No log 1.4286 40 1.2154 -0.0247 1.2154 1.1025
No log 1.5 42 1.2591 0.0 1.2591 1.1221
No log 1.5714 44 1.5619 0.0 1.5619 1.2498
No log 1.6429 46 1.7772 0.0 1.7772 1.3331
No log 1.7143 48 1.6926 0.0 1.6926 1.3010
No log 1.7857 50 1.3587 0.0 1.3587 1.1656
No log 1.8571 52 1.0400 -0.0247 1.0400 1.0198
No log 1.9286 54 0.8784 -0.0200 0.8784 0.9372
No log 2.0 56 0.9408 -0.0385 0.9408 0.9699
No log 2.0714 58 1.1678 -0.0457 1.1678 1.0806
No log 2.1429 60 0.9738 0.0111 0.9738 0.9868
No log 2.2143 62 0.8486 0.0129 0.8486 0.9212
No log 2.2857 64 0.9610 0.0089 0.9610 0.9803
No log 2.3571 66 1.6775 0.0 1.6775 1.2952
No log 2.4286 68 2.0345 0.0 2.0345 1.4263
No log 2.5 70 1.6960 0.0 1.6960 1.3023
No log 2.5714 72 0.9913 -0.0648 0.9913 0.9956
No log 2.6429 74 0.7192 0.0 0.7192 0.8481
No log 2.7143 76 0.7240 0.0 0.7240 0.8509
No log 2.7857 78 0.8239 -0.0753 0.8239 0.9077
No log 2.8571 80 1.1507 0.0065 1.1507 1.0727
No log 2.9286 82 1.4772 0.0 1.4772 1.2154
No log 3.0 84 1.5121 0.0 1.5121 1.2297
No log 3.0714 86 1.2388 -0.0490 1.2388 1.1130
No log 3.1429 88 0.8619 0.0867 0.8619 0.9284
No log 3.2143 90 0.7711 -0.1067 0.7711 0.8781
No log 3.2857 92 0.7763 -0.0499 0.7763 0.8811
No log 3.3571 94 0.7584 -0.0035 0.7584 0.8709
No log 3.4286 96 0.7952 -0.0739 0.7952 0.8917
No log 3.5 98 1.0817 -0.0997 1.0817 1.0400
No log 3.5714 100 1.4947 0.0 1.4947 1.2226
No log 3.6429 102 1.5602 0.0 1.5602 1.2491
No log 3.7143 104 1.4634 0.0 1.4634 1.2097
No log 3.7857 106 1.2619 0.0032 1.2619 1.1233
No log 3.8571 108 1.0654 -0.0686 1.0654 1.0322
No log 3.9286 110 1.0250 0.0006 1.0250 1.0124
No log 4.0 112 1.0063 0.0026 1.0063 1.0032
No log 4.0714 114 1.1073 0.0006 1.1073 1.0523
No log 4.1429 116 1.0283 0.0067 1.0283 1.0141
No log 4.2143 118 0.8550 0.0476 0.8550 0.9246
No log 4.2857 120 0.8535 0.0071 0.8535 0.9239
No log 4.3571 122 0.8225 0.0159 0.8225 0.9069
No log 4.4286 124 0.8770 0.0099 0.8770 0.9365
No log 4.5 126 0.8330 -0.1599 0.8330 0.9127
No log 4.5714 128 0.8953 -0.0204 0.8953 0.9462
No log 4.6429 130 0.9026 -0.0144 0.9026 0.9501
No log 4.7143 132 1.0853 -0.0327 1.0853 1.0418
No log 4.7857 134 1.3112 -0.0746 1.3112 1.1451
No log 4.8571 136 1.0436 -0.0355 1.0436 1.0216
No log 4.9286 138 0.9290 -0.1737 0.9290 0.9638
No log 5.0 140 0.8617 -0.1628 0.8617 0.9283
No log 5.0714 142 0.8067 -0.1148 0.8067 0.8981
No log 5.1429 144 0.8360 -0.0679 0.8360 0.9143
No log 5.2143 146 0.8129 -0.0679 0.8129 0.9016
No log 5.2857 148 0.7987 -0.1220 0.7987 0.8937
No log 5.3571 150 0.8537 -0.1622 0.8537 0.9240
No log 5.4286 152 0.9453 -0.2335 0.9453 0.9723
No log 5.5 154 1.0088 -0.0853 1.0088 1.0044
No log 5.5714 156 0.9184 0.0955 0.9184 0.9583
No log 5.6429 158 0.8941 -0.0089 0.8941 0.9456
No log 5.7143 160 0.8086 -0.1893 0.8086 0.8992
No log 5.7857 162 0.8750 0.0191 0.8750 0.9354
No log 5.8571 164 1.1947 -0.0892 1.1947 1.0930
No log 5.9286 166 1.0628 -0.0877 1.0628 1.0309
No log 6.0 168 0.8180 -0.0711 0.8180 0.9044
No log 6.0714 170 0.8524 -0.0892 0.8524 0.9233
No log 6.1429 172 0.9325 -0.1301 0.9325 0.9657
No log 6.2143 174 0.9280 -0.1643 0.9280 0.9633
No log 6.2857 176 0.9077 -0.1722 0.9077 0.9527
No log 6.3571 178 0.8847 -0.1795 0.8847 0.9406
No log 6.4286 180 0.8702 -0.0550 0.8702 0.9328
No log 6.5 182 0.9051 0.0152 0.9051 0.9514
No log 6.5714 184 0.8713 -0.1395 0.8713 0.9335
No log 6.6429 186 0.8743 -0.1795 0.8743 0.9350
No log 6.7143 188 0.8734 -0.1054 0.8734 0.9346
No log 6.7857 190 0.9064 0.0095 0.9064 0.9520
No log 6.8571 192 0.9167 0.0068 0.9167 0.9574
No log 6.9286 194 0.8160 -0.2439 0.8160 0.9033
No log 7.0 196 0.8828 -0.1606 0.8828 0.9396
No log 7.0714 198 0.8607 -0.1753 0.8607 0.9277
No log 7.1429 200 0.7719 0.0035 0.7719 0.8786
No log 7.2143 202 0.7247 -0.0035 0.7247 0.8513
No log 7.2857 204 0.7492 0.0334 0.7492 0.8655
No log 7.3571 206 0.7610 -0.0160 0.7610 0.8724
No log 7.4286 208 0.7998 -0.1992 0.7998 0.8943
No log 7.5 210 0.9124 0.0315 0.9124 0.9552
No log 7.5714 212 0.9384 0.0734 0.9384 0.9687
No log 7.6429 214 0.8582 -0.0248 0.8582 0.9264
No log 7.7143 216 0.8144 -0.0407 0.8144 0.9024
No log 7.7857 218 0.7966 -0.0204 0.7966 0.8925
No log 7.8571 220 0.7882 0.0260 0.7882 0.8878
No log 7.9286 222 0.7602 -0.0188 0.7602 0.8719
No log 8.0 224 0.7368 0.0506 0.7368 0.8584
No log 8.0714 226 0.7308 -0.0679 0.7308 0.8549
No log 8.1429 228 0.7456 -0.1094 0.7456 0.8635
No log 8.2143 230 0.7958 0.1080 0.7958 0.8921
No log 8.2857 232 0.8331 0.1494 0.8331 0.9128
No log 8.3571 234 0.8055 0.0053 0.8055 0.8975
No log 8.4286 236 0.7917 -0.0465 0.7917 0.8898
No log 8.5 238 0.7898 -0.1033 0.7898 0.8887
No log 8.5714 240 0.8249 -0.0870 0.8249 0.9082
No log 8.6429 242 0.8352 -0.0488 0.8352 0.9139
No log 8.7143 244 0.8389 -0.1531 0.8389 0.9159
No log 8.7857 246 0.8618 -0.0921 0.8618 0.9283
No log 8.8571 248 0.8781 -0.0921 0.8781 0.9371
No log 8.9286 250 0.8982 -0.0524 0.8982 0.9477
No log 9.0 252 1.0144 -0.0870 1.0144 1.0072
No log 9.0714 254 1.0521 -0.0518 1.0521 1.0257
No log 9.1429 256 0.9554 -0.0985 0.9554 0.9774
No log 9.2143 258 0.9103 -0.1659 0.9103 0.9541
No log 9.2857 260 0.8921 -0.0692 0.8921 0.9445
No log 9.3571 262 0.8332 -0.1542 0.8332 0.9128
No log 9.4286 264 0.7961 -0.1163 0.7961 0.8922
No log 9.5 266 0.8095 -0.0711 0.8095 0.8997
No log 9.5714 268 0.8302 -0.1106 0.8302 0.9111
No log 9.6429 270 0.8751 -0.1994 0.8751 0.9355
No log 9.7143 272 0.9101 -0.1841 0.9101 0.9540
No log 9.7857 274 0.9281 -0.1334 0.9281 0.9634
No log 9.8571 276 0.9526 -0.0695 0.9526 0.9760
No log 9.9286 278 0.9363 -0.0576 0.9363 0.9676
No log 10.0 280 0.8948 0.0570 0.8948 0.9460
No log 10.0714 282 0.8560 -0.0831 0.8560 0.9252
No log 10.1429 284 0.8485 -0.0643 0.8485 0.9212
No log 10.2143 286 0.8589 -0.0295 0.8589 0.9268
No log 10.2857 288 0.8301 -0.0595 0.8301 0.9111
No log 10.3571 290 0.8386 -0.1088 0.8386 0.9157
No log 10.4286 292 0.8458 -0.1168 0.8458 0.9197
No log 10.5 294 0.9138 0.0095 0.9138 0.9559
No log 10.5714 296 0.9158 -0.0295 0.9158 0.9570
No log 10.6429 298 0.8774 0.0 0.8774 0.9367
No log 10.7143 300 0.8355 -0.0334 0.8355 0.9140
No log 10.7857 302 0.8193 -0.0984 0.8193 0.9051
No log 10.8571 304 0.8904 -0.0240 0.8904 0.9436
No log 10.9286 306 0.9050 -0.0602 0.9050 0.9513
No log 11.0 308 0.9002 -0.0522 0.9002 0.9488
No log 11.0714 310 0.8965 -0.0861 0.8965 0.9468
No log 11.1429 312 0.8612 -0.1638 0.8612 0.9280
No log 11.2143 314 0.8011 -0.0628 0.8011 0.8950
No log 11.2857 316 0.7432 -0.0179 0.7432 0.8621
No log 11.3571 318 0.7516 -0.0179 0.7516 0.8670
No log 11.4286 320 0.7356 0.0863 0.7356 0.8577
No log 11.5 322 0.7507 -0.0406 0.7507 0.8664
No log 11.5714 324 0.8074 0.0654 0.8074 0.8985
No log 11.6429 326 0.8159 0.1033 0.8159 0.9033
No log 11.7143 328 0.8461 -0.1066 0.8461 0.9198
No log 11.7857 330 0.9572 0.0456 0.9572 0.9783
No log 11.8571 332 0.9327 0.0490 0.9327 0.9658
No log 11.9286 334 0.8767 -0.0870 0.8767 0.9363
No log 12.0 336 0.8457 -0.0585 0.8457 0.9196
No log 12.0714 338 0.8265 0.0673 0.8265 0.9091
No log 12.1429 340 0.7976 -0.0366 0.7976 0.8931
No log 12.2143 342 0.7710 -0.0065 0.7710 0.8781
No log 12.2857 344 0.8183 0.0225 0.8183 0.9046
No log 12.3571 346 0.8711 0.0512 0.8711 0.9333
No log 12.4286 348 0.8644 0.0095 0.8644 0.9297
No log 12.5 350 0.8387 -0.0103 0.8387 0.9158
No log 12.5714 352 0.8459 0.0547 0.8459 0.9197
No log 12.6429 354 0.8581 0.0570 0.8581 0.9263
No log 12.7143 356 0.8494 0.0377 0.8494 0.9216
No log 12.7857 358 0.8546 -0.0163 0.8546 0.9245
No log 12.8571 360 0.8766 0.0639 0.8766 0.9363
No log 12.9286 362 0.8850 0.1243 0.8850 0.9407
No log 13.0 364 0.8220 0.0225 0.8220 0.9066
No log 13.0714 366 0.8055 -0.0091 0.8055 0.8975
No log 13.1429 368 0.8138 -0.0506 0.8138 0.9021
No log 13.2143 370 0.8184 -0.0506 0.8184 0.9047
No log 13.2857 372 0.8110 -0.0675 0.8110 0.9005
No log 13.3571 374 0.8320 0.0670 0.8320 0.9122
No log 13.4286 376 0.8471 0.0670 0.8471 0.9204
No log 13.5 378 0.8309 -0.0228 0.8309 0.9115
No log 13.5714 380 0.8256 -0.0228 0.8256 0.9086
No log 13.6429 382 0.8559 0.0214 0.8559 0.9251
No log 13.7143 384 0.8089 -0.0204 0.8089 0.8994
No log 13.7857 386 0.8045 -0.0675 0.8045 0.8969
No log 13.8571 388 0.8543 -0.0228 0.8543 0.9243
No log 13.9286 390 0.8913 0.0152 0.8913 0.9441
No log 14.0 392 0.9386 0.0016 0.9386 0.9688
No log 14.0714 394 0.9218 0.0068 0.9218 0.9601
No log 14.1429 396 0.9641 -0.0008 0.9641 0.9819
No log 14.2143 398 1.0091 0.0711 1.0091 1.0045
No log 14.2857 400 1.0131 0.0377 1.0131 1.0065
No log 14.3571 402 0.9313 0.0442 0.9313 0.9650
No log 14.4286 404 0.9295 0.0377 0.9295 0.9641
No log 14.5 406 0.8364 -0.0711 0.8364 0.9145
No log 14.5714 408 0.8155 -0.0711 0.8155 0.9030
No log 14.6429 410 0.8220 -0.0711 0.8220 0.9066
No log 14.7143 412 0.8825 0.0999 0.8825 0.9394
No log 14.7857 414 0.9846 0.0346 0.9846 0.9923
No log 14.8571 416 0.9766 0.0456 0.9766 0.9882
No log 14.9286 418 0.9607 -0.0991 0.9607 0.9802
No log 15.0 420 0.9600 -0.1077 0.9600 0.9798
No log 15.0714 422 0.9514 -0.1227 0.9514 0.9754
No log 15.1429 424 0.9783 0.0065 0.9783 0.9891
No log 15.2143 426 0.9850 0.0793 0.9850 0.9925
No log 15.2857 428 0.9158 0.0490 0.9158 0.9570
No log 15.3571 430 0.8557 -0.1116 0.8557 0.9250
No log 15.4286 432 0.8411 -0.1599 0.8411 0.9171
No log 15.5 434 0.8361 -0.2022 0.8361 0.9144
No log 15.5714 436 0.8406 0.0225 0.8406 0.9168
No log 15.6429 438 0.8324 0.0225 0.8324 0.9123
No log 15.7143 440 0.8222 -0.0240 0.8222 0.9067
No log 15.7857 442 0.8183 -0.0215 0.8183 0.9046
No log 15.8571 444 0.8247 -0.1682 0.8247 0.9081
No log 15.9286 446 0.8506 -0.1466 0.8506 0.9223
No log 16.0 448 0.8342 -0.1538 0.8342 0.9133
No log 16.0714 450 0.8057 -0.1682 0.8057 0.8976
No log 16.1429 452 0.7949 -0.1230 0.7949 0.8916
No log 16.2143 454 0.8089 -0.0215 0.8089 0.8994
No log 16.2857 456 0.8405 -0.0704 0.8405 0.9168
No log 16.3571 458 0.8588 -0.0992 0.8588 0.9267
No log 16.4286 460 0.9039 -0.0629 0.9039 0.9508
No log 16.5 462 0.9440 0.0140 0.9440 0.9716
No log 16.5714 464 0.9384 -0.1026 0.9384 0.9687
No log 16.6429 466 0.9149 -0.0915 0.9149 0.9565
No log 16.7143 468 0.9034 -0.1334 0.9034 0.9505
No log 16.7857 470 0.8971 -0.1515 0.8971 0.9471
No log 16.8571 472 0.8790 -0.0690 0.8790 0.9375
No log 16.9286 474 0.8371 -0.1026 0.8371 0.9149
No log 17.0 476 0.8293 -0.1542 0.8293 0.9107
No log 17.0714 478 0.8212 -0.1542 0.8212 0.9062
No log 17.1429 480 0.8246 -0.0675 0.8246 0.9081
No log 17.2143 482 0.8714 0.0512 0.8714 0.9335
No log 17.2857 484 0.8904 0.1243 0.8904 0.9436
No log 17.3571 486 0.8462 0.0549 0.8462 0.9199
No log 17.4286 488 0.8219 -0.0690 0.8219 0.9066
No log 17.5 490 0.8308 -0.1682 0.8308 0.9115
No log 17.5714 492 0.8574 -0.2446 0.8574 0.9259
No log 17.6429 494 0.8567 -0.1653 0.8567 0.9256
No log 17.7143 496 0.8373 -0.0999 0.8373 0.9150
No log 17.7857 498 0.8332 0.0236 0.8332 0.9128
0.3309 17.8571 500 0.8439 0.0205 0.8439 0.9186
0.3309 17.9286 502 0.8387 0.0123 0.8387 0.9158
0.3309 18.0 504 0.8531 0.1291 0.8531 0.9237
0.3309 18.0714 506 0.8628 0.1291 0.8628 0.9289
0.3309 18.1429 508 0.8534 0.0205 0.8534 0.9238
0.3309 18.2143 510 0.8734 -0.1464 0.8734 0.9346
0.3309 18.2857 512 0.9376 0.0418 0.9376 0.9683
0.3309 18.3571 514 0.9365 0.0703 0.9365 0.9677
0.3309 18.4286 516 0.8820 -0.1355 0.8820 0.9391
0.3309 18.5 518 0.8705 -0.1140 0.8705 0.9330
0.3309 18.5714 520 0.9098 0.0456 0.9098 0.9539
0.3309 18.6429 522 0.8826 0.0095 0.8826 0.9395
0.3309 18.7143 524 0.8554 0.0205 0.8554 0.9249
0.3309 18.7857 526 0.8213 -0.0557 0.8213 0.9063
0.3309 18.8571 528 0.8258 -0.0370 0.8258 0.9087
0.3309 18.9286 530 0.8578 0.0705 0.8578 0.9262
0.3309 19.0 532 0.8565 0.0715 0.8565 0.9255
0.3309 19.0714 534 0.8276 0.0610 0.8276 0.9097
0.3309 19.1429 536 0.8176 0.0289 0.8176 0.9042
0.3309 19.2143 538 0.9134 0.0871 0.9134 0.9557
0.3309 19.2857 540 0.9810 0.0984 0.9810 0.9905
0.3309 19.3571 542 0.9255 0.1107 0.9255 0.9620
0.3309 19.4286 544 0.8267 0.0183 0.8267 0.9092
0.3309 19.5 546 0.8039 -0.0889 0.8039 0.8966
0.3309 19.5714 548 0.8096 -0.0831 0.8096 0.8998
0.3309 19.6429 550 0.8093 -0.0831 0.8093 0.8996
0.3309 19.7143 552 0.8006 -0.0992 0.8006 0.8948

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
9
Safetensors
Model size
135M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k11_task3_organization

Finetuned
(3994)
this model