ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k20_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0756
  • Qwk: -0.0182
  • Mse: 1.0756
  • Rmse: 1.0371

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.04 2 3.6495 -0.0051 3.6495 1.9104
No log 0.08 4 2.1092 0.0643 2.1092 1.4523
No log 0.12 6 2.2727 0.0076 2.2727 1.5076
No log 0.16 8 1.4068 0.0045 1.4068 1.1861
No log 0.2 10 1.7743 -0.0431 1.7743 1.3320
No log 0.24 12 1.9622 -0.0241 1.9622 1.4008
No log 0.28 14 0.9517 -0.0143 0.9517 0.9755
No log 0.32 16 0.7105 -0.0695 0.7105 0.8429
No log 0.36 18 0.7815 0.0159 0.7815 0.8840
No log 0.4 20 1.1852 -0.0468 1.1852 1.0887
No log 0.44 22 1.8096 -0.0265 1.8096 1.3452
No log 0.48 24 1.7371 -0.0265 1.7371 1.3180
No log 0.52 26 1.1175 0.0 1.1175 1.0571
No log 0.56 28 0.8423 -0.0852 0.8423 0.9178
No log 0.6 30 0.8201 0.0129 0.8201 0.9056
No log 0.64 32 0.9250 -0.0638 0.9250 0.9618
No log 0.68 34 0.8575 0.0287 0.8575 0.9260
No log 0.72 36 1.0037 -0.0398 1.0037 1.0019
No log 0.76 38 1.0246 -0.0411 1.0246 1.0122
No log 0.8 40 1.2545 -0.0468 1.2545 1.1200
No log 0.84 42 1.3908 -0.0468 1.3908 1.1793
No log 0.88 44 1.4291 0.0048 1.4291 1.1955
No log 0.92 46 1.1245 -0.0446 1.1245 1.0604
No log 0.96 48 0.8784 0.0346 0.8784 0.9372
No log 1.0 50 0.7699 -0.0160 0.7699 0.8774
No log 1.04 52 0.8468 0.0442 0.8468 0.9202
No log 1.08 54 0.8950 0.0233 0.8950 0.9460
No log 1.12 56 0.9383 -0.0545 0.9383 0.9686
No log 1.16 58 1.0168 -0.0930 1.0168 1.0084
No log 1.2 60 1.8268 -0.0241 1.8268 1.3516
No log 1.24 62 2.2176 -0.0263 2.2176 1.4892
No log 1.28 64 1.9999 -0.0247 1.9999 1.4142
No log 1.32 66 1.5452 -0.0221 1.5452 1.2431
No log 1.3600 68 0.8990 -0.0240 0.8990 0.9482
No log 1.4 70 0.8560 -0.0892 0.8560 0.9252
No log 1.44 72 0.8460 -0.0958 0.8460 0.9198
No log 1.48 74 1.0325 -0.1279 1.0325 1.0161
No log 1.52 76 1.7710 -0.0014 1.7710 1.3308
No log 1.56 78 1.8453 -0.0264 1.8453 1.3584
No log 1.6 80 1.2716 -0.0133 1.2716 1.1276
No log 1.6400 82 0.8392 -0.0215 0.8392 0.9161
No log 1.6800 84 0.7946 -0.0520 0.7946 0.8914
No log 1.72 86 0.7793 -0.0035 0.7793 0.8828
No log 1.76 88 0.8384 -0.0331 0.8384 0.9156
No log 1.8 90 1.0952 -0.0677 1.0952 1.0465
No log 1.8400 92 1.2324 -0.0145 1.2324 1.1101
No log 1.88 94 0.9740 0.0233 0.9740 0.9869
No log 1.92 96 0.7798 0.0479 0.7798 0.8831
No log 1.96 98 0.8266 -0.0644 0.8266 0.9092
No log 2.0 100 1.0445 0.0217 1.0445 1.0220
No log 2.04 102 1.3980 -0.0629 1.3980 1.1824
No log 2.08 104 1.1779 -0.1196 1.1779 1.0853
No log 2.12 106 0.9919 -0.0440 0.9919 0.9959
No log 2.16 108 0.9616 -0.1159 0.9616 0.9806
No log 2.2 110 0.8873 0.0303 0.8873 0.9420
No log 2.24 112 0.9226 -0.0699 0.9226 0.9605
No log 2.2800 114 1.2133 -0.1192 1.2133 1.1015
No log 2.32 116 2.0415 -0.0638 2.0415 1.4288
No log 2.36 118 1.9093 -0.1544 1.9093 1.3818
No log 2.4 120 1.2113 -0.0804 1.2113 1.1006
No log 2.44 122 0.9697 -0.1755 0.9697 0.9847
No log 2.48 124 0.9891 -0.0186 0.9891 0.9946
No log 2.52 126 1.1478 -0.0832 1.1478 1.0714
No log 2.56 128 1.0629 -0.0376 1.0629 1.0310
No log 2.6 130 0.9789 -0.1280 0.9789 0.9894
No log 2.64 132 1.0017 -0.1632 1.0017 1.0008
No log 2.68 134 0.9990 -0.0764 0.9990 0.9995
No log 2.7200 136 1.4073 -0.0817 1.4073 1.1863
No log 2.76 138 2.4080 -0.0777 2.4080 1.5518
No log 2.8 140 2.2970 -0.0428 2.2970 1.5156
No log 2.84 142 1.5426 -0.0613 1.5426 1.2420
No log 2.88 144 0.9136 0.0 0.9136 0.9558
No log 2.92 146 0.9447 -0.0648 0.9447 0.9719
No log 2.96 148 0.9580 -0.0989 0.9580 0.9788
No log 3.0 150 0.8688 -0.1331 0.8688 0.9321
No log 3.04 152 0.9786 -0.1671 0.9786 0.9892
No log 3.08 154 1.2215 -0.0285 1.2215 1.1052
No log 3.12 156 1.1538 -0.0269 1.1538 1.0742
No log 3.16 158 0.9406 -0.0801 0.9406 0.9699
No log 3.2 160 0.8492 0.0318 0.8492 0.9215
No log 3.24 162 0.8890 -0.0252 0.8890 0.9429
No log 3.2800 164 0.9594 -0.1148 0.9594 0.9795
No log 3.32 166 0.9263 -0.0723 0.9263 0.9624
No log 3.36 168 0.9278 -0.0723 0.9278 0.9632
No log 3.4 170 0.9181 -0.1121 0.9181 0.9582
No log 3.44 172 0.9269 -0.0731 0.9269 0.9628
No log 3.48 174 1.2121 -0.0175 1.2121 1.1010
No log 3.52 176 1.3553 -0.0561 1.3553 1.1642
No log 3.56 178 1.0985 -0.1175 1.0985 1.0481
No log 3.6 180 0.9060 -0.0274 0.9060 0.9519
No log 3.64 182 0.8027 -0.0493 0.8027 0.8959
No log 3.68 184 0.7656 -0.0033 0.7656 0.8750
No log 3.7200 186 0.8400 0.0129 0.8400 0.9165
No log 3.76 188 1.1516 -0.0345 1.1516 1.0731
No log 3.8 190 1.2439 -0.0348 1.2439 1.1153
No log 3.84 192 0.9898 -0.0474 0.9898 0.9949
No log 3.88 194 0.8088 -0.0125 0.8088 0.8994
No log 3.92 196 0.8252 -0.1111 0.8252 0.9084
No log 3.96 198 0.9461 -0.0767 0.9461 0.9727
No log 4.0 200 0.9543 -0.0767 0.9543 0.9769
No log 4.04 202 0.9418 -0.2270 0.9418 0.9705
No log 4.08 204 1.0056 -0.1026 1.0056 1.0028
No log 4.12 206 1.0497 -0.1088 1.0497 1.0245
No log 4.16 208 1.0113 -0.0495 1.0113 1.0056
No log 4.2 210 0.9711 -0.0915 0.9711 0.9854
No log 4.24 212 0.9246 -0.1333 0.9246 0.9616
No log 4.28 214 0.9676 0.0017 0.9676 0.9837
No log 4.32 216 0.9893 -0.0008 0.9893 0.9946
No log 4.36 218 0.9695 -0.0230 0.9695 0.9846
No log 4.4 220 0.9248 -0.2295 0.9248 0.9617
No log 4.44 222 0.9039 -0.1047 0.9039 0.9508
No log 4.48 224 1.0088 -0.1209 1.0088 1.0044
No log 4.52 226 0.8847 0.0214 0.8847 0.9406
No log 4.5600 228 0.7700 -0.1697 0.7700 0.8775
No log 4.6 230 0.7517 -0.0591 0.7517 0.8670
No log 4.64 232 0.7882 -0.0695 0.7882 0.8878
No log 4.68 234 0.7865 -0.1158 0.7865 0.8869
No log 4.72 236 0.8106 -0.1616 0.8106 0.9003
No log 4.76 238 0.8914 -0.1126 0.8914 0.9442
No log 4.8 240 1.0892 -0.0079 1.0892 1.0436
No log 4.84 242 1.0347 -0.0746 1.0347 1.0172
No log 4.88 244 0.9854 -0.0373 0.9854 0.9927
No log 4.92 246 1.0290 -0.0163 1.0290 1.0144
No log 4.96 248 0.9577 -0.0459 0.9577 0.9786
No log 5.0 250 0.8718 0.0071 0.8718 0.9337
No log 5.04 252 0.8464 -0.0488 0.8464 0.9200
No log 5.08 254 0.8931 -0.1211 0.8931 0.9450
No log 5.12 256 0.8987 -0.1211 0.8987 0.9480
No log 5.16 258 0.9134 -0.0082 0.9134 0.9557
No log 5.2 260 0.9305 -0.0483 0.9305 0.9646
No log 5.24 262 0.9319 -0.1214 0.9319 0.9653
No log 5.28 264 0.9686 -0.0262 0.9686 0.9842
No log 5.32 266 1.1915 -0.0912 1.1915 1.0916
No log 5.36 268 1.2247 -0.0937 1.2247 1.1066
No log 5.4 270 1.0593 -0.1212 1.0593 1.0292
No log 5.44 272 0.9324 -0.0930 0.9324 0.9656
No log 5.48 274 0.9687 -0.0240 0.9687 0.9842
No log 5.52 276 0.9586 -0.0240 0.9586 0.9791
No log 5.5600 278 0.9566 -0.0262 0.9566 0.9781
No log 5.6 280 0.9869 -0.0390 0.9869 0.9934
No log 5.64 282 0.9835 -0.0833 0.9835 0.9917
No log 5.68 284 0.9899 -0.0833 0.9899 0.9949
No log 5.72 286 0.9362 0.0159 0.9362 0.9676
No log 5.76 288 0.9767 -0.1255 0.9767 0.9883
No log 5.8 290 1.0224 0.0618 1.0224 1.0111
No log 5.84 292 0.8834 -0.0145 0.8834 0.9399
No log 5.88 294 0.8723 -0.0612 0.8723 0.9340
No log 5.92 296 0.8770 -0.0309 0.8770 0.9365
No log 5.96 298 0.9055 -0.0371 0.9055 0.9516
No log 6.0 300 0.9294 -0.0823 0.9294 0.9640
No log 6.04 302 0.8206 -0.0160 0.8206 0.9059
No log 6.08 304 0.8094 -0.1538 0.8094 0.8996
No log 6.12 306 0.8121 -0.0541 0.8121 0.9012
No log 6.16 308 0.8412 0.0129 0.8412 0.9172
No log 6.2 310 0.9466 -0.0008 0.9466 0.9729
No log 6.24 312 0.9645 -0.0008 0.9645 0.9821
No log 6.28 314 0.8904 0.0159 0.8904 0.9436
No log 6.32 316 0.9321 -0.0371 0.9321 0.9654
No log 6.36 318 1.0219 -0.0490 1.0219 1.0109
No log 6.4 320 0.9539 -0.0371 0.9539 0.9767
No log 6.44 322 0.9273 0.1141 0.9273 0.9630
No log 6.48 324 0.9501 -0.1302 0.9501 0.9747
No log 6.52 326 0.9399 -0.2017 0.9399 0.9695
No log 6.5600 328 0.9637 0.0412 0.9637 0.9817
No log 6.6 330 1.0949 -0.0409 1.0949 1.0464
No log 6.64 332 1.0495 0.0016 1.0495 1.0244
No log 6.68 334 0.9679 -0.0027 0.9679 0.9838
No log 6.72 336 0.9576 -0.0831 0.9576 0.9786
No log 6.76 338 0.9104 -0.1106 0.9104 0.9541
No log 6.8 340 0.9470 -0.0033 0.9470 0.9732
No log 6.84 342 1.0056 -0.0236 1.0056 1.0028
No log 6.88 344 0.9820 0.0089 0.9820 0.9910
No log 6.92 346 0.9136 -0.0442 0.9136 0.9558
No log 6.96 348 0.9043 -0.0755 0.9043 0.9510
No log 7.0 350 0.9520 -0.0799 0.9520 0.9757
No log 7.04 352 1.0846 -0.0200 1.0846 1.0414
No log 7.08 354 1.1523 -0.0276 1.1523 1.0735
No log 7.12 356 0.9855 -0.0852 0.9855 0.9927
No log 7.16 358 0.8895 -0.0389 0.8895 0.9431
No log 7.2 360 0.9124 -0.0393 0.9124 0.9552
No log 7.24 362 0.9109 -0.0528 0.9109 0.9544
No log 7.28 364 0.8280 -0.0949 0.8280 0.9099
No log 7.32 366 0.8053 -0.0520 0.8053 0.8974
No log 7.36 368 0.7901 -0.0033 0.7901 0.8889
No log 7.4 370 0.8094 -0.0711 0.8094 0.8997
No log 7.44 372 0.9169 -0.0812 0.9169 0.9576
No log 7.48 374 0.8921 -0.0351 0.8921 0.9445
No log 7.52 376 0.9066 -0.0823 0.9066 0.9522
No log 7.5600 378 0.8737 -0.0778 0.8737 0.9347
No log 7.6 380 0.8962 -0.0823 0.8962 0.9467
No log 7.64 382 0.9164 -0.0056 0.9164 0.9573
No log 7.68 384 0.9284 -0.0143 0.9284 0.9635
No log 7.72 386 0.8844 -0.0033 0.8844 0.9404
No log 7.76 388 0.8514 -0.0658 0.8514 0.9227
No log 7.8 390 0.8311 -0.0506 0.8311 0.9117
No log 7.84 392 0.8150 -0.0513 0.8150 0.9028
No log 7.88 394 0.8027 -0.0091 0.8027 0.8960
No log 7.92 396 0.8607 -0.0390 0.8607 0.9278
No log 7.96 398 1.0547 0.0353 1.0547 1.0270
No log 8.0 400 1.0834 0.0353 1.0834 1.0409
No log 8.04 402 0.9134 -0.0861 0.9134 0.9557
No log 8.08 404 0.8281 0.0488 0.8281 0.9100
No log 8.12 406 0.8926 -0.1532 0.8926 0.9448
No log 8.16 408 0.9469 -0.1372 0.9469 0.9731
No log 8.2 410 0.9152 -0.0300 0.9152 0.9566
No log 8.24 412 0.9856 -0.0122 0.9856 0.9928
No log 8.28 414 1.3918 -0.1194 1.3918 1.1797
No log 8.32 416 1.5161 -0.0890 1.5161 1.2313
No log 8.36 418 1.2696 -0.0978 1.2696 1.1268
No log 8.4 420 0.9617 -0.0163 0.9617 0.9807
No log 8.44 422 0.8123 -0.0215 0.8123 0.9013
No log 8.48 424 0.7922 -0.0513 0.7922 0.8901
No log 8.52 426 0.7758 -0.0560 0.7758 0.8808
No log 8.56 428 0.7776 -0.0711 0.7776 0.8818
No log 8.6 430 0.8561 -0.0033 0.8561 0.9253
No log 8.64 432 0.9725 0.0458 0.9725 0.9862
No log 8.68 434 0.9675 0.0576 0.9675 0.9836
No log 8.72 436 0.9282 -0.0033 0.9282 0.9634
No log 8.76 438 0.9211 -0.1148 0.9211 0.9598
No log 8.8 440 0.9183 -0.0723 0.9183 0.9583
No log 8.84 442 0.8747 -0.0195 0.8747 0.9352
No log 8.88 444 0.8379 0.0225 0.8379 0.9154
No log 8.92 446 0.8665 0.0099 0.8665 0.9309
No log 8.96 448 0.8932 -0.0861 0.8932 0.9451
No log 9.0 450 0.9628 -0.0143 0.9628 0.9812
No log 9.04 452 1.0268 0.0458 1.0268 1.0133
No log 9.08 454 0.9664 -0.0122 0.9664 0.9830
No log 9.12 456 0.8709 0.0129 0.8709 0.9332
No log 9.16 458 0.8483 0.0660 0.8483 0.9210
No log 9.2 460 0.8426 0.1047 0.8426 0.9179
No log 9.24 462 0.9036 -0.0143 0.9036 0.9506
No log 9.28 464 0.9245 0.0182 0.9245 0.9615
No log 9.32 466 0.9986 0.0134 0.9986 0.9993
No log 9.36 468 0.9749 0.0182 0.9749 0.9873
No log 9.4 470 0.8639 0.0525 0.8639 0.9295
No log 9.44 472 0.8912 0.1315 0.8912 0.9440
No log 9.48 474 0.9242 -0.0159 0.9242 0.9614
No log 9.52 476 0.8742 -0.1461 0.8742 0.9350
No log 9.56 478 0.8532 0.0129 0.8532 0.9237
No log 9.6 480 1.0438 0.0458 1.0438 1.0217
No log 9.64 482 1.1729 -0.0345 1.1729 1.0830
No log 9.68 484 1.0772 0.0067 1.0772 1.0379
No log 9.72 486 0.9962 0.0157 0.9962 0.9981
No log 9.76 488 0.8715 -0.0371 0.8715 0.9335
No log 9.8 490 0.8490 0.0129 0.8490 0.9214
No log 9.84 492 0.8317 0.0225 0.8317 0.9120
No log 9.88 494 0.8604 -0.0008 0.8604 0.9276
No log 9.92 496 0.8512 -0.0008 0.8512 0.9226
No log 9.96 498 0.8029 0.0759 0.8029 0.8960
0.3416 10.0 500 0.8047 0.0436 0.8047 0.8970
0.3416 10.04 502 0.8242 0.0454 0.8242 0.9079
0.3416 10.08 504 0.8648 -0.1515 0.8648 0.9299
0.3416 10.12 506 0.9382 -0.1203 0.9382 0.9686
0.3416 10.16 508 1.0438 -0.0182 1.0438 1.0217
0.3416 10.2 510 1.1098 -0.0182 1.1098 1.0535
0.3416 10.24 512 1.0756 -0.0182 1.0756 1.0371

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
12
Safetensors
Model size
135M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k20_task3_organization

Finetuned
(3994)
this model