openai-fineweb-edu-scorer-mdeberta-binary-lr5e-05-20250411_132948

This model is a fine-tuned version of microsoft/mdeberta-v3-base on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.1606
  • Precision: 1.0
  • Recall: 1.0
  • F1 Macro: 1.0
  • Accuracy: 1.0

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 32
  • eval_batch_size: 64
  • seed: 0
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • num_epochs: 20

Training results

Training Loss Epoch Step Validation Loss Precision Recall F1 Macro Accuracy
No log 0 0 0.3573 1.0 1.0 1.0 1.0
0.1215 0.3908 1000 0.1103 1.0 1.0 1.0 1.0
0.1107 0.7816 2000 0.1143 1.0 1.0 1.0 1.0
0.0892 1.1723 3000 0.1120 1.0 1.0 1.0 1.0
0.0893 1.5631 4000 0.1125 1.0 1.0 1.0 1.0
0.0888 1.9539 5000 0.1101 1.0 1.0 1.0 1.0
0.0433 2.3447 6000 0.1286 1.0 1.0 1.0 1.0
0.0524 2.7354 7000 0.1353 1.0 1.0 1.0 1.0
0.0302 3.1262 8000 0.1432 1.0 1.0 1.0 1.0
0.0344 3.5170 9000 0.1357 1.0 1.0 1.0 1.0
0.0331 3.9078 10000 0.1378 1.0 1.0 1.0 1.0
0.0227 4.2986 11000 0.1442 1.0 1.0 1.0 1.0
0.0211 4.6893 12000 0.1437 1.0 1.0 1.0 1.0
0.0184 5.0801 13000 0.1472 1.0 1.0 1.0 1.0
0.0188 5.4709 14000 0.1452 1.0 1.0 1.0 1.0
0.0238 5.8617 15000 0.1410 1.0 1.0 1.0 1.0
0.0142 6.2524 16000 0.1488 1.0 1.0 1.0 1.0
0.0125 6.6432 17000 0.1568 1.0 1.0 1.0 1.0
0.0119 7.0340 18000 0.1507 1.0 1.0 1.0 1.0
0.0177 7.4248 19000 0.1593 1.0 1.0 1.0 1.0
0.0136 7.8156 20000 0.1544 1.0 1.0 1.0 1.0
0.0118 8.2063 21000 0.1641 1.0 1.0 1.0 1.0
0.0152 8.5971 22000 0.1524 1.0 1.0 1.0 1.0
0.0113 8.9879 23000 0.1554 1.0 1.0 1.0 1.0
0.0114 9.3787 24000 0.1482 1.0 1.0 1.0 1.0
0.0121 9.7694 25000 0.1451 1.0 1.0 1.0 1.0
0.0089 10.1602 26000 0.1561 1.0 1.0 1.0 1.0
0.0087 10.5510 27000 0.1621 1.0 1.0 1.0 1.0
0.0114 10.9418 28000 0.1553 1.0 1.0 1.0 1.0
0.0061 11.3326 29000 0.1547 1.0 1.0 1.0 1.0
0.0075 11.7233 30000 0.1578 1.0 1.0 1.0 1.0
0.0063 12.1141 31000 0.1575 1.0 1.0 1.0 1.0
0.0081 12.5049 32000 0.1587 1.0 1.0 1.0 1.0
0.0088 12.8957 33000 0.1592 1.0 1.0 1.0 1.0
0.0062 13.2864 34000 0.1614 1.0 1.0 1.0 1.0
0.0085 13.6772 35000 0.1563 1.0 1.0 1.0 1.0
0.0056 14.0680 36000 0.1585 1.0 1.0 1.0 1.0
0.0046 14.4588 37000 0.1608 1.0 1.0 1.0 1.0
0.0066 14.8496 38000 0.1697 1.0 1.0 1.0 1.0
0.0068 15.2403 39000 0.1570 1.0 1.0 1.0 1.0
0.0046 15.6311 40000 0.1745 1.0 1.0 1.0 1.0
0.0062 16.0219 41000 0.1617 1.0 1.0 1.0 1.0
0.0056 16.4127 42000 0.1653 1.0 1.0 1.0 1.0
0.0053 16.8034 43000 0.1551 1.0 1.0 1.0 1.0
0.0026 17.1942 44000 0.1638 1.0 1.0 1.0 1.0
0.005 17.5850 45000 0.1555 1.0 1.0 1.0 1.0
0.0033 17.9758 46000 0.1577 1.0 1.0 1.0 1.0
0.0049 18.3665 47000 0.1545 1.0 1.0 1.0 1.0
0.0043 18.7573 48000 0.1589 1.0 1.0 1.0 1.0
0.0046 19.1481 49000 0.1588 1.0 1.0 1.0 1.0
0.0047 19.5389 50000 0.1596 1.0 1.0 1.0 1.0
0.0041 19.9297 51000 0.1606 1.0 1.0 1.0 1.0

Framework versions

  • Transformers 4.49.0
  • Pytorch 2.6.0+cu124
  • Datasets 3.3.2
  • Tokenizers 0.21.1
Downloads last month
0
Safetensors
Model size
279M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for whoisjones/openai-fineweb-edu-scorer-mdeberta-binary-lr5e-05-20250411_132948

Finetuned
(165)
this model