Edit model card

bert-uncased-L10-H128-A2

This is one of 24 smaller BERT models (English only, uncased, trained with WordPiece masking) released by google-research/bert.

These BERT models was released as TensorFlow checkpoints, however, this is the converted version to PyTorch. More information can be found in google-research/bert or lyeoni/convert-tf-to-pytorch.

Evaluation

Here are the evaluation scores (F1/Accuracy) for the MPRC task.

Model MRPC
BERT-Tiny 81.22/68.38
BERT-Mini 81.43/69.36
BERT-Small 81.41/70.34
BERT-Medium 83.33/73.53
BERT-Base 85.62/78.19

References

@article{turc2019,
  title={Well-Read Students Learn Better: On the Importance of Pre-training Compact Models},
  author={Turc, Iulia and Chang, Ming-Wei and Lee, Kenton and Toutanova, Kristina},
  journal={arXiv preprint arXiv:1908.08962v2 },
  year={2019}
}
Downloads last month
3
Inference API
Unable to determine this model’s pipeline type. Check the docs .