metadata
library_name: transformers
license: apache-2.0
metrics:
- perplexity
base_model:
- facebook/esm1b_t33_650M_UR50S
Pretraining on Combined Phosphosite Data
ESM-1b is trained from scratch by Masked Language Modeling objective. The data is combination of phosphosite data which are used to train isikz/esm1b_msa_mlm_pt_phosphosite and isikz/esm1b_mlm_pt_phosphosite. The total number of data is 1055221.