Fill-Mask
Transformers
Safetensors
roberta
ChemBERTa
cheminformatics
Eval Results
eacortes's picture
Update README.md
2da3150 verified
metadata
license: apache-2.0
datasets:
  - Derify/augmented_canonical_pubchem_13m
metrics:
  - roc_auc
  - rmse
library_name: transformers
tags:
  - ChemBERTa
  - cheminformatics
pipeline_tag: fill-mask
model-index:
  - name: Derify/augmented_canonical_pubchem_13m
    results:
      - task:
          type: text-classification
          name: Classification (ROC AUC)
        dataset:
          name: BACE
          type: BACE
        metrics:
          - type: roc_auc
            value: 0.8008
      - task:
          type: text-classification
          name: Classification (ROC AUC)
        dataset:
          name: BBBP
          type: BBBP
        metrics:
          - type: roc_auc
            value: 0.7418
      - task:
          type: text-classification
          name: Classification (ROC AUC)
        dataset:
          name: TOX21
          type: TOX21
        metrics:
          - type: roc_auc
            value: 0.7548
      - task:
          type: text-classification
          name: Classification (ROC AUC)
        dataset:
          name: HIV
          type: HIV
        metrics:
          - type: roc_auc
            value: 0.7744
      - task:
          type: text-classification
          name: Classification (ROC AUC)
        dataset:
          name: SIDER
          type: SIDER
        metrics:
          - type: roc_auc
            value: 0.6313
      - task:
          type: text-classification
          name: Classification (ROC AUC)
        dataset:
          name: CLINTOX
          type: CLINTOX
        metrics:
          - type: roc_auc
            value: 0.9621
      - task:
          type: regression
          name: Regression (RMSE)
        dataset:
          name: ESOL
          type: ESOL
        metrics:
          - type: rmse
            value: 0.8798
      - task:
          type: regression
          name: Regression (RMSE)
        dataset:
          name: FREESOLV
          type: FREESOLV
        metrics:
          - type: rmse
            value: 0.5282
      - task:
          type: regression
          name: Regression (RMSE)
        dataset:
          name: LIPO
          type: LIPO
        metrics:
          - type: rmse
            value: 0.6853
      - task:
          type: regression
          name: Regression (RMSE)
        dataset:
          name: BACE
          type: BACE
        metrics:
          - type: rmse
            value: 0.9554
      - task:
          type: regression
          name: Regression (RMSE)
        dataset:
          name: CLEARANCE
          type: CLEARANCE
        metrics:
          - type: rmse
            value: 45.4362

This model is a ChemBERTa model trained on the augmented_canonical_pubchem_13m dataset.

The model was trained for 24 epochs using NVIDIA Apex's FusedAdam optimizer with a reduce-on-plateau learning rate scheduler. To improve performance, mixed precision (fp16), TF32, and torch.compile were enabled. Training used gradient accumulation (16 steps) and batch size of 128 for efficient resource utilization. Evaluation was performed at regular intervals, with the best model selected based on validation performance.

Benchmarks

Classification Datasets (ROC AUC - Higher is better)

Model BACE↑ BBBP↑ TOX21↑ HIV↑ SIDER↑ CLINTOX↑
Tasks 1 1 12 1 27 2
Derify/ChemBERTa_augmented_pubchem_13m 0.8008 0.7418 0.7548 0.7744 0.6313 0.9621

Regression Datasets (RMSE - Lower is better)

Model ESOL↓ FREESOLV↓ LIPO↓ BACE↓ CLEARANCE↓
Tasks 1 1 1 1 1
Derify/ChemBERTa_augmented_pubchem_13m 0.8798 0.5282 0.6853 0.9554 45.4362

Benchmarks were conducted using the chemberta3 framework. Datasets were split with DeepChem’s scaffold splits and filtered to include only molecules with SMILES length ≤200, following MolFormer paper's recommendation. The model was fine-tuned for 100 epochs with a learning rate of 3e-5 and batch size of 32. Each task was run with 3 different random seeds, and the mean performance is reported.

References

ChemBERTa Series

@misc{chithrananda2020chembertalargescaleselfsupervisedpretraining,
      title={ChemBERTa: Large-Scale Self-Supervised Pretraining for Molecular Property Prediction}, 
      author={Seyone Chithrananda and Gabriel Grand and Bharath Ramsundar},
      year={2020},
      eprint={2010.09885},
      archivePrefix={arXiv},
      primaryClass={cs.LG},
      url={https://arxiv.org/abs/2010.09885}, 
}
@misc{ahmad2022chemberta2chemicalfoundationmodels,
      title={ChemBERTa-2: Towards Chemical Foundation Models}, 
      author={Walid Ahmad and Elana Simon and Seyone Chithrananda and Gabriel Grand and Bharath Ramsundar},
      year={2022},
      eprint={2209.01712},
      archivePrefix={arXiv},
      primaryClass={cs.LG},
      url={https://arxiv.org/abs/2209.01712}, 
}
@misc{singh2025chemberta3opensource,
  title={ChemBERTa-3: An Open Source Training Framework for Chemical Foundation Models},
  author={Singh, R. and Barsainyan, A. A. and Irfan, R. and Amorin, C. J. and He, S. and Davis, T. and others},
  year={2025},
  howpublished={ChemRxiv},
  doi={10.26434/chemrxiv-2025-4glrl-v2},
  note={This content is a preprint and has not been peer-reviewed},
  url={https://doi.org/10.26434/chemrxiv-2025-4glrl-v2}
}