Fill-Mask
Transformers
Safetensors
roberta
ChemBERTa
cheminformatics
Eval Results
eacortes's picture
Update README.md
2da3150 verified
---
license: apache-2.0
datasets:
- Derify/augmented_canonical_pubchem_13m
metrics:
- roc_auc
- rmse
library_name: transformers
tags:
- ChemBERTa
- cheminformatics
pipeline_tag: fill-mask
model-index:
- name: Derify/augmented_canonical_pubchem_13m
results:
- task:
type: text-classification
name: Classification (ROC AUC)
dataset:
name: BACE
type: BACE
metrics:
- type: roc_auc
value: 0.8008
- task:
type: text-classification
name: Classification (ROC AUC)
dataset:
name: BBBP
type: BBBP
metrics:
- type: roc_auc
value: 0.7418
- task:
type: text-classification
name: Classification (ROC AUC)
dataset:
name: TOX21
type: TOX21
metrics:
- type: roc_auc
value: 0.7548
- task:
type: text-classification
name: Classification (ROC AUC)
dataset:
name: HIV
type: HIV
metrics:
- type: roc_auc
value: 0.7744
- task:
type: text-classification
name: Classification (ROC AUC)
dataset:
name: SIDER
type: SIDER
metrics:
- type: roc_auc
value: 0.6313
- task:
type: text-classification
name: Classification (ROC AUC)
dataset:
name: CLINTOX
type: CLINTOX
metrics:
- type: roc_auc
value: 0.9621
- task:
type: regression
name: Regression (RMSE)
dataset:
name: ESOL
type: ESOL
metrics:
- type: rmse
value: 0.8798
- task:
type: regression
name: Regression (RMSE)
dataset:
name: FREESOLV
type: FREESOLV
metrics:
- type: rmse
value: 0.5282
- task:
type: regression
name: Regression (RMSE)
dataset:
name: LIPO
type: LIPO
metrics:
- type: rmse
value: 0.6853
- task:
type: regression
name: Regression (RMSE)
dataset:
name: BACE
type: BACE
metrics:
- type: rmse
value: 0.9554
- task:
type: regression
name: Regression (RMSE)
dataset:
name: CLEARANCE
type: CLEARANCE
metrics:
- type: rmse
value: 45.4362
---
This model is a ChemBERTa model trained on the augmented_canonical_pubchem_13m dataset.
The model was trained for 24 epochs using NVIDIA Apex's FusedAdam optimizer with a reduce-on-plateau learning rate scheduler.
To improve performance, mixed precision (fp16), TF32, and torch.compile were enabled. Training used gradient accumulation (16 steps) and batch size of 128 for efficient resource utilization.
Evaluation was performed at regular intervals, with the best model selected based on validation performance.
## Benchmarks
### Classification Datasets (ROC AUC - Higher is better)
| Model | BACE↑ | BBBP↑ | TOX21↑ | HIV↑ | SIDER↑ | CLINTOX↑ |
| ------------------------- | ------ | ------ | ------ | ------ | ------ | -------- |
| **Tasks** | 1 | 1 | 12 | 1 | 27 | 2 |
| Derify/ChemBERTa_augmented_pubchem_13m | 0.8008 | 0.7418 | 0.7548 | 0.7744 | 0.6313 | 0.9621 |
### Regression Datasets (RMSE - Lower is better)
| Model | ESOL↓ | FREESOLV↓ | LIPO↓ | BACE↓ | CLEARANCE↓ |
| ------------------------- | ------ | --------- | ------ | ------ | ---------- |
| **Tasks** | 1 | 1 | 1 | 1 | 1 |
| Derify/ChemBERTa_augmented_pubchem_13m | 0.8798 | 0.5282 | 0.6853 | 0.9554 | 45.4362 |
Benchmarks were conducted using the [chemberta3](https://github.com/deepforestsci/chemberta3) framework.
Datasets were split with DeepChem’s scaffold splits and filtered to include only molecules with SMILES length ≤200, following MolFormer paper's recommendation.
The model was fine-tuned for 100 epochs with a learning rate of 3e-5 and batch size of 32.
Each task was run with 3 different random seeds, and the mean performance is reported.
## References
### ChemBERTa Series
```
@misc{chithrananda2020chembertalargescaleselfsupervisedpretraining,
title={ChemBERTa: Large-Scale Self-Supervised Pretraining for Molecular Property Prediction},
author={Seyone Chithrananda and Gabriel Grand and Bharath Ramsundar},
year={2020},
eprint={2010.09885},
archivePrefix={arXiv},
primaryClass={cs.LG},
url={https://arxiv.org/abs/2010.09885},
}
```
```
@misc{ahmad2022chemberta2chemicalfoundationmodels,
title={ChemBERTa-2: Towards Chemical Foundation Models},
author={Walid Ahmad and Elana Simon and Seyone Chithrananda and Gabriel Grand and Bharath Ramsundar},
year={2022},
eprint={2209.01712},
archivePrefix={arXiv},
primaryClass={cs.LG},
url={https://arxiv.org/abs/2209.01712},
}
```
```
@misc{singh2025chemberta3opensource,
title={ChemBERTa-3: An Open Source Training Framework for Chemical Foundation Models},
author={Singh, R. and Barsainyan, A. A. and Irfan, R. and Amorin, C. J. and He, S. and Davis, T. and others},
year={2025},
howpublished={ChemRxiv},
doi={10.26434/chemrxiv-2025-4glrl-v2},
note={This content is a preprint and has not been peer-reviewed},
url={https://doi.org/10.26434/chemrxiv-2025-4glrl-v2}
}
```