synthetic_relex model for biomedical relation extraction

This is a relation extraction model that is distilled from Llama 3.3 70B down to a BERT model. It is a microsoft/BiomedNLP-BiomedBERT-base-uncased-abstract-fulltext model that has been fine-tuned on synthetic labels created with Llama 3.3 70B when prompted with sentences from PubTator Central. The dataset is available here.

Note: No humans were involved in annotating the dataset used, so there may be erroneous annotations. Detailed evaluation by human experts would be needed to gain an accurate view of the model's accuracy. The dataset and model offer a starting point for understanding and development of biomedical relation extraction models.

More information about the model and dataset can be found at the project repo: https://github.com/Glasgow-AI4BioMed/synthetic_relex

πŸš€ Example Usage

The model can classify the relationship between two entities into one of X labels. The labels are:

To use the model, take the input text and wrap the first entity in [E1][/E1] tags and second entity in [E2][/E2] tags as in the example below. The classifier then outputs the predicted relation label with an associated score.

from transformers import pipeline

classifier = pipeline("text-classification", model="Glasgow-AI4BioMed/synthetic_relex")

classifier("[E1]Paclitaxel[/E1] is a common chemotherapy used for [E2]lung cancer[/E2].")

# Output:
# [{'label': 'treats', 'score': 0.9868311882019043}]

πŸ“ˆ Performance

The results on the test set are reported below:

Label Precision Recall F1-score Support
affects_efficacy_of 0.473 0.296 0.364 1127
binds_to 0.541 0.266 0.357 492
biomarker_for 0.455 0.621 0.525 314
causes 0.667 0.571 0.615 3400
co_expressed_with 0.440 0.473 0.456 131
downregulates 0.472 0.481 0.477 106
inhibits 0.460 0.251 0.324 1429
interacts_with 0.469 0.310 0.373 1588
none 0.936 0.961 0.948 76442
plays_causal_role_in 0.343 0.426 0.380 202
precursor_of 0.462 0.212 0.291 113
prevents 0.602 0.504 0.548 135
regulates 0.504 0.509 0.506 116
subtype_of 0.382 0.521 0.441 286
treats 0.630 0.702 0.664 1000
upregulates 0.564 0.549 0.557 224
macro avg 0.525 0.478 0.489 87105
weighted avg 0.889 0.898 0.892 87105
Downloads last month
70
Safetensors
Model size
109M params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for Glasgow-AI4BioMed/synthetic_relex