Update README.md
Browse files
README.md
CHANGED
@@ -2,28 +2,46 @@
|
|
2 |
license: unknown
|
3 |
base_model: microsoft/deberta-v3-base
|
4 |
tags:
|
5 |
-
- generated_from_trainer
|
6 |
- medical
|
7 |
-
|
8 |
-
-
|
9 |
-
|
10 |
-
|
11 |
-
-
|
12 |
-
example_title: Example-1
|
13 |
-
- text: A 30-year-old female (65 kg) underwent rhinoplasty under general anaesthesia, which was induced using a combination of a bolus of Remifentanyl (0.5 μg/kg) and Propofol 2 mg/kg.
|
14 |
-
example_title: Example-2
|
15 |
-
- text: >-
|
16 |
-
An 18-year-old male was diagnosed with attention-deficit hyperactivity disorder (ADHD) in 2005.He was overweight with a body mass index (BMI) of 40.
|
17 |
-
example_title: example 3
|
18 |
pipeline_tag: token-classification
|
|
|
|
|
19 |
---
|
20 |
|
21 |
|
22 |
### **BIOMed_NER: Named Entity Recognition for Biomedical Entities**
|
23 |
|
24 |
**Model Overview:**
|
25 |
-
BIOMed_NER is a Named Entity Recognition (NER) model which identifies
|
|
|
26 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
27 |
**Hyperparameters:**
|
28 |
- **Base Model**: `microsoft/deberta-v3-base`
|
29 |
- **Learning Rate**: `3e-5`
|
|
|
2 |
license: unknown
|
3 |
base_model: microsoft/deberta-v3-base
|
4 |
tags:
|
|
|
5 |
- medical
|
6 |
+
- biology
|
7 |
+
- NER
|
8 |
+
- Biomedical
|
9 |
+
- deberta
|
10 |
+
- dataset
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
pipeline_tag: token-classification
|
12 |
+
language:
|
13 |
+
- en
|
14 |
---
|
15 |
|
16 |
|
17 |
### **BIOMed_NER: Named Entity Recognition for Biomedical Entities**
|
18 |
|
19 |
**Model Overview:**
|
20 |
+
BIOMed_NER is a Named Entity Recognition (NER) model which identifies biomedical entities using DeBERTaV3. This model is useful for extracting structured information from clinical text, such as diseases, procedures, medications, and anatomical terms.
|
21 |
+
Here's a more detailed and enthusiastic introduction emphasizing the strengths of **DeBERTa** and why it's an excellent choice for your **BIOMed_NER** model:
|
22 |
|
23 |
+
---
|
24 |
+
|
25 |
+
### Why DeBERTa for Biomedical NER?
|
26 |
+
|
27 |
+
DeBERTa (Decoding-enhanced BERT with Disentangled Attention) represents a significant leap forward in NLP model architecture, particularly for nuanced tasks like Named Entity Recognition (NER) in complex domains such as biomedical texts. Here’s why DeBERTa was the ideal choice for **BIOMed_NER**:
|
28 |
+
|
29 |
+
1. **Advanced Disentangled Attention Mechanism**:
|
30 |
+
- DeBERTa goes beyond traditional transformers by using a unique disentangled attention mechanism that separately encodes word content and word position. This allows DeBERTa to capture the contextual meaning of biomedical terms and understand complex sentence structures, which is essential for accurately tagging biomedical entities that often have overlapping or highly specific terms.
|
31 |
+
|
32 |
+
2. **Enhanced Embedding for Richer Contextual Understanding**:
|
33 |
+
- Biomedical text often contains long sentences, specialized terminology, and hierarchical relationships between entities (e.g., "diabetes" vs. "Type 1 diabetes"). DeBERTa’s improved embedding layer allows it to capture these nuanced relationships better than traditional transformer models, making it especially effective in understanding context-rich medical documents.
|
34 |
+
|
35 |
+
3. **Superior Performance on Downstream NLP Tasks**:
|
36 |
+
- DeBERTa consistently ranks among the top models on NLP benchmarks like GLUE and SQuAD, which is a testament to its ability to generalize across tasks. This high performance is especially beneficial for **BIOMed_NER**, where accurate recognition of subtle differences between biomedical entities can significantly enhance the quality of structured data extracted from unstructured clinical notes.
|
37 |
+
|
38 |
+
4. **Pre-trained for Optimal Transfer Learning**:
|
39 |
+
- Leveraging the "base" DeBERTaV3 variant allows us to tap into a model pre-trained on vast amounts of text, thus providing an excellent foundation for fine-tuning on domain-specific biomedical data. This pre-training, combined with the fine-tuning on the dataset, allows **BIOMed_NER** to accurately distinct biomedical entities, from diseases and medications to clinical events and anatomical structures.
|
40 |
+
|
41 |
+
5. **Efficient Fine-Tuning for Large Biomedical Datasets**:
|
42 |
+
- DeBERTa is optimized for both accuracy and efficiency, making it easier to train on large and complex datasets without needing excessive computational resources. This means faster iterations during model development and a more accessible deployment pipeline.
|
43 |
+
|
44 |
+
By selecting DeBERTa for **BIOMed_NER**, we've built a model that excels in understanding the intricate language of medicine, providing high accuracy and contextual depth essential for healthcare applications. Whether for researchers analyzing clinical data or applications structuring patient records, DeBERTa enables **BIOMed_NER** to extract, tag, and organize critical medical information effectively.
|
45 |
**Hyperparameters:**
|
46 |
- **Base Model**: `microsoft/deberta-v3-base`
|
47 |
- **Learning Rate**: `3e-5`
|