Files changed (1) hide show
  1. README.md +108 -97
README.md CHANGED
@@ -1,98 +1,109 @@
1
- ---
2
- license: apache-2.0
3
- language:
4
- - en
5
- - zh
6
- base_model:
7
- - Qwen/Qwen2.5-3B-Instruct
8
- tags:
9
- - medical
10
- - cancer
11
- - Onco
12
- ---
13
- # OncoCareBrain-GPT
14
-
15
- ## Model Description
16
-
17
- OncoCareBrain-GPT is a specialized large language model fine-tuned for oncology applications. Built upon the powerful Qwen2.5-3B foundation model, it has undergone supervised fine-tuning (SFT) with tens of thousands of multi-omics data samples, including genomic, pathological, and clinical data. This model is specifically designed to serve the cancer care domain with advanced reasoning capabilities.
18
-
19
- ## Key Features
20
-
21
- - **Intelligent Medical Q&A**: Quickly answers complex questions about cancer, leveraging a deep understanding of oncology concepts
22
- - **Precision Decision Support**: Recommends optimal treatment plans based on multi-dimensional data analysis
23
- - **Transparent Reasoning Process**: Generates detailed chains of thought to ensure model explainability and trust in clinical settings
24
-
25
- ## Intended Uses
26
-
27
- - **Clinical Decision Support**: Assists healthcare providers in evaluating treatment options
28
- - **Patient Education**: Helps patients better understand their condition and treatment plans
29
- - **Medical Research**: Supports researchers in analyzing cancer data and generating insights
30
-
31
- ## Training Data
32
-
33
- OncoCareBrain-GPT was fine-tuned on a diverse dataset comprising:
34
- - Genomic data
35
- - Pathological samples
36
- - Clinical records and case studies
37
-
38
- The model was trained to generate detailed reasoning chains, provide personalized prognostic assessments, and suggest evidence-based treatment recommendations.
39
-
40
- ## Technical Specifications
41
-
42
- - **Base Model**: Qwen2.5-3B
43
- - **Parameters**: 3 billion
44
- - **Training Method**: Supervised Fine-Tuning (SFT)
45
- - **Language Capabilities**: English, Chinese
46
- - **Input Format**: Natural language
47
- - **Output Format**: Detailed explanations with chain-of-thought reasoning
48
-
49
- ## Limitations
50
-
51
- - The model should be used as a clinical decision support tool and not as a replacement for professional medical judgment
52
- - Recommendations should be verified by qualified healthcare professionals
53
- - Performance may vary depending on the complexity and rarity of cancer cases
54
- - While the model supports English and Chinese, performance might vary between languages
55
-
56
- ## Ethical Considerations
57
-
58
- - **Privacy**: The model operates on input data and does not store patient information
59
- - **Bias**: While efforts have been made to minimize biases, users should be aware of potential biases in training data
60
- - **Transparency**: The model provides reasoning chains to ensure transparency in its decision-making process
61
-
62
- ## How to Use
63
-
64
- ```python
65
- # Example code for model inference
66
- from transformers import AutoTokenizer, AutoModelForCausalLM
67
-
68
- tokenizer = AutoTokenizer.from_pretrained("DXCLab/OncoCareBrain-GPT")
69
- model = AutoModelForCausalLM.from_pretrained("DXCLab/OncoCareBrain-GPT")
70
-
71
- input_text = "Could you analyze this genomic profile and suggest potential treatment options for breast cancer with BRCA1 mutation?"
72
- inputs = tokenizer(input_text, return_tensors="pt")
73
- outputs = model.generate(**inputs, max_length=1000)
74
- response = tokenizer.decode(outputs[0])
75
- print(response)
76
- ```
77
-
78
- ## Citation
79
-
80
- If you use OncoCareBrain-GPT in your research, please cite:
81
-
82
- ```
83
- @misc{OncoCareBrain-GPT,
84
- author = {DXCLab},
85
- title = {OncoCareBrain-GPT: A Specialized Language Model for Oncology},
86
- year = {2025},
87
- publisher = {Hugging Face},
88
- howpublished = {\url{https://huggingface.co/DXCLab/OncoCareBrain-GPT}}
89
- }
90
- ```
91
-
92
- ## License
93
-
94
- This model is licensed under the Apache License 2.0. See the [LICENSE](LICENSE) file for details.
95
-
96
- ## Contact
97
-
 
 
 
 
 
 
 
 
 
 
 
98
  For questions or feedback about OncoCareBrain-GPT, please visit our Hugging Face page at https://huggingface.co/DXCLab or open an issue in the repository.
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - zho
5
+ - eng
6
+ - fra
7
+ - spa
8
+ - por
9
+ - deu
10
+ - ita
11
+ - rus
12
+ - jpn
13
+ - kor
14
+ - vie
15
+ - tha
16
+ - ara
17
+ base_model:
18
+ - Qwen/Qwen2.5-3B-Instruct
19
+ tags:
20
+ - medical
21
+ - cancer
22
+ - Onco
23
+ ---
24
+ # OncoCareBrain-GPT
25
+
26
+ ## Model Description
27
+
28
+ OncoCareBrain-GPT is a specialized large language model fine-tuned for oncology applications. Built upon the powerful Qwen2.5-3B foundation model, it has undergone supervised fine-tuning (SFT) with tens of thousands of multi-omics data samples, including genomic, pathological, and clinical data. This model is specifically designed to serve the cancer care domain with advanced reasoning capabilities.
29
+
30
+ ## Key Features
31
+
32
+ - **Intelligent Medical Q&A**: Quickly answers complex questions about cancer, leveraging a deep understanding of oncology concepts
33
+ - **Precision Decision Support**: Recommends optimal treatment plans based on multi-dimensional data analysis
34
+ - **Transparent Reasoning Process**: Generates detailed chains of thought to ensure model explainability and trust in clinical settings
35
+
36
+ ## Intended Uses
37
+
38
+ - **Clinical Decision Support**: Assists healthcare providers in evaluating treatment options
39
+ - **Patient Education**: Helps patients better understand their condition and treatment plans
40
+ - **Medical Research**: Supports researchers in analyzing cancer data and generating insights
41
+
42
+ ## Training Data
43
+
44
+ OncoCareBrain-GPT was fine-tuned on a diverse dataset comprising:
45
+ - Genomic data
46
+ - Pathological samples
47
+ - Clinical records and case studies
48
+
49
+ The model was trained to generate detailed reasoning chains, provide personalized prognostic assessments, and suggest evidence-based treatment recommendations.
50
+
51
+ ## Technical Specifications
52
+
53
+ - **Base Model**: Qwen2.5-3B
54
+ - **Parameters**: 3 billion
55
+ - **Training Method**: Supervised Fine-Tuning (SFT)
56
+ - **Language Capabilities**: English, Chinese
57
+ - **Input Format**: Natural language
58
+ - **Output Format**: Detailed explanations with chain-of-thought reasoning
59
+
60
+ ## Limitations
61
+
62
+ - The model should be used as a clinical decision support tool and not as a replacement for professional medical judgment
63
+ - Recommendations should be verified by qualified healthcare professionals
64
+ - Performance may vary depending on the complexity and rarity of cancer cases
65
+ - While the model supports English and Chinese, performance might vary between languages
66
+
67
+ ## Ethical Considerations
68
+
69
+ - **Privacy**: The model operates on input data and does not store patient information
70
+ - **Bias**: While efforts have been made to minimize biases, users should be aware of potential biases in training data
71
+ - **Transparency**: The model provides reasoning chains to ensure transparency in its decision-making process
72
+
73
+ ## How to Use
74
+
75
+ ```python
76
+ # Example code for model inference
77
+ from transformers import AutoTokenizer, AutoModelForCausalLM
78
+
79
+ tokenizer = AutoTokenizer.from_pretrained("DXCLab/OncoCareBrain-GPT")
80
+ model = AutoModelForCausalLM.from_pretrained("DXCLab/OncoCareBrain-GPT")
81
+
82
+ input_text = "Could you analyze this genomic profile and suggest potential treatment options for breast cancer with BRCA1 mutation?"
83
+ inputs = tokenizer(input_text, return_tensors="pt")
84
+ outputs = model.generate(**inputs, max_length=1000)
85
+ response = tokenizer.decode(outputs[0])
86
+ print(response)
87
+ ```
88
+
89
+ ## Citation
90
+
91
+ If you use OncoCareBrain-GPT in your research, please cite:
92
+
93
+ ```
94
+ @misc{OncoCareBrain-GPT,
95
+ author = {DXCLab},
96
+ title = {OncoCareBrain-GPT: A Specialized Language Model for Oncology},
97
+ year = {2025},
98
+ publisher = {Hugging Face},
99
+ howpublished = {\url{https://huggingface.co/DXCLab/OncoCareBrain-GPT}}
100
+ }
101
+ ```
102
+
103
+ ## License
104
+
105
+ This model is licensed under the Apache License 2.0. See the [LICENSE](LICENSE) file for details.
106
+
107
+ ## Contact
108
+
109
  For questions or feedback about OncoCareBrain-GPT, please visit our Hugging Face page at https://huggingface.co/DXCLab or open an issue in the repository.