Introduction
Train LLM to be neuroscientist. It expected to work in Chinese and Engliah environment.
Data
- FreedomIntelligence/Huatuo26M-Lite. We select neuroscience(神经科学) label as train data.
- CocoNutZENG/NeuroQABenchmark
Train Detail
We fine-tuned the Qwen2.5 model using supervised fine-tuning (SFT) with LoRA for efficient parameter optimization. The LoRA configuration employed a rank of 8 (R=8) to
balance adaptation quality with computational efficiency. Training was conducted for 1 epoch (approximately 1 hour duration) using two NVIDIA A40 GPUs with DeepSpeed’s Stage 2 optimization
for memory efficiency. We adopted the Adam optimizer with an initial learning rate of 5e-5 and a
cosine learning rate scheduler for smooth decay. This configuration achieved effective model adaptation while maintaining computational tractability on our hardware setup. Our model’s loss drop as
expected, see figure below for loss detail.
Evalution
Model | Acc |
---|---|
Qwen2.5-3b | 0.788 |
Qwen2.5-7b | 0.820 |
+HuatuoLite | 0.832 |
+Full data | 0.848 |
- Downloads last month
- 12