Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up
nicolay-rΒ 
posted an update 10 days ago
Post
3508
πŸš€ For those who interested in summarization of the long textual reports in medical domain πŸ“πŸ©Ί, @Xiaolihai and I delighted to share that we experiment with distillation tuning adaptation for Qwen-2.5 0.5B. We use reports from the MultiClinSum dataset and pass it through 72B version to retrieve report explanations in order to initiate ditillation tuning for 0.5B model. We experiment with passages written in English, French, Portuguese, and Spanish.

πŸ”‘ We find that using distil-technique results in 2-4% performance increment on fine-tuning and similar improvements for reports in English (non-official and official evaluation). For the other it results in systems that perform similar to the convential tuning (standard) (see result below).

Dataset: https://zenodo.org/records/15459174
Competition: https://participants-area.bioasq.org/general_information/MultiClinSum/
Github: https://github.com/nicolay-r/distil-tuning-llm
model: nicolay-r/qwen25-05b-multiclinsum-distil

Interesting, how does it compare to larger models?

Β·

@Doctor-Chad-PhD , there is a definitely a room of further improvements, including adaptation of the larger instance of the Qwen and other models. Although such details goes beyound the experiments, I might still navigate you to more detailed overview, in which my observations on how to use this framework at larger scale:

https://www.youtube.com/watch?v=uOAiUvLghuE&t=960s

Spoiler: 1️⃣ necessity for more coplex summaries evaluation and 2️⃣ more precise declaration for the extracted clinical key reports.

Finally, organizers of the workshop still keep unpublished results of the other systems (planned toπŸ”“ by the end of August), so and probably the leaderboard of other systems might sort out your question.