Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
48.7
TFLOPS
7
23
40
Nicolay Rusnachenko
nicolay-r
Follow
Luis-fer's profile picture
alessio4552's profile picture
yeiad07's profile picture
210 followers
·
9 following
https://nicolayr.com/
nicolayr_
nicolay-r
nicolay-r
nicolayr.com
AI & ML interests
Information Retrieval・Medical Multimodal NLP (🖼+📝) Research Fellow @BU_Research・software developer http://arekit.io・PhD in NLP
Recent Activity
posted
an
update
2 days ago
🚀 For those who interested in multilingual clinical case report sukmmarization 🩺📋, deligned to share a video-update to the earlier post on Qwen2.5 model family adaptation: 🎬 Video: https://www.youtube.com/watch?v=uOAiUvLghuE This is 15-min skimming of the study (+ 5 mins for code) in which we overview the application of Qwen model family (72B as a teacher and 0.5B as a student) in summarization of the clinical reports, including detaied overview of the experiments organization. In particular, attempted to cover: 1. Background of previous Seq2Seq models to conclude their limitations 2. ChatML roles exploiting for distilation tuning in clinical report summarization 3. Known limitation of work and unleashing full capabilities As in previous post, there is a model card that is also covered in video. 🤗 Huggingface: https://huggingface.co/nicolay-r/qwen25-05b-multiclinsum-standar
replied
to
their
post
3 days ago
🚀 For those who interested in summarization of the long textual reports in medical domain 📝🩺, @Xiaolihai and I delighted to share that we experiment with distillation tuning adaptation for Qwen-2.5 0.5B. We use reports from the MultiClinSum dataset and pass it through 72B version to retrieve report explanations in order to initiate ditillation tuning for 0.5B model. We experiment with passages written in English, French, Portuguese, and Spanish. 🔑 We find that using distil-technique results in 2-4% performance increment on fine-tuning and similar improvements for reports in English (non-official and official evaluation). For the other it results in systems that perform similar to the convential tuning (standard) (see result below). Dataset: https://zenodo.org/records/15459174 Competition: https://participants-area.bioasq.org/general_information/MultiClinSum/ Github: https://github.com/nicolay-r/distil-tuning-llm model: https://huggingface.co/nicolay-r/qwen25-05b-multiclinsum-distil
updated
a model
3 days ago
nicolay-r/qwen25-05b-multiclinsum-standard
View all activity
Organizations
None yet
nicolay-r
's models
7
Sort: Recently updated
nicolay-r/qwen25-05b-multiclinsum-standard
Text Generation
•
0.5B
•
Updated
3 days ago
•
33
nicolay-r/qwen25-05b-multiclinsum-distil
Text Generation
•
0.5B
•
Updated
3 days ago
•
48
•
•
1
nicolay-r/flan-t5-tsa-thor-large
Text Generation
•
0.8B
•
Updated
Feb 23
•
21
nicolay-r/flan-t5-emotion-cause-thor-base
Text Generation
•
0.2B
•
Updated
Feb 23
•
218
•
1
nicolay-r/flan-t5-tsa-thor-xl
Text Generation
•
3B
•
Updated
Feb 23
•
13
•
3
nicolay-r/flan-t5-tsa-prompt-xl
Text Generation
•
3B
•
Updated
Feb 23
•
14
•
1
nicolay-r/flan-t5-tsa-thor-base
Text Generation
•
0.2B
•
Updated
Feb 23
•
16
•
1