@OmniSVG
Nicolay Rusnachenko
AI & ML interests
Recent Activity
Organizations
nicolay-r's activity

@OmniSVG

https://github.com/nicolay-r/nlp-thirdgate/blob/master/llm/replicate_104.py
bulk-chain: https://github.com/nicolay-r/bulk-chain-shell
Model: meta-llama/Llama-4-Maverick-17B-128E-Original

https://github.com/nicolay-r/bulk-chain/releases/tag/0.25.3
The latest release brings huge updates on:
β Reforged mechanism of models inference that work in steraming mode.
- Callbacks support for streaming mode (earlier only in demo)
- Deployment of various clients (shell, tksheet; see attachment)
β Support for batching (earlier in API mode only)
β Optional caching of inferred data in SQlite (always enabled earlier)
- This now makes possible to faster launch small (but mighty) LLMs
π Project: https://github.com/nicolay-r/bulk-chain
π Proviers: https://github.com/nicolay-r/nlp-thirdgate

Model: NX-AI/xLSTM-7b
Paper: https://arxiv.org/abs/2503.13427

π€ https://github.com/nicolay-r/nlp-thirdgate/blob/master/llm/transformers_phi4.py
π https://github.com/nicolay-r/nlp-thirdgate/blob/master/tutorials/llm_phi4.py
Findings on adaptation: I was able to reproduce only the pipeline based model launching. This version is for textual llm only. Microsoft also released multimodal Phi-4 which is out of scope of this wrapper.
π nlp-thirdgate: https://lnkd.in/ef-wBnNn

https://github.com/nicolay-r/bulk-chain/releases/tag/0.25.2
π§ Fixes:
- Fixed issues with batching mode
- Fixed problem with parsing and passing args in shell mode
β οΈ Limitation: bathing mode is still available only via API.
π Quick Start with Gemma-3 in batching mode: https://github.com/nicolay-r/nlp-thirdgate/blob/master/tutorials/llm_gemma_3.ipynb

The important comment is to use the very latest version of the bulk-chain from github which fixes the bug for double-inference in batching.

https://github.com/nicolay-r/nlp-thirdgate/blob/master/tutorials/llm_gemma_3.ipynb
Limitation: schema supports texts only (for now), while gemma-3 is a text+image to text.
Model: google/gemma-3-1b-it
Provider: https://github.com/nicolay-r/nlp-thirdgate/blob/master/llm/transformers_gemma3.py

This makes it particularly mysterious what went into QwQ-32B? Why did it work so well? Was it trained from scratch? Anyone has insights about this?
onekq-ai/WebApp1K-models-leaderboard
@ritvik77 , sounds good on your plans! Meanwhile looking forward to adapt 7B version to experiment in radiology domain. Happy to read more on that and once and if it gets to the paper, so I can populate the survey of the related advances.
@ritvik77 , excited to run into this! Is the paper and studies behind it on arxiv or elsewhere?

π©Ί Medical Diagnosis AI Model - Powered by Mistral-7B & LoRA π
πΉ Model Overview:
Base Model: Mistral-7B (7.7 billion parameters)
Fine-Tuning Method: LoRA (Low-Rank Adaptation)
Quantization: bnb_4bit (reduces memory footprint while retaining performance)
πΉ Parameter Details:
Original Mistral-7B Parameters: 7.7 billion
LoRA Fine-Tuned Parameters: 4.48% of total model parameters (340 million) Final Merged Model Size (bnb_4bit Quantized): ~4.5GB
This can help you in making a AI agent for healthcare, if you need to finetune it for JSON function/tool calling format you can use some medical function calling dataset to again fine fine tine on it.


Code: https://github.com/Jaykef/ai-algorithms/blob/main/hybrid_normalization.ipynb
@ychen , I see. I was expecting your findings were a part of the phd program. Take your time with publications then, since it is common while at Phd. It would be great to have a paper during the masters and all the best with it!
@ychen Good luck with your studies and pleased for affecting on your advances in it. Are you on google scholar or github with personal advances in this domain?

https://gist.github.com/nicolay-r/c8cfe7df1bef0c14f77760fa78ae5b5c
Why it might be intersted to check? The provided supports batching mode for a quck inference. In the case of Flan-T5-base that would be the quickest option via LLM.
π Evaluation results are available in model card:
nicolay-r/flan-t5-emotion-cause-thor-base

This is a part of the most recent release of the bulk-chain 0.25.0.
β https://github.com/nicolay-r/bulk-chain/releases/tag/0.25.1
How it works: it launches your CoT by asking missed parameters if necessary. For each item of the chain you receive input prompt and streamed output of your LLM.
To settle onto certain parameters, you can pass them via
--src
:- TXT files (using filename as a parameter name)
- JSON dictionaries for multiple
π€ Model: meta-llama/Llama-3.3-70B-Instruct
π Other models: https://github.com/nicolay-r/nlp-thirdgate
And to clarify your findings on those words you can measure such degree with tf-idf application for your annotated texts. Basically, if you have a set of positive and negative responses from GPT-4o, you can calculate so-called Semantic Orientation (SO) based on Pointwise Mutual Information (PMI). This would give a consistecy to your observations.
This comes from the relatively old classics: https://arxiv.org/pdf/cs/0212032