
Nicolay Rusnachenko
AI & ML interests
Recent Activity
Organizations
nicolay-r's activity


mistralai/Mistral-Small-3.1-24B-Instruct-2503

π€ https://github.com/nicolay-r/nlp-thirdgate/blob/master/llm/transformers_phi4.py
π https://github.com/nicolay-r/nlp-thirdgate/blob/master/tutorials/llm_phi4.py
Findings on adaptation: I was able to reproduce only the pipeline based model launching. This version is for textual llm only. Microsoft also released multimodal Phi-4 which is out of scope of this wrapper.
π nlp-thirdgate: https://lnkd.in/ef-wBnNn

https://github.com/nicolay-r/bulk-chain/releases/tag/0.25.2
π§ Fixes:
- Fixed issues with batching mode
- Fixed problem with parsing and passing args in shell mode
β οΈ Limitation: bathing mode is still available only via API.
π Quick Start with Gemma-3 in batching mode: https://github.com/nicolay-r/nlp-thirdgate/blob/master/tutorials/llm_gemma_3.ipynb

The important comment is to use the very latest version of the bulk-chain from github which fixes the bug for double-inference in batching.

https://github.com/nicolay-r/nlp-thirdgate/blob/master/tutorials/llm_gemma_3.ipynb
Limitation: schema supports texts only (for now), while gemma-3 is a text+image to text.
Model: google/gemma-3-1b-it
Provider: https://github.com/nicolay-r/nlp-thirdgate/blob/master/llm/transformers_gemma3.py

This makes it particularly mysterious what went into QwQ-32B? Why did it work so well? Was it trained from scratch? Anyone has insights about this?
onekq-ai/WebApp1K-models-leaderboard
@ritvik77 , sounds good on your plans! Meanwhile looking forward to adapt 7B version to experiment in radiology domain. Happy to read more on that and once and if it gets to the paper, so I can populate the survey of the related advances.
@ritvik77 , excited to run into this! Is the paper and studies behind it on arxiv or elsewhere?

π©Ί Medical Diagnosis AI Model - Powered by Mistral-7B & LoRA π
πΉ Model Overview:
Base Model: Mistral-7B (7.7 billion parameters)
Fine-Tuning Method: LoRA (Low-Rank Adaptation)
Quantization: bnb_4bit (reduces memory footprint while retaining performance)
πΉ Parameter Details:
Original Mistral-7B Parameters: 7.7 billion
LoRA Fine-Tuned Parameters: 4.48% of total model parameters (340 million) Final Merged Model Size (bnb_4bit Quantized): ~4.5GB
This can help you in making a AI agent for healthcare, if you need to finetune it for JSON function/tool calling format you can use some medical function calling dataset to again fine fine tine on it.
how to perform batch inference for llama-3.2-1B-Instruct


Code: https://github.com/Jaykef/ai-algorithms/blob/main/hybrid_normalization.ipynb
@ychen , I see. I was expecting your findings were a part of the phd program. Take your time with publications then, since it is common while at Phd. It would be great to have a paper during the masters and all the best with it!
@ychen Good luck with your studies and pleased for affecting on your advances in it. Are you on google scholar or github with personal advances in this domain?