
Nicolay Rusnachenko
AI & ML interests
Recent Activity
Organizations


nicolay-r/flan-t5-tsa-prompt-xl

nicolay-r/flan-t5-tsa-thor-xl

nicolay-r/flan-t5-tsa-thor-large

nicolay-r/flan-t5-tsa-thor-base

Excited to share the ultimate guide for dealing with ATAS refusals and how to write effective rebuttal letters.
π¬ https://youtu.be/bfknM3n-SHs
π From the video you will find:
1. Why appealing an ATAS decision matters even if your visa is approved
2. Which docments to use in understanding the principles behind sponsorship decisions
3. Key tips for proper rebuttal letter structuring

Sounds good, all the best in this route!

here are 4 types of context failure and why each one happens:
1. context poisoning: if hallucination finds its way into your context, the agent will rely on that false information to make its future moves. for example if the agent hallucinates about the "task description", all of its planning to solve the task would also be corrupt.
2. context distraction: when the context becomes too bloated, the model focuses too much on it rather than come up with novel ideas or to follow what it has learned during training. as Gemini 2.5 Pro technical report points out, as context grows significantly from 100K tokens, "the agent showed a tendency toward favoring repeating actions from its vast history rather than synthesizing novel plans".
3. context confusion: everyone lost it when MCPs became popular, it seemed like AGI was achieved. I suspected there is something wrong and there was: it's not just about providing tools, bloating the context with tool use derails the model from selecting the right one! even if you can fit all your tool metadata in the context, as their number grows, the model gets confused over which one to pick.
4. Context Clash: if you exchange conversation with a model step by step and provide information as you go along, chances are you get worse performance rather than providing all the useful information at once. one the model's context fills with wrong information, it's more difficult to guide it to embrace the right info. agents pull information from tools, documents, user queries, etc. and there is a chance that some of these information contradict each other, and it's not good new for agentic applications.
check this article by Drew Breunig for deeper read: https://www.dbreunig.com/2025/06/26/how-to-fix-your-context.html?ref=blog.langchain.com

Thanks for that grouped findings!
In addition, lost in the middle would fitted around 1-3 for sure:
https://arxiv.org/abs/2307.03172

Title:
π§ β―Understanding the Perceptual-Interface Protocol: Structured Sensory Modules for Cognitive Input Parsing
πβ―Read it here: https://huggingface.co/blog/kanaria007/understanding-the-perceptual-interface-protocol
---
Summary:
What if artificial intelligence systems could βsenseβ inputs β not physically, but structurally?
This article introduces the *Five Sense Protocol*, a *theoretical blueprint* for embedding cognitive input structuring into AI systems.
Inspired by human perceptual organization, it proposes *five abstract sensory layers* to refine reasoning through structured perception.
---
Why It Matters:
Current LLMs treat input as undifferentiated text streams. But humans donβt.
We segment, anticipate, detect contradiction, monitor coherence, and adjust ethically β *before* reasoning begins.
The Perceptual-Interface protocol brings this pre-reasoning perceptual organization into AI cognition design.
---
Core Layers:
β’ Surface Syntax Detection β using formatting as reasoning cues
β’ Temporal Rhythm Tracking β anticipating structural pacing
β’ Conflict Sensitivity β flagging contradictions and dissonance
β’ Structural Coherence Mapping β detecting fragmented logic
β’ Ethical Context Filters β triggering meta-awareness in sensitive domains
---
π§© *Think of it as structured input for structured output.*
Reasoning is only as good as the way it begins.
---
Relevant For:
β’ AI researchers exploring perceptual architecture in cognition
β’ Developers building input-sensitive autonomous systems
β’ Cognitive scientists bridging human and artificial attention models
---
π§ Protocol Dataset: kanaria007/agi-structural-intelligence-protocols
---
*This isnβt perception emulation.*
Itβs *perception structuring* β for the next layer of intelligent reasoning.

What a detailed and vebose update! I see, there is a lot of happening here both on a side of the architectural solutions, organization of the fine-tuning, related experiments. My comment on that would be very broad as willing to have a closer look when highlighted ideas would be supported by the related tables as an outcome of experiments. At first glance, the idea of generatlized pre-training following by domain specific fine-tuning sounds quite promising. Besides that, curious if that DHM model and other cliincal-tasks related models would be publicly available

True, especially with NER as a task, BERT fits too well both from the side of the required resources and annotation accuracy.

OpenMed is back! π₯ Finding the right biomedical NER model just became as precise as a PCR assay!
I'm thrilled to unveil my comprehensive OpenMed Named Entity Recognition Model Discovery App that puts 384 specialized biomedical AI models at your fingertips.
π― Why This Matters in Healthcare AI:
Traditional clinical text mining required hours of manual model evaluation. My Discovery App instantly connects researchers, clinicians, and data scientists with the exact NER models they need for their biomedical entity extraction tasks.
π¬ What You Can Discover:
β Pharmacological Models - Extract "chemical compounds", "drug interactions", and "pharmaceutical" entities from clinical notes
β Genomics & Proteomics - Identify "DNA sequences", "RNA transcripts", "gene variants", "protein complexes", and "cell lines"
β Pathology & Disease Detection - Recognize "pathological formations", "cancer types", and "disease entities" in medical literature
β Anatomical Recognition - Map "anatomical systems", "tissue types", "organ structures", and "cellular components"
β Clinical Entity Extraction - Detect "organism species", "amino acids", 'protein families", and "multi-tissue structures"
π‘ Advanced Features:
π Intelligent Entity Search - Find models by specific biomedical entities (e.g., "Show me models detecting CHEM + DNA + Protein")
π₯ Domain-Specific Filtering - Browse by Oncology, Pharmacology, Genomics, Pathology, Hematology, and more
π Model Architecture Insights - Compare BERT, RoBERTa, and DeBERTa implementations
β‘ Real-Time Search - Auto-filtering as you type, no search buttons needed
π¨ Clinical-Grade UI - Beautiful, intuitive interface designed for medical professionals
Ready to revolutionize your biomedical NLP pipeline?
π Try it now: OpenMed/openmed-ner-models
𧬠Built with: Gradio, Transformers, Advanced Entity Mapping

Pleased for the lightweight architecture implementaiton and filtering feature.
Thank you for sharing this! π

Well done to Healthcare contribution and to field of ECG data processing in particular. π
Looks like base model mention a lot of concepts (CNN, ViT, GNN) that could individually represent a solution for that propblem or form a processing pipeline. Would it be possible to ask on having a look on the framework diagram / preprint studies for a greater detail?

I am excited to announce the launch of research on the Digital Heart Model (DHM), an AI-driven digital twin designed to transform personalized cardiovascular care. DHM integrates multimodal data, focusing initially on cardiac imaging, histopathological imaging, and ECG data, to predict patient outcomes and optimize interventions.
Initial Model and Dataset Overview:
Base Model: Multimodal AI foundation combining Convolutional Neural Networks (CNN), Vision Transformers (ViT), and Graph Neural Networks (GNN).
Datasets: Cardiac MRI and CT imaging datasets, histopathological cardiac tissue images, and extensive ECG waveform data.
Expected Results from First Iteration:
Cardiac event prediction (e.g., myocardial infarction) accuracy: AUC β₯ 0.90
Arrhythmia detection and classification accuracy: AUC β₯ 0.88
Enhanced segmentation accuracy for cardiac imaging: Dice Score β₯ 0.85
π Next Steps:
Conducting initial retrospective validation.
Preparing for prospective clinical validation.
Stay tuned for updates as we redefine cardiovascular precision medicine!
Connect with us for collaboration and insights!


π¬ Video: https://www.youtube.com/watch?v=uOAiUvLghuE
This is 15-min skimming of the study (+ 5 mins for code) in which we overview the application of Qwen model family (72B as a teacher and 0.5B as a student) in summarization of the clinical reports, including detaied overview of the experiments organization. In particular, attempted to cover:
1. Background of previous Seq2Seq models to conclude their limitations
2. ChatML roles exploiting for distilation tuning in clinical report summarization
3. Known limitation of work and unleashing full capabilities
As in previous post, there is a model card that is also covered in video.
π€ Huggingface: https://huggingface.co/nicolay-r/qwen25-05b-multiclinsum-standar

@Doctor-Chad-PhD , there is a definitely a room of further improvements, including adaptation of the larger instance of the Qwen and other models. Although such details goes beyound the experiments, I might still navigate you to more detailed overview, in which my observations on how to use this framework at larger scale:
https://www.youtube.com/watch?v=uOAiUvLghuE&t=960s
Spoiler: 1οΈβ£ necessity for more coplex summaries evaluation and 2οΈβ£ more precise declaration for the extracted clinical key reports.
Finally, organizers of the workshop still keep unpublished results of the other systems (planned toπ by the end of August), so and probably the leaderboard of other systems might sort out your question.