Citrus: Leveraging Expert Cognitive Pathways in a Medical Language Model for Advanced Medical Decision Support

📑Paper🤗Github Page🤗Model📚Medical Reasoning Data📚Evaluation Data

The Introduction to Our Work

1. Main approaches

image

2. Overview of training stages and training data pipeline

Citrus is a medical language model that bridges the gap between clinical expertise and AI reasoning by emulating the cognitive processes of medical experts. The model is trained on a large corpus of simulated expert disease reasoning data in sft-stage-3, synthesized using a novel approach that accurately captures the decision-making pathways of clinicians.

The contributions of this work are as follows:

  1. We propose a training-free reasoning approach that emulates the cognitive processes of medical experts, enabling large language models to enhance their medical capabilities in clinical diagnosis and treatment.

  2. In conjunction with the data construction method, we introduce a multi-stage post-training approach to further improve the model’s medical performance.

  3. We have made the Citrus model and its training data publicly available as open-source resources to advance research in AI-driven medical decision-making.

  4. We have developed and open-sourced a large-scale, updatable clinical practice evaluation dataset based on real-world data, accurately reflecting the distribution of patients in real-world settings.

Notice

  1. Our model is built with Llama3.1-70B, Llama 3.1 is licensed under the Llama 3.1 Community License, Copyright © Meta Platforms, Inc. All Rights Reserved.
  2. Our default license is MIT, provided it does not conflict with the Llama license.
Downloads last month
348
Safetensors
Model size
70.6B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for jdh-algo/Citrus1.0-llama-70B

Quantizations
2 models