Baichuan-M3-235B
From Inquiry to Decision: Building Trustworthy Medical AI
π₯ Experience AI-Powered Medical Inquiry: ying.ai
π Model Overview
Baichuan-M3 is Baichuan AI's new-generation medical-enhanced large language model, a major milestone following Baichuan-M2.
In contrast to prior approaches that primarily focus on static question answering or superficial role-playing, Baichuan-M3 is trained to explicitly model the clinical decision-making process, aiming to improve usability and reliability in real-world medical practice. Rather than merely producing "plausible-sounding answers" or high-frequency vague recommendations like "you should see a doctor soon," the model is trained to proactively acquire critical clinical information, construct coherent medical reasoning pathways, and systematically constrain hallucination-prone behaviors.
Core Highlights
- π Surpasses GPT-5.2: Outperforms OpenAI's latest model across HealthBench, HealthBench-Hard, hallucination evaluation, and SCAN-bench, establishing a new SOTA in medical AI
- π©Ί High-Fidelity Clinical Inquiry: The only model to rank first across all three SCAN-bench dimensionsβClinical Inquiry, Laboratory Testing, and Diagnosis
- π§ Low Hallucination, High Reliability: Achieves lower hallucination rate than GPT-5.2 through Fact-Aware RL, even without external tools
- β‘ Efficient Deployment: W4 quantization reduces memory to 26% of original; Gated Eagle3 speculative decoding achieves 96% speedup
π Performance
HealthBench & Hallucination Evaluation
HealthBench is OpenAI's authoritative medical benchmark, constructed by 262 practicing physicians from 60 countries, comprising 5,000 high-fidelity multi-turn clinical conversations.
Compared to Baichuan-M2, Baichuan-M3 improves by 28 percentage points on HealthBench-Hard, reaching 44.4 and surpassing GPT-5.2. It also ranks first on the HealthBench Total leaderboard.
For hallucination evaluation, we decompose long-form responses into fine-grained, verifiable atomic medical claims and validate each against authoritative medical evidence. Even in a tool-free setting, Baichuan-M3 achieves lower hallucination rate than GPT-5.2.
SCAN-bench Evaluation
SCAN-bench is our end-to-end clinical decision-making benchmark that simulates the complete clinical workflow from patient encounter to final diagnosis, evaluating models' high-fidelity clinical inquiry capabilities through three stations: History Taking, Ancillary Investigations, and Final Diagnosis.
Baichuan-M3 ranks first across all three core dimensions, outperforming the second-best model by 12.4 points in Clinical Inquiry.
π’ The SCAN-bench will be open-sourced soon. Stay tuned.
π¬ Technical Features
π For detailed technical information, please refer to: Tech Blog
SPAR: Segmented Pipeline Reinforcement Learning
To address reward sparsity and credit assignment challenges in long clinical interactions, we propose SPAR (Step-Penalized Advantage with Relative baseline): it decomposes clinical workflows into four stagesβhistory taking, differential diagnosis, laboratory testing, and final diagnosisβeach with independent rewards, combined with process-level rewards for precise credit assignment, driving the model to construct auditable and complete decision logic.
Fact-Aware Reinforcement Learning
By integrating factual verification directly into the RL loop, we build an online hallucination detection module that validates model-generated medical claims against authoritative medical evidence in real-time, supported by efficient caching mechanisms for online RL training. A dynamic reward aggregation strategy adaptively balances task learning and factual constraints based on the model's capability stage, significantly enhancing medical factual reliability without sacrificing reasoning depth.
Efficient Training and Inference
Adopts a three-stage multi-expert fusion training paradigm (Domain-Specific RL β Offline Distillation β MOPD), combined with Gated Eagle3 speculative decoding (96% speedup) and W4 quantization (only 26% memory) for efficient deployment.
π§ Quick Start
from transformers import AutoTokenizer, AutoModelForCausalLM
model = AutoModelForCausalLM.from_pretrained("baichuan-inc/Baichuan-M3-235B", trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained("baichuan-inc/Baichuan-M3-235B")
messages = [{"role": "user", "content": "I've been having headaches lately, especially worse in the afternoon. What should I do?"}]
text = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True,
thinking_mode='on'
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
generated_ids = model.generate(
**model_inputs,
max_new_tokens=32768,
temperature=0.6
)
response = tokenizer.decode(generated_ids[0][len(model_inputs.input_ids[0]):], skip_special_tokens=True)
print(response)
Deployment
Create an OpenAI-compatible API endpoint using sglang>=0.4.6.post1 or vllm>=0.9.0:
# SGLang
python -m sglang.launch_server --model-path baichuan-inc/Baichuan-M3-235B --reasoning-parser qwen3
# vLLM
vllm serve baichuan-inc/Baichuan-M3-235B --reasoning-parser qwen3
Speculative Decoding with SGlang (>=0.5.5.post3)
- To support the gated attention in eagle3 draft model, just replace the code llama_eagle3.py in the sglang installation directory with our draft/llama_eagle3.py, for example:
cp -f /path/to/draft/llama_eagle3.py /sgl-workspace/sglang/python/sglang/srt/models/
- Launch sglang (taking 8 * H20(96G) deployment as an example):
python3 -m sglang.launch_server \
--model-path baichuan-inc/Baichuan-M3-235B \
--tensor-parallel-size 8 \
--trust-remote-code \
--mem-fraction-static 0.8 \
--host 0.0.0.0 \
--port 80 \
--speculative-algorithm EAGLE3 \
--speculative-draft-model-path baichuan-inc/Baichuan-M3-235B/draft \
--speculative-num-steps 5 \
--speculative-eagle-topk 8 \
--speculative-num-draft-tokens 32 \
--reasoning-parser qwen3
β οΈ Usage Notices
- Medical Disclaimer: For research and reference only; cannot replace professional medical diagnosis or treatment
- Intended Use Cases: Medical education, health consultation, clinical decision support
- Safe Use: Recommended under guidance of medical professionals
π License
Licensed under the Apache License 2.0. Research and commercial use permitted.
π€ Acknowledgements
- Base Model: Qwen3
- Training Framework: verl
- Inference Engines: vLLM, SGLang
Thank you to the open-source community. We commit to continuous contribution and advancement of healthcare AI.
π Contact Us
- Official Website: Baichuan AI
- Technical Support: GitHub
Advancing Medical AI from "Answering Correctly" to "Supporting Decisions"
π Citation
@misc{baichuan-m3,
title={Baichuan-M3: Modeling Clinical Inquiry for Reliable Medical Decision-Making},
author={Baichuan M3 Team},
year={2025},
url={https://github.com/baichuan-inc/Baichuan-M3-235B},
}
- Downloads last month
- 849