Baichuan-M3-235B

License Hugging Face M3 GPTQ-4bit Tech Blog

From Inquiry to Decision: Building Trustworthy Medical AI

πŸ₯ Experience AI-Powered Medical Inquiry: ying.ai

🌟 Model Overview

Baichuan-M3 is Baichuan AI's new-generation medical-enhanced large language model, a major milestone following Baichuan-M2.

In contrast to prior approaches that primarily focus on static question answering or superficial role-playing, Baichuan-M3 is trained to explicitly model the clinical decision-making process, aiming to improve usability and reliability in real-world medical practice. Rather than merely producing "plausible-sounding answers" or high-frequency vague recommendations like "you should see a doctor soon," the model is trained to proactively acquire critical clinical information, construct coherent medical reasoning pathways, and systematically constrain hallucination-prone behaviors.

Core Highlights

  • πŸ† Surpasses GPT-5.2: Outperforms OpenAI's latest model across HealthBench, HealthBench-Hard, hallucination evaluation, and SCAN-bench, establishing a new SOTA in medical AI
  • 🩺 High-Fidelity Clinical Inquiry: The only model to rank first across all three SCAN-bench dimensionsβ€”Clinical Inquiry, Laboratory Testing, and Diagnosis
  • 🧠 Low Hallucination, High Reliability: Achieves lower hallucination rate than GPT-5.2 through Fact-Aware RL, even without external tools
  • ⚑ Efficient Deployment: W4 quantization reduces memory to 26% of original; Gated Eagle3 speculative decoding achieves 96% speedup

πŸ“Š Performance

HealthBench & Hallucination Evaluation

HealthBench is OpenAI's authoritative medical benchmark, constructed by 262 practicing physicians from 60 countries, comprising 5,000 high-fidelity multi-turn clinical conversations.

HealthBench Performance

Compared to Baichuan-M2, Baichuan-M3 improves by 28 percentage points on HealthBench-Hard, reaching 44.4 and surpassing GPT-5.2. It also ranks first on the HealthBench Total leaderboard.

For hallucination evaluation, we decompose long-form responses into fine-grained, verifiable atomic medical claims and validate each against authoritative medical evidence. Even in a tool-free setting, Baichuan-M3 achieves lower hallucination rate than GPT-5.2.

SCAN-bench Evaluation

SCAN-bench is our end-to-end clinical decision-making benchmark that simulates the complete clinical workflow from patient encounter to final diagnosis, evaluating models' high-fidelity clinical inquiry capabilities through three stations: History Taking, Ancillary Investigations, and Final Diagnosis.

SCAN-bench Performance

Baichuan-M3 ranks first across all three core dimensions, outperforming the second-best model by 12.4 points in Clinical Inquiry.

πŸ“’ The SCAN-bench will be open-sourced soon. Stay tuned.

πŸ”¬ Technical Features

πŸ“– For detailed technical information, please refer to: Tech Blog

SPAR: Segmented Pipeline Reinforcement Learning

To address reward sparsity and credit assignment challenges in long clinical interactions, we propose SPAR (Step-Penalized Advantage with Relative baseline): it decomposes clinical workflows into four stagesβ€”history taking, differential diagnosis, laboratory testing, and final diagnosisβ€”each with independent rewards, combined with process-level rewards for precise credit assignment, driving the model to construct auditable and complete decision logic.

Fact-Aware Reinforcement Learning

By integrating factual verification directly into the RL loop, we build an online hallucination detection module that validates model-generated medical claims against authoritative medical evidence in real-time, supported by efficient caching mechanisms for online RL training. A dynamic reward aggregation strategy adaptively balances task learning and factual constraints based on the model's capability stage, significantly enhancing medical factual reliability without sacrificing reasoning depth.

Efficient Training and Inference

Adopts a three-stage multi-expert fusion training paradigm (Domain-Specific RL β†’ Offline Distillation β†’ MOPD), combined with Gated Eagle3 speculative decoding (96% speedup) and W4 quantization (only 26% memory) for efficient deployment.

πŸ”§ Quick Start

from transformers import AutoTokenizer, AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained("baichuan-inc/Baichuan-M3-235B", trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained("baichuan-inc/Baichuan-M3-235B")

messages = [{"role": "user", "content": "I've been having headaches lately, especially worse in the afternoon. What should I do?"}]
text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True,
    thinking_mode='on'
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)

generated_ids = model.generate(
    **model_inputs,
    max_new_tokens=32768,
    temperature=0.6
)
response = tokenizer.decode(generated_ids[0][len(model_inputs.input_ids[0]):], skip_special_tokens=True)
print(response)

Deployment

Create an OpenAI-compatible API endpoint using sglang>=0.4.6.post1 or vllm>=0.9.0:

# SGLang
python -m sglang.launch_server --model-path baichuan-inc/Baichuan-M3-235B --reasoning-parser qwen3

# vLLM
vllm serve baichuan-inc/Baichuan-M3-235B --reasoning-parser qwen3

Speculative Decoding with SGlang (>=0.5.5.post3)

  1. To support the gated attention in eagle3 draft model, just replace the code llama_eagle3.py in the sglang installation directory with our draft/llama_eagle3.py, for example:
cp -f /path/to/draft/llama_eagle3.py /sgl-workspace/sglang/python/sglang/srt/models/
  1. Launch sglang (taking 8 * H20(96G) deployment as an example):
python3 -m sglang.launch_server \
   --model-path baichuan-inc/Baichuan-M3-235B \
   --tensor-parallel-size 8 \
   --trust-remote-code \
   --mem-fraction-static 0.8 \
   --host 0.0.0.0 \
   --port 80 \
   --speculative-algorithm EAGLE3 \
   --speculative-draft-model-path baichuan-inc/Baichuan-M3-235B/draft \
   --speculative-num-steps 5 \
   --speculative-eagle-topk 8 \
   --speculative-num-draft-tokens 32 \
   --reasoning-parser qwen3

⚠️ Usage Notices

  1. Medical Disclaimer: For research and reference only; cannot replace professional medical diagnosis or treatment
  2. Intended Use Cases: Medical education, health consultation, clinical decision support
  3. Safe Use: Recommended under guidance of medical professionals

πŸ“„ License

Licensed under the Apache License 2.0. Research and commercial use permitted.

🀝 Acknowledgements

  • Base Model: Qwen3
  • Training Framework: verl
  • Inference Engines: vLLM, SGLang

Thank you to the open-source community. We commit to continuous contribution and advancement of healthcare AI.

πŸ“ž Contact Us

Advancing Medical AI from "Answering Correctly" to "Supporting Decisions"

πŸ“š Citation

@misc{baichuan-m3,
    title={Baichuan-M3: Modeling Clinical Inquiry for Reliable Medical Decision-Making},
    author={Baichuan M3 Team},
    year={2025},
    url={https://github.com/baichuan-inc/Baichuan-M3-235B},
}
Downloads last month
849
Safetensors
Model size
235B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ 4 Ask for provider support

Model tree for baichuan-inc/Baichuan-M3-235B

Finetuned
(33)
this model
Quantizations
6 models

Space using baichuan-inc/Baichuan-M3-235B 1

Collection including baichuan-inc/Baichuan-M3-235B