๐ŸŒŸ Qwen3.5-2B-Claude-4.6-Opus-Reasoning-Distilled

๐Ÿ“ข Announcement

Update: This model has been further enhanced with additional reasoning data distilled from Qwen3.5-27B.

The new training data introduces higher-quality reasoning trajectories across domains such as science, instruction-following, and mathematics.

Part of the data comes from Jackrong/Qwen3.5-reasoning-700x, a curated dataset designed to improve structured step-by-step reasoning and reasoning diversity.

HCaJnUQaoAAaMIc

๐Ÿ’ก Model Introduction

Qwen3.5-2B-Claude-4.6-Opus-Reasoning-Distilled is a highly capable reasoning model fine-tuned on top of the Qwen3.5-2B dense architecture. The model's core directive is to leverage state-of-the-art Chain-of-Thought (CoT) distillation primarily sourced from Claude-4.6 Opus interactions.

Through Supervised Fine-Tuning (SFT) focusing specifically on structured reasoning logic, this model excels in breaking down complex user problems, planning step-by-step methodologies within strictly formatted <think> tags, and ultimately delivering precise, nuanced solutions.

๐Ÿ—บ๏ธ Training Pipeline Overview

Base Model (Qwen3.5-2B)
 โ”‚
 โ–ผ
Supervised Fine-Tuning (SFT) + LoRA
(Response-Only Training masked on "<|im_start|>assistant\n<think>")
 โ”‚
 โ–ผ
Final Model Text Only (Qwen3.5-2B-Claude-4.6-Opus-Reasoning-Distilled)

๐Ÿง  Example of Learned Reasoning Scaffold๏ผˆExample๏ผ‰

The model includes targeted optimizations addressing Qwen3.5โ€™s tendency toward excessive transitional or repetitive reasoning on simple queries. Through deep distillation and structural imitation of Claude-4.6-Opus reasoning chains, the model adopts a more efficient structured thinking pattern:
โ€œLet me analyze this request carefully: 1..2..3...โ€.
This streamlined reasoning paradigm significantly reduces redundant cognitive loops while preserving deep analytical capacity, resulting in substantially improved inference efficiency.

Let me analyze this request carefully:

1. Identify the core objective of the problem.
2. Break the task into clearly defined subcomponents.
3. Evaluate constraints and edge cases.
4. Formulate a step-by-step solution plan.
5. Execute the reasoning sequentially and verify consistency.
            .
            .
            .

๐Ÿ”น Supervised Fine-Tuning (SFT)

  • Objective: To inject high-density reasoning logic and establish a strict format for problem-solving involving an internal thinking state prior to outputting the final response.
  • Method: We utilized Unsloth for highly efficient memory and compute optimization. A critical component of this stage is the train_on_responses_only strategy, masking instructions so the loss is purely calculated over the generation of the <think> sequences and the subsequent solutions.
  • Format Enforcement: All training samples were systematically normalized so the model strictly abides by the structure <think> {internal reasoning} </think>\n {final answer}.

๐Ÿ“ˆ Training Loss Curve

The training loss showed a strong and healthy downward trend throughout the entire 3-epoch run, demonstrating effective knowledge distillation. Starting from an initial loss of 0.730115, the model converged steadily to a final loss of 0.186790 โ€” indicating the model successfully internalized the structured <think> reasoning patterns from the Claude 4.6 Opus teacher data.

๐Ÿ“š All Datasets Used

The dataset consists of high-quality, filtered reasoning distillation data:

Dataset Name Description / Purpose
nohurry/Opus-4.6-Reasoning-3000x-filtered Provides comprehensive Claude 4.6 Opus reasoning trajectories.
TeichAI/claude-4.5-opus-high-reasoning-250x Injecting high-intensity, structured reasoning instances.
Jackrong/Qwen3.5-reasoning-700x Additional curated reasoning samples designed to strengthen structured step-by-step problem solving and improve reasoning diversity.

๐ŸŒŸ Core Skills & Capabilities

  1. Modular & Structured Thinking: Inheriting traits from Opus-level reasoning, the model demonstrates confident parsing of the prompt, establishing an outlined plan in its <think> block sequentially rather than exploratory "trial-and-error" self-doubt.
  2. Extended Context Support: Fine-tuned smoothly with a 16,384 token context window allowing complex multi-step reasoning traces to exist gracefully within memory limits.

โš ๏ธ Limitations & Intended Use

  • Hallucination Risk: While reasoning is strong, the model remains an autoregressive LLM; external facts provided during the thinking sequence may occasionally contain hallucinations if verifying real-world events.
  • Intended Scenario: Best suited for offline analytical tasks, coding, math, and heavy logic-dependent prompting where the user needs to transparently follow the AI's internal logic.
  • This model is a test version intended solely for learning and demonstration purposes, and is for academic research and technical exploration use only.

๐Ÿ™ Acknowledgements

Significant thanks to the Unsloth AI team for making rapid fine-tuning of large LLM models accessible. Additionally, we acknowledge Qwen internally, and the open-source community developers producing exceptional distilled datasets (nohurry and TeichAI).

Downloads last month
8,532
GGUF
Model size
2B params
Architecture
qwen35
Hardware compatibility
Log In to add your hardware

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ 1 Ask for provider support

Collection including Jackrong/Qwen3.5-2B-Claude-4.6-Opus-Reasoning-Distilled-GGUF