Qwen3.5-27B Claude Opus Reasoning Distilled
Community fine-tune that distills Claude Opus 4.6 reasoning into Qwen3.5-27B via LoRA. 28B parameters, Apache 2.0, no published benchmarks.

Overview
Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled is a community fine-tune by HuggingFace user Jackrong that applies Claude Opus 4.6 reasoning patterns to the Qwen3.5-27B base model. The distillation uses supervised fine-tuning with LoRA (rank 64) on roughly 3,950 reasoning trace samples created by Claude Opus 4.6.
TL;DR
- Community LoRA fine-tune of Qwen3.5-27B using Claude Opus 4.6 reasoning traces
- Uses
<think>...</think>tags for chain-of-thought output - text-only, 8K context - No published benchmarks - quality is unverified beyond community anecdotes
- Apache 2.0 licensed, 7 quantized variants available
The model outputs structured reasoning in <think> blocks before providing answers, mimicking Claude's reasoning style. The collection has racked up an estimated 57,000+ downloads across 13 variants (the GGUF version alone hit 20,500) and 134 likes, but comes with important caveats: no published benchmarks, a dramatic reduction in context window from the base model's 262K to just 8K, and loss of multimodal capabilities.
This is a preview release. The creator acknowledges potential bugs, compatibility inconsistencies, and integration edge cases. It's best understood as a research experiment in reasoning distillation, not a production-ready model.
Key Specifications
| Specification | Details |
|---|---|
| Provider | Jackrong (Community) |
| Base Model | Qwen3.5-27B |
| Architecture | Gated DeltaNet + Dense FFN (LoRA fine-tuned) |
| Total Parameters | 28B |
| LoRA Rank | 64 |
| Context Window | 8,192 tokens |
| Input Modalities | Text only |
| Output Format | <think> reasoning + final answer |
| Training Samples | ~3,950 |
| Fine-tuning Framework | Unsloth 2026.3.3 |
| Tensor Format | Safetensors (BF16, F32) |
| License | Apache 2.0 |
| Quantized Variants | 7 available |
Benchmark Performance
No benchmarks have been published. This is the most significant data gap. The model card provides qualitative descriptions of capabilities but no scores on standard evaluations.
For reference, here are the base model's known scores versus frontier reasoning models:
| Benchmark | Qwen3.5-27B (base) | Claude Opus 4.6 | Notes |
|---|---|---|---|
| SWE-bench Verified | 72.4 | 80.8 | Coding |
| MMLU-Pro | Not published | 85.3 | Knowledge |
| GPQA Diamond | Not published | 82.7 | Science reasoning |
| IFEval | 95.0 | 90.1 | Instruction following |
| LiveCodeBench | 80.7 | Not published | Live coding |
Without scores for the distilled model, there's no way to assess whether the LoRA fine-tuning improved reasoning capabilities beyond the base Qwen3.5-27B, degraded other capabilities, or had negligible effect. The training set of ~3,280 samples is small by distillation standards - DeepSeek-R1's distilled Qwen variants used 800,000 high-quality reasoning samples with full fine-tuning, orders of magnitude more data and compute.
Key Capabilities
The model is designed for chain-of-thought reasoning tasks: complex Q&A, logical analysis, step-by-step problem solving, and nuanced conversations. The <think> block output format shows the model's reasoning process before delivering a final answer.
Intended use cases include offline analytical tasks, coding assistance, math, and logic-dependent prompting. The creator specifically notes it's not suited for production deployment and recommends it for research and experimentation.
The 8K context window is a significant practical limitation. The base Qwen3.5-27B supports 262K tokens natively and up to 1M with YaRN extension. The distilled version can't handle the long documents, codebases, or extended agent conversations that the base model supports.
Pricing and Availability
The model is free under Apache 2.0. Weights are available on HuggingFace in safetensors format. Seven quantized variants are available for running on consumer hardware - at 27B parameters, the model fits on a single GPU with 4-bit quantization (about 16GB VRAM).
No API hosting is available. This is a self-hosted model only.
Strengths
- Chain-of-thought reasoning traces from Claude Opus 4.6
- Runs on consumer hardware with quantization
- Apache 2.0 - fully open for commercial use
- Shows reasoning process via
<think>tags
Weaknesses
- No published benchmarks - quality is unverified
- 8K context window (vs 262K for base model)
- Text-only (base model supports image and video)
- Only ~3,280 training samples - shallow distillation (vs 800K for DeepSeek-R1)
- Preview status - potential bugs and compatibility issues
- Legal gray area - Anthropic's TOS prohibits using outputs to train AI models without permission
Related Coverage
- Claude Opus Reasoning Distilled Into Open 27B Model - Our news coverage
- Qwen3.5-27B - Base model profile
- Distilled vs Base: What Reasoning Distillation Adds - Head-to-head comparison
- Anthropic Distillation Attacks - Background on distillation controversy
Sources:
