Qwen3.5-27B Claude Opus Reasoning Distilled

Community fine-tune that distills Claude Opus 4.6 reasoning into Qwen3.5-27B via LoRA. 28B parameters, Apache 2.0, no published benchmarks.

Qwen3.5-27B Claude Opus Reasoning Distilled

Overview

Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled is a community fine-tune by HuggingFace user Jackrong that applies Claude Opus 4.6 reasoning patterns to the Qwen3.5-27B base model. The distillation uses supervised fine-tuning with LoRA (rank 64) on roughly 3,950 reasoning trace samples created by Claude Opus 4.6.

TL;DR

  • Community LoRA fine-tune of Qwen3.5-27B using Claude Opus 4.6 reasoning traces
  • Uses <think>...</think> tags for chain-of-thought output - text-only, 8K context
  • No published benchmarks - quality is unverified beyond community anecdotes
  • Apache 2.0 licensed, 7 quantized variants available

The model outputs structured reasoning in <think> blocks before providing answers, mimicking Claude's reasoning style. The collection has racked up an estimated 57,000+ downloads across 13 variants (the GGUF version alone hit 20,500) and 134 likes, but comes with important caveats: no published benchmarks, a dramatic reduction in context window from the base model's 262K to just 8K, and loss of multimodal capabilities.

This is a preview release. The creator acknowledges potential bugs, compatibility inconsistencies, and integration edge cases. It's best understood as a research experiment in reasoning distillation, not a production-ready model.

Key Specifications

SpecificationDetails
ProviderJackrong (Community)
Base ModelQwen3.5-27B
ArchitectureGated DeltaNet + Dense FFN (LoRA fine-tuned)
Total Parameters28B
LoRA Rank64
Context Window8,192 tokens
Input ModalitiesText only
Output Format<think> reasoning + final answer
Training Samples~3,950
Fine-tuning FrameworkUnsloth 2026.3.3
Tensor FormatSafetensors (BF16, F32)
LicenseApache 2.0
Quantized Variants7 available

Benchmark Performance

No benchmarks have been published. This is the most significant data gap. The model card provides qualitative descriptions of capabilities but no scores on standard evaluations.

For reference, here are the base model's known scores versus frontier reasoning models:

BenchmarkQwen3.5-27B (base)Claude Opus 4.6Notes
SWE-bench Verified72.480.8Coding
MMLU-ProNot published85.3Knowledge
GPQA DiamondNot published82.7Science reasoning
IFEval95.090.1Instruction following
LiveCodeBench80.7Not publishedLive coding

Without scores for the distilled model, there's no way to assess whether the LoRA fine-tuning improved reasoning capabilities beyond the base Qwen3.5-27B, degraded other capabilities, or had negligible effect. The training set of ~3,280 samples is small by distillation standards - DeepSeek-R1's distilled Qwen variants used 800,000 high-quality reasoning samples with full fine-tuning, orders of magnitude more data and compute.

Key Capabilities

The model is designed for chain-of-thought reasoning tasks: complex Q&A, logical analysis, step-by-step problem solving, and nuanced conversations. The <think> block output format shows the model's reasoning process before delivering a final answer.

Intended use cases include offline analytical tasks, coding assistance, math, and logic-dependent prompting. The creator specifically notes it's not suited for production deployment and recommends it for research and experimentation.

The 8K context window is a significant practical limitation. The base Qwen3.5-27B supports 262K tokens natively and up to 1M with YaRN extension. The distilled version can't handle the long documents, codebases, or extended agent conversations that the base model supports.

Pricing and Availability

The model is free under Apache 2.0. Weights are available on HuggingFace in safetensors format. Seven quantized variants are available for running on consumer hardware - at 27B parameters, the model fits on a single GPU with 4-bit quantization (about 16GB VRAM).

No API hosting is available. This is a self-hosted model only.

Strengths

  • Chain-of-thought reasoning traces from Claude Opus 4.6
  • Runs on consumer hardware with quantization
  • Apache 2.0 - fully open for commercial use
  • Shows reasoning process via <think> tags

Weaknesses

  • No published benchmarks - quality is unverified
  • 8K context window (vs 262K for base model)
  • Text-only (base model supports image and video)
  • Only ~3,280 training samples - shallow distillation (vs 800K for DeepSeek-R1)
  • Preview status - potential bugs and compatibility issues
  • Legal gray area - Anthropic's TOS prohibits using outputs to train AI models without permission

Sources:

Qwen3.5-27B Claude Opus Reasoning Distilled
About the author AI Benchmarks & Tools Analyst

James is a software engineer turned tech writer who spent six years building backend systems at a fintech startup in Chicago before pivoting to full-time analysis of AI tools and infrastructure.