OPEN_SOURCE ↗
REDDIT · REDDIT// 21d agoMODEL RELEASE
Qwen 3.5 9B merge distills Claude 4.6 logic
A high-performance community merge combining Qwen 3.5’s 9B architecture with distilled Claude 4.6 Opus reasoning patterns. Optimized for local inference, it delivers an uncensored, frontier-intelligent experience for consumer GPUs.
// ANALYSIS
This 9B merge punches well above its weight, successfully bridging the gap between small local models and frontier-tier reasoning.
- –Float32 precision merging preserves weight integrity, preventing the "lobotomization" common in lower-precision merges.
- –Distillation from Claude 4.6 Opus trajectories drastically reduces repetitive "thinking loops" found in base models.
- –The 262k context window and zero-refusal tuning make it a top contender for complex, uncensored creative work.
- –Achieving 42 tokens per second on an RTX 3060 makes high-tier intelligence accessible on budget hardware.
- –Full support requires the latest LM Studio 0.4.7+ or llama.cpp to handle the hybrid Gated DeltaNet architecture.
// TAGS
qwen3-5-9b-claude-4-6-opus-uncensored-v2llmreasoningopen-weightsself-hostedai-coding
DISCOVERED
21d ago
2026-03-22
PUBLISHED
21d ago
2026-03-22
RELEVANCE
9/ 10
AUTHOR
EvilEnginer