BACK_TO_FEEDAICRIER_2
Qwen-Claude 9B uncensored distill GGUF drops
OPEN_SOURCE ↗
REDDIT · REDDIT// 27d agoOPENSOURCE RELEASE

Qwen-Claude 9B uncensored distill GGUF drops

A community member on r/LocalLLaMA released a GGUF model that merges an uncensored Qwen 3.5 9B checkpoint with a Claude Opus 4.6 reasoning distillation, targeting local inference users who want creative, refusal-free output. The merge extracts modified tensors from an aggressive uncensoring fork and combines them with a Claude-distilled base, with thinking disabled by default via a patched chat template.

// ANALYSIS

The LocalLLaMA community keeps proving that model merging is a legitimate alternative to full fine-tuning for capability and alignment modifications — and this release is a textbook example of iterative community remixing.

  • Tensor-diff merging (extracting only the modified weights from the uncensored fork) is a cleaner approach than naive SLERP, preserving more of the reasoning distillation from the Claude base
  • Disabling the thinking chain by default via the GGUF chat template is a practical UX decision for roleplay and image-prompt use cases where CoT output is noise
  • The model targets RTX 3060 12GB users — a very common hobbyist GPU — making this a real run-anywhere release, not a demo
  • Score of 120 with 23 comments on LocalLLaMA suggests genuine community interest, not spam
  • No eval benchmarks provided; "zero refusals and creative responses" is anecdotal — YMMV depending on use case
// TAGS
llmopen-weightsopen-sourcefine-tuningself-hosted

DISCOVERED

27d ago

2026-03-15

PUBLISHED

27d ago

2026-03-15

RELEVANCE

5/ 10

AUTHOR

EvilEnginer