BACK_TO_FEEDAICRIER_2
Qwen3.6 35B Heretic drops with near-zero KLD
OPEN_SOURCE ↗
REDDIT · REDDIT// 4h agoMODEL RELEASE

Qwen3.6 35B Heretic drops with near-zero KLD

Qwen3.6-35B-A3B-Heretic is a decensored version of Qwen's latest 35B Mixture-of-Experts (MoE) model, engineered to maintain its high-end reasoning and coding capabilities without the restrictive safety filters. Developed using the Magnitude-Preserving Orthogonal Ablation (MPOA) method, the model achieves a remarkably low KL Divergence of 0.0015, ensuring it stays true to the original's intelligence while slashing refusal rates by nearly 90%.

// ANALYSIS

Heretic is a masterclass in surgical model abliteration, proving that safety filters can be removed without sacrificing the underlying intelligence of a high-parameter MoE architecture.

  • KL Divergence of 0.0015 is remarkably low, indicating the model's weights remain statistically nearly identical to the original despite the "lobotomy" reversal.
  • The 88% reduction in refusals makes it a viable choice for complex, uncensored coding and multi-turn reasoning tasks where standard Qwen models often "preach."
  • With only 3B activated parameters, it delivers high-end reasoning performance while remaining runnable on consumer hardware like a single 24GB VRAM GPU.
  • Preservation of the "Thinking" context is critical for developers using it as an agentic backend for repo-level refactoring.
// TAGS
qwen3.6-35b-a3b-hereticllmopen-weightsmixture-of-expertsdecensoredai-coding

DISCOVERED

4h ago

2026-04-26

PUBLISHED

4h ago

2026-04-26

RELEVANCE

9/ 10

AUTHOR

My_Unbiased_Opinion