BACK_TO_FEEDAICRIER_2
Qwen3.6 think tags stick in prompts
OPEN_SOURCE ↗
REDDIT · REDDIT// 2d agoMODEL RELEASE

Qwen3.6 think tags stick in prompts

A LocalLLaMA report says Qwen3.6-35B-A3B still injects `<think>` tags into the prompt even after several no-thinking flags are tried. The output can be stripped cleanly, but the prompt-side behavior looks like a template contract issue rather than a weights problem.

// ANALYSIS

This looks less like a broken model and more like a runtime mismatch with Qwen3.6’s newer thinking template. The official docs show Qwen3.6 thinks by default and expose separate `enable_thinking` and `preserve_thinking` controls, so output parsing alone will not rewrite how history is fed back in.

  • Qwen’s docs say Qwen3.6 operates in thinking mode by default and that `preserve_thinking` can feed prior reasoning back into the next turn.
  • `--reasoning-format deepseek` can strip what downstream consumers see, but it does not necessarily change the tokenizer/chat-template state that gets reinserted upstream.
  • If Unsloth behaves and APEX does not, the difference is probably in template implementation, not model quality.
  • For agents and tools like Frigate, the real fix is likely a nonthinking or custom chat template, or explicit prompt cleanup after template application.
  • This is the kind of change that breaks integrations silently: the assistant output looks fine, but the next-turn context still carries hidden reasoning scaffolding.
// TAGS
qwen3.6-35b-a3bllmreasoningopen-weightsinferenceagent

DISCOVERED

2d ago

2026-04-18

PUBLISHED

2d ago

2026-04-18

RELEVANCE

9/ 10

AUTHOR

Bulky-Priority6824