BACK_TO_FEEDAICRIER_2
Qwen3.5 reasoning budget stumps vLLM, SGLang
OPEN_SOURCE ↗
REDDIT · REDDIT// 22d agoTUTORIAL

Qwen3.5 reasoning budget stumps vLLM, SGLang

A Reddit user asks how to cap Qwen3.5’s reasoning budget when serving through vLLM or SGLang, saying the model keeps chewing through roughly 1,500 thinking tokens no matter what they try. The thread points to a familiar pain point: Qwen’s thinking controls exist, but they’re exposed differently across serving stacks.

// ANALYSIS

Hot take: this looks less like a model bug and more like a docs-and-runtime mismatch. Qwen3.5 has explicit thinking controls, but vLLM and SGLang expose them as parser/template flags rather than a portable “reasoning-budget” primitive, so it’s easy to misconfigure.

  • Qwen’s repo says `enable_thinking=False` is the strict switch, and `/think` / `/no_think` can toggle behavior turn by turn.
  • vLLM documents the `qwen3` reasoning parser and server defaults via `--default-chat-template-kwargs '{"enable_thinking": false}'`, with request-level kwargs overriding the default.
  • SGLang’s Qwen 3.5 guide shows `--reasoning-parser qwen3` and `--tool-call-parser qwen3_coder`, but no separate budget knob in the basic launch flow.
  • If the engine still allows free-form thinking, prompt tweaks alone probably won’t stop the model from burning tokens.
  • For production use, this really wants a single compatibility matrix for “thinking on/off,” “budget cap,” and “framework support.”
// TAGS
qwen-3.5vllmsglangreasoningllminference

DISCOVERED

22d ago

2026-03-20

PUBLISHED

22d ago

2026-03-20

RELEVANCE

8/ 10

AUTHOR

DingyAtoll