BACK_TO_FEEDAICRIER_2
Qwen 3.5 tool calling needs client-side safety nets
OPEN_SOURCE ↗
REDDIT · REDDIT// 6d agoTUTORIAL

Qwen 3.5 tool calling needs client-side safety nets

Qwen 3.5 is a top-tier model for agentic tasks, but its native XML tool-calling format often breaks standard local server parsers. Community fixes involving regex fallbacks and custom Jinja templates can restore near-perfect reliability for local LLM users.

// ANALYSIS

Qwen 3.5 is brilliant but its "plumbing" is currently broken in almost every major local inference engine.

  • XML tool calls frequently leak as plain text or get buried inside thinking blocks, causing agent loops to hang or crash.
  • Stock Jinja templates fail on basic argument filtering; switching to "barubary-attuned" or Unsloth templates is a non-negotiable requirement for stability.
  • Servers like llama.cpp suffer from "thinking leaks" where internal reasoning tags poison multi-turn context, requiring aggressive client-side stripping.
  • LM Studio v0.4.9 currently leads the pack by natively handling Qwen's specific parsing quirks that vLLM and Ollama still struggle with.
  • The model's reasoning capabilities are elite, but developers must treat it as a "raw" output source rather than relying on server-provided JSON fields.
// TAGS
qwen-3.5llmagentai-codingopen-sourceprompt-engineering

DISCOVERED

6d ago

2026-04-06

PUBLISHED

6d ago

2026-04-05

RELEVANCE

9/ 10

AUTHOR

FigZestyclose7787