OPEN_SOURCE ↗
REDDIT · REDDIT// 1h agoMODEL RELEASE
Qwen3.6-Plus repeats Gemma 4 research bugs
A Reddit user says Qwen3.6-Plus feels uncomfortably similar to Gemma 4 in research workflows: it keeps searching without converging, repeats its plan between tool calls, and never seems to use page fetches. The complaint lands hard because Alibaba is pitching the model as an agentic upgrade with stronger coding and multimodal reasoning.
// ANALYSIS
The sharp takeaway is that users do not care how ambitious the launch copy sounds if the agent loop feels sloppy in practice. If this anecdote reflects a broader pattern, Qwen’s latest release may be trading benchmark polish for worse tool discipline.
- –The official Qwen3.6-Plus launch positions it as an agentic coding model with a 1M-token context window and multimodal reasoning, so the expectation bar is high.
- –The post’s core complaint is not accuracy alone, but execution: repeated thinking blocks, search-only behavior, and failure to stop when enough evidence is already gathered.
- –That makes the comparison to Gemma 4 especially pointed, because both models are being judged on workflow quality rather than raw chat competence.
- –The user’s Qwen3.5 comparison suggests the regression is visible in real research tasks, where tool selection and persistence matter more than single-turn response quality.
- –One Reddit thread is not a benchmark, but it is the kind of anecdotal signal that shapes developer trust fast when a model is meant for agentic use.
// TAGS
qwen3.6-plusllmagentreasoningsearch
DISCOVERED
1h ago
2026-04-17
PUBLISHED
5h ago
2026-04-17
RELEVANCE
9/ 10
AUTHOR
Daniel_H212