BACK_TO_FEEDAICRIER_2
Qwen3.6-35B-A3B targets Mac coding workflows
OPEN_SOURCE ↗
REDDIT · REDDIT// 6h agoMODEL RELEASE

Qwen3.6-35B-A3B targets Mac coding workflows

A 64GB Mac user is weighing Qwen3.6-35B-A3B for local coding and app development, while keeping Claude for planning and final review. The model’s appeal is its sparse MoE design, agentic-coding focus, and official Apple Silicon support through local serving stacks.

// ANALYSIS

For this setup, this is probably the right class of model to try first, because it balances local practicality with enough coding quality to be useful.

  • Qwen3.6-35B-A3B is open-weight and explicitly aimed at agentic coding, which fits web/mobile app execution work better than a generic chat model.
  • Apple Silicon support is a real advantage here: official docs call out MLX and llama.cpp, so a Mac-local workflow is not an edge case.
  • On 64GB unified memory, a quantized deployment is realistic, which makes this far more workable than larger dense models.
  • Keeping Claude for architecture and review is a sensible split: use the local model for generation and iteration, and Claude for audit and final judgment.
  • If latency becomes the bottleneck, a smaller Qwen coding model is the natural fallback; if quality matters more, stay in this 35B-class sparse MoE range.
// TAGS
qwenlocal-llmapple-siliconcodingagentic-developmentopen-weight

DISCOVERED

6h ago

2026-04-20

PUBLISHED

6h ago

2026-04-20

RELEVANCE

8/ 10

AUTHOR

skyyyy007