BACK_TO_FEEDAICRIER_2
Qwen3.6-35B-A3B JANG lands on Apple Silicon
OPEN_SOURCE ↗
REDDIT · REDDIT// 1h agoMODEL RELEASE

Qwen3.6-35B-A3B JANG lands on Apple Silicon

bearzi shipped a full 15-profile JANG quantization sweep for Qwen3.6-35B-A3B, spanning extreme compression to near-lossless quality. The suite is tuned for Apple Silicon and already loads in vmlx, MLX Studio, and oMLX with a patch pending.

// ANALYSIS

This is more than another quant pack: it’s a practical argument for layer-aware compression on MoE models, where preserving attention precision matters a lot more than squeezing every last weight uniformly.

  • The 15 profiles give Mac users a real memory/performance ladder instead of a single compromise build
  • Activation-aware calibration plus MSE-all optimization points to quality-first quantization, not just size-chasing
  • MoE models are especially sensitive to naive quantization, so JANG’s higher-precision attention treatment is the key technical bet
  • Native support in vmlx and MLX Studio lowers the friction for local deployment; oMLX support broadens the ecosystem if the patch lands
  • The release also sets up a clear follow-on: if Qwen3-Coder-Next gets the same treatment, local coding workflows could benefit a lot
// TAGS
llminferenceopen-sourceself-hostedqwen3.6-35b-a3b-jang

DISCOVERED

1h ago

2026-04-17

PUBLISHED

3h ago

2026-04-17

RELEVANCE

8/ 10

AUTHOR

PiccoloAcceptable922