OPEN_SOURCE ↗
REDDIT · REDDIT// 1h agoMODEL RELEASE
Qwen3.6-35B-A3B JANG lands on Apple Silicon
bearzi shipped a full 15-profile JANG quantization sweep for Qwen3.6-35B-A3B, spanning extreme compression to near-lossless quality. The suite is tuned for Apple Silicon and already loads in vmlx, MLX Studio, and oMLX with a patch pending.
// ANALYSIS
This is more than another quant pack: it’s a practical argument for layer-aware compression on MoE models, where preserving attention precision matters a lot more than squeezing every last weight uniformly.
- –The 15 profiles give Mac users a real memory/performance ladder instead of a single compromise build
- –Activation-aware calibration plus MSE-all optimization points to quality-first quantization, not just size-chasing
- –MoE models are especially sensitive to naive quantization, so JANG’s higher-precision attention treatment is the key technical bet
- –Native support in vmlx and MLX Studio lowers the friction for local deployment; oMLX support broadens the ecosystem if the patch lands
- –The release also sets up a clear follow-on: if Qwen3-Coder-Next gets the same treatment, local coding workflows could benefit a lot
// TAGS
llminferenceopen-sourceself-hostedqwen3.6-35b-a3b-jang
DISCOVERED
1h ago
2026-04-17
PUBLISHED
3h ago
2026-04-17
RELEVANCE
8/ 10
AUTHOR
PiccoloAcceptable922