OPEN_SOURCE ↗
REDDIT · REDDIT// 22d agoBENCHMARK RESULT
Qwen 3.5 uncensored tops 89% MMLU
This JANG-based, Apple Silicon-focused uncensored variant of Qwen3.5-VL-397B-A17B claims 88.9% MMLU on 112 GB Macs. It runs through MLX Studio and is aimed at squeezing a 397B MoE model onto high-memory Mac hardware.
// ANALYSIS
This looks less like a clean intelligence jump and more like a benchmark boost from removing refusals, but it is still a serious engineering feat to make a model this large usable on local Mac hardware.
- –The key technical story is JANG mixed-precision quantization, which keeps a 397B model runnable in 112 GB while still hitting useful token speeds.
- –The 1.9-point MMLU lift over the base JANG_1L build is real on paper, but uncensoring can inflate multiple-choice benchmarks without improving downstream coding or tool use as much.
- –MLX Studio is doing a lot of the heavy lifting here; this is a very Apple-Silicon-specific win, not a general-purpose deployment recipe.
- –For local model enthusiasts, the value is obvious: more capability per watt and per GB, with multimodal support still intact.
- –For everyone else, the safer read is "better benchmark score on a derestricted variant," not "new frontier model breakthrough."
// TAGS
qwen3-5-vl-397b-a17b-jang-1l-crackllmbenchmarkopen-weightsinferenceself-hosted
DISCOVERED
22d ago
2026-03-21
PUBLISHED
22d ago
2026-03-21
RELEVANCE
9/ 10
AUTHOR
HealthyCommunicat