OPEN_SOURCE ↗
REDDIT · REDDIT// 12d agoRESEARCH PAPER
Isotropic Tradeoff exposes rotation's sparsity tax
The Isotropic Tradeoff is a draft paper and evaluation suite that argues rotation-based LLM quantization improves outlier reconstruction while creating hundreds of thousands of ghost activations. Its Qwen2.5-1.5B 3-bit runs show better cosine similarity and lower outlier MSE, but a worse low-activation noise floor.
// ANALYSIS
Provocative framing aside, the useful takeaway is that rotation-based quantization may optimize the wrong proxy if your real concern is preserving sparse internal structure.
- –TurboQuant, QuaRot, and QuIP all try to smooth outlier channels so low-bit scalar quantization behaves better; this project argues the hidden cost is semantic sparsity loss, not just numerical distortion.
- –The metrics move in opposite directions: outlier reconstruction and cosine similarity improve, while the quiet end of the activation distribution gets noisier.
- –The big unanswered question is downstream effect. If ghost activations hurt reasoning or calibration, this is a real eval gap; if not, it is mostly a representational artifact.
- –The notebook, fixed seeds, and published draft make it easy to rerun, but the claim still needs broader cross-model replication before it becomes a general indictment.
// TAGS
llmresearchbenchmarkopen-sourceisotropic-tradeoffturboquant
DISCOVERED
12d ago
2026-03-30
PUBLISHED
12d ago
2026-03-30
RELEVANCE
8/ 10
AUTHOR
D_E_V_25