OPEN_SOURCE ↗
REDDIT · REDDIT// 8d agoBENCHMARK RESULT
Qwen 3.5 122B beats Gemma 4 in meeting summaries
A real-world evaluation comparing Google’s new Gemma 4 (31B Dense) and Alibaba’s Qwen 3.5 (122B MoE) for meeting summarization reveals that Qwen’s larger parameter pool captures significantly more detail. While Gemma 4 offers high-precision reasoning in a smaller footprint, Qwen 3.5 proves superior for exhaustive information extraction from long-form audio transcriptions.
// ANALYSIS
Qwen 3.5 122B (10B active) demonstrates that total parameter count remains a critical factor for "world knowledge" and detail retention in complex tasks.
- –Qwen 3.5 122B Q4 quantization outperformed Gemma 4 Q8 in capturing meeting nuances, despite Gemma's higher precision.
- –Gemma 4 31B remains the "sweet spot" for 48GB VRAM setups, providing frontier-level performance for users with limited local hardware.
- –The MoE architecture in Qwen 3.5 provides a substantial advantage in summarizing dense context without the VRAM penalty of a 100B+ dense model.
- –This comparison highlights a shift where MoE models are increasingly viable for local "prosumer" workflows requiring maximum detail.
// TAGS
llmbenchmarkopen-weightsgemma-4qwen-3-5
DISCOVERED
8d ago
2026-04-03
PUBLISHED
8d ago
2026-04-03
RELEVANCE
8/ 10
AUTHOR
Terminator857