OPEN_SOURCE ↗
REDDIT · REDDIT// 1d agoNEWS
Gemma 4, small models fail basic grammar checks
Users report that recent open-weights models like Gemma 4 E4B and Qwen 3 Next fail at basic grammar checking by over-correcting valid text. The issue highlights a persistent fluency bias and prompt brittleness in smaller parameter architectures.
// ANALYSIS
Despite high reasoning benchmarks, small models remain surprisingly bad at surgical text editing because they are trained to rewrite rather than correct.
- –Gemma 4 E4B "corrects" valid words to themselves, demonstrating an inability to execute a simple no-op when text is already correct
- –Instruct-tuning conditions these models to always be "helpful," leading them to force unnecessary edits and add conversational filler
- –Smaller architectures struggle to maintain strict negative constraints (like "do not change the meaning") across longer sentences
- –The over-correction issue is widespread across new releases, affecting the recently launched GPT-OSS 20B and Qwen 3 Next 80B models
// TAGS
gemma-4llmopen-weightsreasoning
DISCOVERED
1d ago
2026-04-13
PUBLISHED
1d ago
2026-04-13
RELEVANCE
8/ 10
AUTHOR
Character-Blood3482