BACK_TO_FEEDAICRIER_2
Gemma 4, small models fail basic grammar checks
OPEN_SOURCE ↗
REDDIT · REDDIT// 1d agoNEWS

Gemma 4, small models fail basic grammar checks

Users report that recent open-weights models like Gemma 4 E4B and Qwen 3 Next fail at basic grammar checking by over-correcting valid text. The issue highlights a persistent fluency bias and prompt brittleness in smaller parameter architectures.

// ANALYSIS

Despite high reasoning benchmarks, small models remain surprisingly bad at surgical text editing because they are trained to rewrite rather than correct.

  • Gemma 4 E4B "corrects" valid words to themselves, demonstrating an inability to execute a simple no-op when text is already correct
  • Instruct-tuning conditions these models to always be "helpful," leading them to force unnecessary edits and add conversational filler
  • Smaller architectures struggle to maintain strict negative constraints (like "do not change the meaning") across longer sentences
  • The over-correction issue is widespread across new releases, affecting the recently launched GPT-OSS 20B and Qwen 3 Next 80B models
// TAGS
gemma-4llmopen-weightsreasoning

DISCOVERED

1d ago

2026-04-13

PUBLISHED

1d ago

2026-04-13

RELEVANCE

8/ 10

AUTHOR

Character-Blood3482