OPEN_SOURCE ↗
REDDIT · REDDIT// 9d agoMODEL RELEASE
Gemma 4 E4B hits MLX for Apple Silicon
Google's Gemma 4 E4B-it model, featuring 4.5B effective parameters and multimodal reasoning, is now optimized for Apple Silicon via the MLX framework. This community-driven conversion brings advanced "thinking" capabilities and native audio-visual understanding to local Mac hardware with a 128K context window.
// ANALYSIS
Gemma 4's "Effective" parameter architecture represents a significant shift toward intelligence-per-parameter, punching well above its weight class in reasoning tasks.
- –Per-Layer Embeddings (PLE) allow the 4.5B model to rival the performance of much larger predecessors like Gemma 3 27B, specifically in complex logic and planning.
- –The inclusion of native audio input and a "Thinking Mode" step-by-step reasoning process makes this a tier-one candidate for building autonomous on-device agents.
- –MLX optimization leverages unified memory to manage the 128K context window efficiently, enabling advanced RAG and long-document analysis on consumer-grade Macs.
- –Community tooling from repositories like `localllm-gemma4-mlx` provides immediate OpenAI-compatible server support, accelerating local developer adoption.
// TAGS
gemma-4llmmlxapple-siliconopen-weightsedge-aimultimodal
DISCOVERED
9d ago
2026-04-02
PUBLISHED
9d ago
2026-04-02
RELEVANCE
9/ 10
AUTHOR
Pathfinder-electron