OPEN_SOURCE ↗
REDDIT · REDDIT// 8d agoOPENSOURCE RELEASE
Gemma 4 Local Chat lands on Mac
This open-source Mac app wraps Google’s Gemma 4 in a minimal Flask plus vanilla HTML/CSS/JS chat UI powered by MLX, so Apple Silicon users can run it locally after a one-time model download. It adds a persistent system prompt, full session memory, a token counter, and a desktop launcher for a one-double-click workflow.
// ANALYSIS
This is less a flashy demo than a credible local-LLM UX package: the model matters, but the real product is the frictionless offline workflow.
- –MLX on Apple Silicon plus an M4 16GB benchmark of roughly 19 tok/s makes the experience sound practical, not just theoretical
- –Persistent system prompts and full conversation memory push it beyond a basic wrapper and toward a useful writing or roleplay assistant
- –The desktop shortcut is the right kind of polish for local AI, because setup friction is usually the thing that kills adoption
- –Full-history memory will hit context limits fast, so the token counter is doing real work here, not just decoration
- –MIT licensing and offline operation after setup make it an attractive privacy-first reference implementation for local AI apps
// TAGS
gemma-4-local-chatllmchatbotopen-sourceself-hostededge-ai
DISCOVERED
8d ago
2026-04-03
PUBLISHED
9d ago
2026-04-03
RELEVANCE
8/ 10
AUTHOR
Polstick1971