OPEN_SOURCE ↗
REDDIT · REDDIT// 34d agoOPENSOURCE RELEASE
Crystalys launches as local Ollama CLI
Crystalys is a new open-source AI CLI that runs against local or self-hosted Ollama models and exposes built-in tool calling for filesystem access, shell commands, and code search. It looks like an early solo-dev release, but it squarely targets developers who want an offline-friendly assistant without API rate limits.
// ANALYSIS
This is the scrappy end of the local-AI tooling wave: less polished than big-name coding agents, but pointed directly at a real developer pain point.
- –The project is built as an npm CLI with interactive chat, model switching, prompt loading, and config management
- –Its built-in filesystem, shell, and search tools push it beyond simple chat into lightweight agent territory
- –Local Ollama support makes it appealing for developers who care about privacy, cost control, or avoiding hosted-model limits
- –The repo is extremely new and still rough around the edges, so the story here is promising open-source momentum more than production readiness
// TAGS
crystalysclillmopen-sourceautomation
DISCOVERED
34d ago
2026-03-08
PUBLISHED
35d ago
2026-03-08
RELEVANCE
7/ 10
AUTHOR
Inevitable_Back3319