OPEN_SOURCE ↗
REDDIT · REDDIT// 37d agoOPENSOURCE RELEASE
ScrapChat brings self-hosted llama.cpp chat, web tools
ScrapChat is a new open-source Node.js web UI for local llama.cpp servers that adds autonomous web search and page fetching, multimodal image input, live reasoning stream rendering, and slot-level context monitoring. It targets privacy-first local inference setups with no cloud dependency for core chat.
// ANALYSIS
This is a strong “power-user local stack” release that closes practical gaps between raw `llama.cpp` and polished AI chat apps.
- –Autonomous 5-step tool chaining (search -> fetch -> reason) makes local models more useful for real-world research workflows
- –Slot pinning plus live `n_ctx` usage is unusually practical for people running long-context or multi-conversation local sessions
- –Multimodal support and live `<think>` rendering show clear focus on modern open models like Qwen3/3.5, not just basic text chat
- –Zero frontend build complexity (Express + vanilla JS + Tailwind v4) lowers barrier for self-hosters who want to customize the stack
// TAGS
scrapchatopen-sourceself-hosteddevtoolagentmultimodal
DISCOVERED
37d ago
2026-03-05
PUBLISHED
37d ago
2026-03-05
RELEVANCE
8/ 10
AUTHOR
ols255