OPEN_SOURCE ↗
REDDIT · REDDIT// 34d agoOPENSOURCE RELEASE
CRW drops lean Rust scraper stack
CRW is a new open-source, self-hosted web scraper and crawler in Rust positioned as a lightweight Firecrawl-compatible alternative for AI workflows. The project pairs a single-binary API server with MCP support, LLM-powered structured extraction, and optional JS rendering while claiming roughly 6 MB idle RAM and much faster scrape latency than Firecrawl.
// ANALYSIS
This is the kind of infra project AI builders keep wanting: less platform, more pipe. CRW matters because it targets the unglamorous bottleneck in RAG and agent systems — reliable web ingestion — without dragging in a heavyweight Node/Redis stack.
- –Firecrawl API compatibility makes CRW unusually easy to trial inside existing scraping, RAG, or agent pipelines
- –Built-in MCP support gives Claude Code, Claude Desktop, Cursor, Windsurf, and other clients a direct path to web scraping tools
- –The repo’s strongest pitch is operational simplicity: single binary, Docker-ready, optional JS rendering, and self-hosted economics
- –CSS/XPath targeting, chunking, BM25/cosine filtering, and schema-based extraction show this is aimed at production ingestion workflows, not just hobby scraping
- –It is still extremely early at v0.0.2, so the real question is whether the benchmark claims and compatibility hold up under messy real-world sites
// TAGS
crwopen-sourceragmcpapiself-hostedagent
DISCOVERED
34d ago
2026-03-08
PUBLISHED
35d ago
2026-03-08
RELEVANCE
8/ 10
AUTHOR
psycenos