BACK_TO_FEEDAICRIER_2
Anthropic Harness Post Spurs Setup Questions
OPEN_SOURCE ↗
REDDIT · REDDIT// 4d agoTUTORIAL

Anthropic Harness Post Spurs Setup Questions

Anthropic’s latest engineering post describes a three-agent harness for long-running coding: planner, generator, evaluator. The Reddit thread asks how to recreate that setup locally, with tools like llama.cpp or oMLX for serving models and agent shells like OpenCode.

// ANALYSIS

The interesting part is not the model stack, it’s the control loop: you’re turning vague intent into a spec, code, and measurable critique. That’s the piece most DIY agent setups miss.

  • The planner is doing scope control, which matters more than raw model quality when a task runs for hours and drifts without structure.
  • The evaluator is the load-bearing role: it turns subjective output into graded criteria plus live testing, which is what makes iteration productive instead of random.
  • Local backends like llama.cpp or oMLX can work fine as infrastructure, but the real bottleneck is usually prompt design, rubric quality, and state handoff between agents.
  • OpenCode is closer to a practical orchestration layer than a model server; the harness still needs explicit contracts, file-based handoffs, and a retry/iteration policy.
  • The article reinforces a broader point for agent builders: reproducibility lives in the methods, not just in the model choice.
// TAGS
anthropicagentai-codingllmself-hostedautomationmcp

DISCOVERED

4d ago

2026-04-07

PUBLISHED

4d ago

2026-04-07

RELEVANCE

7/ 10

AUTHOR

LuJieFei