OPEN_SOURCE ↗
REDDIT · REDDIT// 34d agoOPENSOURCE RELEASE
a-hat-optimizer boosts agent tool calling
Arthur Vigier's new open-source package claims it can extract a linear "agency direction" from LLM hidden states and use it to decide when an agent should call a tool, with reported gains from 26.7% to 85% on Qwen3-1.7B and 52.5% to 76.3% on Qwen3-8B. The project shipped on GitHub and PyPI on March 8 with one-line extraction, hidden-state hooks, and threshold calibration for Hugging Face models.
// ANALYSIS
This is a clever agents hack because it promises much better tool use without finetuning, but the big gains are still self-reported and need outside replication.
- –If the signal generalizes, it suggests models often internally know a tool call is needed before their token output reliably says so.
- –Smaller open-weight models look like the biggest beneficiaries, which could make cheap agent stacks much more usable for search, code, and file workflows.
- –The package is practical rather than purely theoretical, bundling contrastive extraction, calibration strategies, and runtime prediction into a lightweight Python library.
- –The obvious caveat is validation: there is no broad independent benchmark yet, so developers should treat the +26-58% improvement claims as promising, not settled.
// TAGS
a-hat-optimizerllmagentopen-sourceresearch
DISCOVERED
34d ago
2026-03-08
PUBLISHED
35d ago
2026-03-08
RELEVANCE
8/ 10
AUTHOR
Clean-Cardiologist77