> ▌

Theo - t3․gg · 1h ago

WorldofAI · 3h ago
OpenCode v1.14.38 is a maintenance-focused release that changes desktop certificate handling to trust the OS system CA store alongside Node’s defaults, expands the UI CSP to allow external connections, and refreshes Node/Nix packaging plus Zed artifacts. The release reads like infrastructure cleanup with real operational impact rather than a user-facing feature drop.
Anthropic’s Claude Mythos Preview is a limited-release frontier model focused on cybersecurity and agentic coding, and it reportedly can autonomously find and help exploit serious software vulnerabilities. According to Axios, the White House is now considering a new safety-review framework for advanced AI models after Mythos intensified concerns about what highly capable systems could mean for government and critical infrastructure.
Alchemy is a TypeScript IaC library that treats infrastructure as runnable code, with resources modeled as async functions. The post argues that the real pain in IaC is not generation speed but verification, because cloud-backed testing is slow, stateful, and expensive.
Plannotator 0.19.8 is a maintenance release for the AI plan-annotation and code-review workflow tool. It improves detection of code file references inside plans, restores URL display for remote users, fixes overlapping sessions in `/plannotator-last` after clearing or resuming sessions, and repairs a code review link issue. The release keeps pushing on the core promise: making agent plans and diffs easier to inspect, annotate, and steer.
OpenCode v1.14.37 focuses on making the agent feel less brittle in multi-workspace and desktop setups. The release adds session warping between workspaces or back to the local project, improves v2 session rendering with cleaner tool states and better compaction summaries, and makes task cancellation propagate to child subtask sessions. On desktop, it fixes migration after the packaging move and smooths Windows titlebar zoom behavior.
DataBoundary is a prompt-injection benchmark and defense lab that wraps untrusted text in random delimiters and tests whether models keep treating it as data. In its latest run, several weaker models jumped from poor baseline defense to 99-100% once delimiters and a strict boundary prompt were added.
FoodTruck Bench says DeepSeek V4 Pro matches GPT-5.2 on its 30-day agentic food-truck benchmark, with similar median outcomes and better run-to-run consistency. The bigger story is economics: it gets there at a much lower token bill.
This Reddit post says Multi-Token Prediction support is about to land in llama.cpp and lists the model families that appear to support it, including DeepSeekv3 OG, DeepSeekv3.2/4, Qwen3.5, GLM4.5+, MiniMax2.5+, Step3.5Flash, and Mimo v2+. The poster notes that, until native MTP weights are available, users need to pull Hugging Face weights and convert them to GGUF, and they plan to test Qwen3.5-122B or GLM4.5-Air first.
A Reddit post describes running Qwen3.6-27B-FP8 on a single RTX 5000 Pro 48GB with BF16 KV cache, 200k context, and vLLM 0.20.1. The author says the setup looks like a practical local coding stack, but the benchmark is still in progress.

DIY Smart Code · 7h ago

Eric Michaud · 9h ago

Github Awesome · 9h ago

Better Stack · 12h ago

Ben Davis · 12h ago

Rob The AI Guy · 12h ago

Ben Davis · 12h ago

DIY Smart Code · 15h ago

Augment Code · 18h ago

Syntax · 20h ago

Eric Michaud · 20h ago

DesignCourse · 20h ago

Prompt Engineering · 21h ago

Github Awesome · 21h ago

Discover AI · 21h ago

The PrimeTime · 21h ago

The PrimeTime · 21h ago

The PrimeTime · 21h ago