The Inference Report

April 11, 2026
From the Wire

The week's AI stories cluster around a single fracture: the gap between what these systems can do in controlled settings and what happens when they actually operate in the world, where coordination fails, incentives misalign, and the stakes turn violent or criminal.

Start with the operational layer. AWS is launching Agent Registry to manage sprawling fleets of AI agents because enterprises have discovered that multiple agents in production sabotage each other, scheduling conflicts with inquiry handling, stale context corrupting decisions, latency climbing from 200 milliseconds to unacceptable levels. DARPA's MATHBAC project exists for the same reason: agents need a communication protocol to function together. Meta is reportedly pulling engineers into an AI unit to have autonomous agents build software, with humans monitoring, but the industry is still figuring out how to make agents not destroy their own output. This is not a theoretical problem anymore. It is a production problem that money is now being spent to solve.

Parallel to this sits the liability and security layer. A stalking victim is suing OpenAI because ChatGPT ignored three warnings, including its own mass-casualty flag, that a user was dangerous. Anthropic's Mythos model can detect critical software vulnerabilities that legacy systems miss, and cybersecurity stocks fell on the news because the model's capability inverts the traditional moat: the defender's tool is now the attacker's tool. A molotov cocktail was thrown at Sam Altman's home. An npm package was compromised by a nation-state. Hungarian government email passwords are circulating online ahead of elections. These are not separate incidents. They are data points in a market where AI capability has become simultaneously more valuable and more dangerous, and the companies building it have not solved the problem of controlling who uses it or what they do with it.

What ties these together is that regulation and internal controls are not keeping pace with capability deployment. Google is adding end-to-end encryption to Gmail on mobile because WhatsApp's encryption methods drew criticism. Meta's Muse Spark model is offering to analyze users' health data and giving bad medical advice. Onix is launching AI versions of health influencers to sell products 24/7. The legal and safety infrastructure that would normally gate these services either does not exist or is being bypassed because the incentive to ship is higher than the cost of the lawsuit that follows. The question underneath all of this is not whether AI is dangerous. It is whether the companies deploying it at scale can actually control what their own systems do once they leave the lab.

Sloane Duvall