OpenAI is carving out specialized advantage in high-stakes domains by tightening control over access and capability release. The Trusted Access for Cyber program now includes GPT-5.4-Cyber, a model explicitly designed for vetted defenders, a clear signal that OpenAI sees cybersecurity as both a defensible market and a domain where unrestricted access carries real liability. The move pairs capability advancement with gating, which is the opposite of the company's consumer-facing ChatGPT onboarding materials. Google DeepMind's Gemini Robotics ER 1.6 announcement reflects a similar push into embodied reasoning, a space where real-world task performance can be measured and monetized. Meanwhile, NVIDIA is positioning quantum AI models as open-source infrastructure, a different play entirely, one that treats the foundation as a platform layer and lets others build the applications. GitHub's security game and AI21's warning about coding agent benchmark inflation both point to the same emerging problem: agent capabilities are being measured against benchmarks that don't always predict real-world reliability, and the market is starting to notice. The tension here is instructive. Specialized, gated access (OpenAI's cyber model) competes against open infrastructure plays (NVIDIA's quantum models) and against the pressure to prove agent systems actually work (GitHub and AI21's focus on real vulnerabilities and false positives). Security and robotics are where capability claims get tested against reality fastest. That's where the money follows.
Sloane Duvall
A curated reference of models from major AI labs, with open/closed weight status, input modalities, and context window size. American labs tend towards closed weights models and Chinese labs tend toward open weights models.
None
None
None
None
None
None
None