OpenAI is monetizing the consumer moat while arming enterprise with specialized models. The company is testing ads in ChatGPT to offset free-tier costs, launching voice reasoning capabilities into the API for real-time applications, and rolling out Trusted Access for Cyber with dedicated GPT-5.5 variants aimed at security teams and infrastructure defenders. Meanwhile, enterprise is moving fast: Parloa and Simplex are shipping production systems on OpenAI models, reducing deployment friction for customer service and software development. Google DeepMind's AlphaEvolve positions Gemini as a coding agent with cross-domain scope, but the announcement is thin on specifics about competitive differentiation. The real signal in this volume is infrastructure: NVIDIA is consolidating the deployment layer through partnerships with IREN and energy officials, while AMD is fighting for parity by optimizing vLLM and enabling 3DGS workloads on Instinct GPUs. GitHub and Hugging Face are solving operational problems that matter at scale, token efficiency in agentic workflows, clinical fine-tuning without CUDA lock-in. Anthropic is publishing mechanistic work on interpretability and open-source alignment tooling, moves that read as credibility plays in a market where the labs are increasingly differentiated by their relationship to enterprise builders and infrastructure vendors rather than model capability alone. The volume suggests a market settling into layers: OpenAI owns consumer and enterprise API; NVIDIA owns the deployment stack; AMD is fighting for cost parity; and the open ecosystem is becoming the proving ground for operational maturity.
Sloane Duvall
A curated reference of models from major AI labs, with open/closed weight status, input modalities, and context window size. American labs tend towards closed weights models and Chinese labs tend toward open weights models.
None
None
None
None
None
None
None