The labs are converging on a single market opportunity: embedding AI into enterprise workflows through APIs and managed services rather than selling models as standalone products. OpenAI's move to AWS through Amazon Bedrock, combined with AWS's own agent layer (Quick and the four agentic solutions for supply chain, hiring, customer experience, and healthcare), signals that infrastructure providers now control distribution to the customer. NVIDIA's Nemotron 3 Nano Omni addresses a real constraint in multimodal agent systems, the latency and context loss from chaining separate vision, audio, and language models, but the announcement's significance lies not in the model itself but in the efficiency play: agents that run faster consume less compute, which means lower costs for enterprises and higher margins for whoever controls the infrastructure stack. Anthropic, Mistral, and AI21 Labs are positioning Claude and their own models as task-specific tools for creative work and research agents, but they're doing so in a market where AWS, Microsoft, and NVIDIA now own the plumbing. The real competition isn't about model quality anymore; it's about who gets to sit between the enterprise customer and the model provider. OpenAI secured that position at AWS. Microsoft owns it at its own shops. NVIDIA is building it into the hardware layer. The smaller labs are left selling differentiated capabilities, research agents, creative workflows, to customers who access them through someone else's platform. This is how markets consolidate: not through superior technology, but through control of the last mile to the customer.
Sloane Duvall
A curated reference of models from major AI labs, with open/closed weight status, input modalities, and context window size. American labs tend towards closed weights models and Chinese labs tend toward open weights models.
None
None
None
None
None
None
None