The day's announcements reveal a market bifurcating along infrastructure and application lines, with the real competitive energy flowing toward on-device execution and tooling that lets teams move faster. Google DeepMind's Gemma 4 positioning as "byte for byte" the most capable open models, paired with NVIDIA's emphasis on local agentic AI and AMD's push for ready-to-deploy solution blueprints, signals that the infrastructure layer is moving decisively away from cloud-only dependency. OpenAI's acquisition of TBPN and flexible pricing for Codex suggests a different bet: that the moat lives in distribution channels and developer lock-in through ease-of-use rather than raw model capability. IBM and NVIDIA's moves into experiential marketing and consumer gaming (Masters watch party, GeForce NOW titles) feel like plays for mindshare in a market where model commoditization is no longer theoretical. Hugging Face's Gemma 4 support and Anthropic's interpretability work on emotion concepts sit elsewhere entirely, one betting on open model adoption as the default, the other on the research narrative that justifies continued funding. The pattern is clear: builders and chip makers are racing to make small models useful at the edge, cloud vendors are fighting to keep developers dependent on their platforms, and everyone is spending energy on distribution because the models themselves are becoming interchangeable.
Sloane Duvall
A curated reference of models from major AI labs, with open/closed weight status, input modalities, and context window size. American labs tend towards closed weights models and Chinese labs tend toward open weights models.
None
None
None
None
None
None
None