back

Signal/Noise

Get SIGNAL/NOISE in your inbox daily

Signal/Noise

2025-10-30

While tech giants spent yesterday defending AI investment burn rates that would make VCs weep, the real story is playing out in three converging moves: the infrastructure layer consolidating around Nvidia’s $5T throne, the application layer fragmenting into specialized tools, and a quiet reckoning with AI’s fundamental limitations in complex systems prediction.

The Great AI Infrastructure Consolidation

Nvidia hit $5 trillion this week—the first company ever to reach that milestone—while Meta, Microsoft, and Google collectively announced they’re tripling down on AI infrastructure spending despite investor revolt. Meta’s stock cratered 7% after announcing “notably larger” capex for 2026, with Microsoft facing similar punishment for $35B in quarterly AI infrastructure spend. But here’s what Wall Street is missing: this isn’t irrational exuberance, it’s oligopoly formation.

The infrastructure wars are essentially over. Nvidia doesn’t just make the chips—it’s becoming the central nervous system of AI development. When Trump casually mentioned potentially selling China a “scaled-down” version of Nvidia’s Blackwell chip, experts warned it would “dramatically shrink the U.S.’s main advantage” in AI. That’s because Nvidia isn’t just a supplier anymore; it’s the chokepoint.

Meanwhile, the hyperscalers are doubling down not because AI is profitable today, but because they’re buying insurance against being locked out of tomorrow’s digital economy. Google’s first-ever $100B quarter and Microsoft’s record cloud revenues aren’t justifying current AI spend—they’re funding the moats. The companies that control the infrastructure layer will extract rents from every AI application built on top of it. This isn’t a bubble; it’s the world’s most expensive game of musical chairs, and the music is about to stop.

The Application Layer’s Identity Crisis

While infrastructure consolidates, the application layer is having an existential crisis. Canva launched its “Creative Operating System” promising to be your “one-stop-shop for AI design.” PayPal unveiled “agent ready” payments for the “AI shopping future.” Google added AI action chips to Chrome’s homepage. Everyone wants to be the interface between humans and AI—but nobody knows what that actually means yet.

The scramble is revealing AI’s dirty secret: most applications are still solutions in search of problems. Walmart’s new AI tools for suppliers to “better understand customer data” sounds impressive until you realize it’s essentially Excel with a chatbot interface. Character.AI’s decision to ban users under 18 after suicide lawsuits isn’t just a safety move—it’s an admission that their core use case (AI companionship) was fundamentally broken.

The winners emerging aren’t the ones building the flashiest AI features—they’re the ones solving actual workflow problems. Nvidia’s reported $1B investment in coding startup Poolside signals where the real value lies: not in consumer AI companions, but in enterprise tools that can demonstrably replace human labor. The application layer will fragment along industry lines, with specialized AI tools winning specific verticals rather than one platform ruling them all.

The Complexity Wall

Harvard’s Cass Sunstein just published research showing AI faces the same fundamental limitations that doomed central planning: the impossibility of calculating complex systems with interdependent variables. His framework, tested against historical tech bubbles, suggests AI ranks 8/8 on bubble indicators—worse than radio or aviation before the 1929 crash. But Sunstein’s deeper point cuts to AI’s core limitation: it can’t predict outcomes in truly complex systems.

This explains why 95% of companies adopting generative AI saw no profit improvement, according to recent MIT research. It’s not that AI doesn’t work—it’s that most business problems exist in complex systems where small changes cascade unpredictably. Samsung’s “next-gen AI” chips and Alphabet’s record spending are betting that more computational power solves this problem. It doesn’t.

The companies acknowledging this complexity wall are building different strategies. Amazon’s 14,000 layoffs while investing in AI aren’t contradictory—they’re recognition that AI can handle routine tasks but requires human judgment for complex decisions. The real AI winners won’t be the ones with the biggest models, but the ones that best understand which problems AI can actually solve versus which require human insight in complex adaptive systems.

Questions

  • If Nvidia controls AI infrastructure and China gets scaled-down Blackwell chips, who really wins the AI race?
  • Why are companies still burning billions on AI applications when 95% see no profit improvement?
  • When the AI bubble bursts, will it look more like 2000’s dotcom crash or 1929’s everything crash?

Past Briefings

Mar 26, 2026

AI’s Blind Geniuses

Everyone's measuring AI adoption. Nobody's measuring AI results. If Jensen Huang and Alfred Lin can't agree on a scorecard, that tells you more about the state of AI than any benchmark can. THE NUMBER: 0.37% or 100% — the gap between the best score any AI achieved on ARC-AGI-3 (Gemini 3.1 Pro's 0.37%) and Jensen Huang's claim that we've already reached AGI. Even among the most credible voices in AI, nobody can agree on whether we're at the starting line or the finish line. That uncertainty isn't a bug. It's the operating environment. And it's exactly why the question of...

Mar 25, 2026

OpenAI Killed Sora 30 Minutes After a Disney Meeting. The Kill List Is the Strategy Now.

$15M/day to run, $2.1M lifetime revenue. The pivot to Codex puts them behind Claude Code — in a market China is about to commoditize from below. THE NUMBER: $15 million / $2.1 million — the daily operating cost of Sora vs. its lifetime revenue. When a product costs 2,600x more to run per day than it has ever earned, killing it isn't a choice. It's arithmetic. The question is what that arithmetic tells you about everything else OpenAI is doing. OpenAI killed Sora this week. Not quietly — 30 minutes after a working session with Disney, whose $1 billion investment...

Mar 24, 2026

I’m a Mac. I’m a PC. And Only One of Us Is Getting Enterprise Contracts

THE NUMBER: 1,000 — the number of publishable-grade hypotheses an AI model can generate in an afternoon. Terence Tao, the greatest living mathematician, says the bottleneck is no longer ideas. It's knowing which ones are true. Two engineers hacked an inflight entertainment system this week to launch a video game at 35,000 feet. The airline gave them free flights for life. The hacker community on X thought it was the coolest thing they'd seen all month. Every CISO reading this just felt their blood pressure spike. That's the divide. Not between capabilities. Between cultures. Remember those "I'm a Mac, I'm...