Search for a command to run...

Timestamps are as accurate as they can be but may be slightly off. We encourage you to listen to the full context.
In this conversation from a16z's Runtime conference, Gavin Baker, Managing Partner and CIO of Atreides Management, joins David George, General Partner at a16z, to unpack the macro view of AI. The discussion centers on whether we're experiencing an AI bubble, comparing today's infrastructure buildout to the 2000 telecom bubble. (04:00) Baker argues we're not in a bubble, citing that unlike the 2000 era where 97% of fiber was "dark" (unused), today there are "no dark GPUs" - all computing infrastructure is actively utilized. The conversation covers the trillion-dollar data center buildout, competitive dynamics between NVIDIA and Google's TPU, and how AI is reshaping business models across industries.
Managing Partner and Chief Investment Officer of Atreides Management, Baker is a veteran tech investor who lived through the 2000 telecom bubble as an active investor. He's known for his thoughtful analysis on social media, particularly around AI developments, and provides macro-level insights on technology investment cycles.
General Partner at Andreessen Horowitz (a16z), George focuses on technology investments and plays a key role in the firm's AI and infrastructure investing activities. He co-hosts discussions on major technology trends and market dynamics at a16z conferences and events.
Unlike the 2000 telecom bubble where 97% of laid fiber was "dark" (unused), today's AI infrastructure shows full utilization. (04:35) Baker emphasizes that "there are no dark GPUs" - all computing infrastructure is actively being used, with technical papers frequently citing GPU overheating as a major challenge. This utilization pattern, combined with companies achieving 10x increases in return on invested capital since ramping AI spending, suggests genuine economic value creation rather than speculative overinvestment.
Application SaaS companies are making a critical mistake by trying to preserve their traditional 90% gross margins while competing in AI. (14:08) Baker draws parallels to retailers who refused to compete with Amazon due to margin concerns, arguing that "it is definitionally impossible to succeed in AI without gross margin pressure." Companies should view declining margins as "a badge of success rather than something to be feared," using their profitable existing businesses to fund AI initiatives at breakeven while building competitive positions.
Pre-reasoning AI models made frontier labs "the fastest depreciating assets in history" without unique data and distribution. However, reasoning capabilities have fundamentally changed the economics by enabling the classic consumer internet flywheel. (20:48) As Baker explains, "having a big user base now kind of unlocks that flywheel" where good products attract users, users improve algorithms through reinforcement learning, better algorithms improve products, and the cycle accelerates - giving established players with large user bases significant advantages.
AI's measurability enables a fundamental shift from time-based to outcome-based pricing across industries. (26:39) In customer service, this means pricing based on "first call resolution" or "happy customer" metrics rather than agent hours. For consumer applications, AI assistants will likely operate on affiliate fee models, getting paid for successful transactions they facilitate. This represents a move away from the systematic overpaying that occurs in traditional advertising models toward more efficient outcome-driven economics.
The semiconductor competition has consolidated into a battle between NVIDIA's integrated systems approach and Google's TPU ecosystem enabled by Broadcom partnerships. (23:24) Baker predicts many high-profile ASIC programs will be canceled in the next three years, especially if Google begins selling TPUs externally. AMD serves as a necessary "second source" option, while Amazon's Annapurna team shows promise but needs time to mature, following Google's pattern of requiring "three generations to get the TPU right."