Search for a command to run...

Timestamps are as accurate as they can be but may be slightly off. We encourage you to listen to the full context.
Jonathan Ross, founder and CEO of Groq, returns to 20 VC for a comprehensive discussion about the current state of AI infrastructure and compute. (00:21) Ross, who previously led the TPU team at Google, explains that while many question whether there's an AI bubble, the real indicator is what smart money is doing - major tech companies continue doubling down on AI spending. (05:26) He argues that the demand for compute is insatiable, with companies like OpenAI and Anthropic able to nearly double their revenue if given twice their current inference compute capacity. (12:26) The conversation covers everything from NVIDIA's market position and energy requirements to the geopolitical implications of AI compute control.
Jonathan Ross is the founder and CEO of Groq, an AI chip company focused on inference at scale, which has raised over $3 billion with a recent valuation approaching $7 billion. Before founding Groq, Ross led the team that built Google's TPU (Tensor Processing Unit), making him one of the key architects of modern AI hardware infrastructure.
The most striking insight from Ross is that compute availability, not performance, has become the primary value proposition for AI infrastructure companies. (23:17) Ross describes customers requesting 5x their total capacity, which no one in the industry can fulfill. This scarcity means that if major AI companies like OpenAI or Anthropic doubled their inference compute, their revenue would almost double within a month due to current rate limiting constraints. The practical implication is that businesses should prioritize securing compute capacity over optimizing for marginal performance gains.
Ross draws a compelling parallel between AI responsiveness and consumer products, noting that high-margin consumer goods correlate with speed of action - tobacco acts fastest, followed by soft drinks. (13:25) This principle applies directly to AI applications, where every 100 milliseconds of speed improvement results in approximately 8% conversion rate increases. The lesson for professionals is that investing in faster AI infrastructure isn't just about user experience - it's about creating measurable business value through improved engagement and brand affinity.
Countries that control compute will control AI, and compute requires energy infrastructure. (36:41) Ross argues that Europe could compete effectively if it leveraged resources like Norway's wind capacity, which could theoretically provide as much energy as the entire United States. (34:01) For business leaders, this means considering geographic location as a strategic advantage, particularly for companies requiring significant compute resources. The message is clear: locate operations where energy is abundant and cheap.
Contrary to popular fears about AI-driven unemployment, Ross predicts massive labor shortages due to three factors: deflationary pressure making life cheaper, people opting to work less due to lower costs, and entirely new job categories emerging. (44:02) He draws parallels to how 98% of the workforce moved from agriculture to other sectors over the past century. For professionals, this suggests focusing on skills that complement AI rather than compete with it, and preparing for a world where human creativity and strategic thinking become even more valuable.
When Ross explains why companies like OpenAI will build their own chips, the primary benefit isn't cost savings - it's control over destiny. (17:05) Custom infrastructure prevents suppliers from dictating allocation and ensures capacity when needed. Ross shares how Google once built 10,000 AMD servers just to get better Intel pricing, demonstrating that the real value lies in negotiating power and supply chain control. For businesses, this principle applies beyond chips: owning critical infrastructure components provides strategic flexibility that often outweighs pure cost considerations.