Command Palette

Search for a command to run...

PodMine
a16z Podcast
a16z Podcast•October 14, 2025

Is AI Slowing Down? Nathan Labenz Says We're Asking the Wrong Question

Nathan Labenz and Eric discuss the current state of AI, arguing that contrary to claims of slowing progress, AI is continuing to advance rapidly across various domains, including reasoning, scientific discovery, and multimodal capabilities.
AI & Machine Learning
Tech Policy & Ethics
Developer Culture
Sam Altman
Cal Newport
Nathan Labenz
Sergey Brin
Demis Hassabis

Summary Sections

  • Podcast Summary
  • Speakers
  • Key Takeaways
  • Statistics & Facts
  • Compelling StoriesPremium
  • Thought-Provoking QuotesPremium
  • Strategies & FrameworksPremium
  • Similar StrategiesPlus
  • Additional ContextPremium
  • Key Takeaways TablePlus
  • Critical AnalysisPlus
  • Books & Articles MentionedPlus
  • Products, Tools & Software MentionedPlus
0:00/0:00

Timestamps are as accurate as they can be but may be slightly off. We encourage you to listen to the full context.

0:00/0:00

Podcast Summary

Nathan Labenz, host of the Cognitive Revolution podcast, joins to debunk the claim that AI progress has plateaued, making a compelling case that we're witnessing unprecedented advancement across multiple fronts. (00:25) The discussion addresses Cal Newport's recent arguments about AI stagnation while examining the real capabilities emerging in GPT-5, reasoning models, and multimodal AI systems. Key themes include the debate over whether AI innovation is slowing versus our expectations catching up to reality, the transformative potential of agents and automation, and how we should shape AI development toward a future we actually want.

• Main Theme: AI capabilities are advancing rapidly across dimensions beyond chatbots, from scientific discovery to coding assistance, despite perceptions of slowdown

Speakers

Nathan Labenz

Nathan Labenz is the host of the Cognitive Revolution podcast and a leading voice in AI analysis and commentary. He has extensive experience tracking AI developments and capabilities, with deep expertise in evaluating model performance across different benchmarks and use cases. Labenz is known for his nuanced takes on AI progress and his ability to translate technical developments into broader implications for society and the future of work.

Eric Newcomer

Eric Newcomer is a technology journalist and podcast host focusing on AI developments and their business implications. He has extensive experience covering the AI industry and frequently interviews key figures in the space about emerging trends and capabilities.

Key Takeaways

Distinguish Between AI Impact and Capability Progress

Labenz emphasizes the critical importance of separating two distinct questions about AI: whether it's beneficial for us now and in the future, versus whether capabilities continue advancing rapidly. (01:38) While valid concerns exist about AI making students lazier or reducing cognitive strain, this doesn't negate the reality that AI systems are becoming dramatically more capable. Many critics conflate current usage problems with fundamental capability limitations, leading to flawed conclusions about AI's trajectory. This distinction matters because policy and preparation decisions should be based on actual capability trends, not just current adoption challenges.

Context Windows Enable Massive Productivity Gains

The expansion from GPT-4's initial 8,000 token context window to current models handling dozens of papers represents a game-changing improvement. (10:52) This advancement allows AI to work with vast amounts of information while maintaining high fidelity reasoning across long contexts. Labenz notes that this capability shift means smaller, more efficient models can achieve results previously requiring massive parameter counts by leveraging extensive context rather than memorizing facts. For professionals, this translates to AI systems that can analyze entire codebases, multiple research papers, or comprehensive project documentation in a single session.

Mathematical Reasoning Represents a Capability Leap

The progression from GPT-4 struggling with high school math to current models achieving IMO gold medals demonstrates qualitative advancement, not just incremental improvement. (13:37) Recent models have jumped from 2% to 25% on the frontier math benchmark within a year, and some have solved problems that took professional mathematicians 18 months to crack. This mathematical reasoning capability serves as a crucial foundation for scientific discovery and engineering applications, suggesting we're entering a phase where AI can contribute genuinely novel solutions to unsolved problems rather than just processing existing knowledge.

AI Agents Show Exponential Task Duration Growth

The task length capabilities of AI agents are doubling approximately every four months, with current systems handling roughly two hours of continuous work. (59:28) Extrapolating this trend suggests AI could manage two-day tasks within a year and two-week projects within two years. Even with 50% success rates, this would represent transformative automation potential across knowledge work. The key insight is that this growth pattern, if sustained, rapidly approaches the scale of work that defines many professional roles, making AI agents increasingly viable alternatives to human workers for complex, extended projects.

Multimodal Integration Creates New Intelligence Paradigms

AI progress extends far beyond language models into biology, materials science, and other specialized domains that will eventually integrate into unified systems. (53:43) Recent breakthroughs include new antibiotics discovered by AI systems and scientific hypotheses generated that match unpublished human research. The pattern mirrors the evolution from text-only to integrated text-image models, suggesting future AI will possess "sixth senses" for molecular spaces, protein structures, and material properties. This multimodal intelligence represents a path toward capabilities that look genuinely superintelligent in specific domains, even without surpassing humans in all areas.

Statistics & Facts

  1. GPT 4.5 scored 65% on the Simple QA benchmark compared to O3's 50%, representing a 30% improvement in absorbing esoteric factual knowledge. (08:52) This demonstrates continued scaling benefits for larger models in knowledge retention.
  2. The Intercom Fin agent now resolves 65% of customer service tickets automatically, up from 55% just three to four months earlier. (32:38) This rapid improvement trajectory suggests human customer service roles face significant displacement pressure.
  3. OpenAI's O3 model can successfully complete 40% of pull requests by research engineers at OpenAI, compared to low-to-mid single digits for previous models. (38:35) This represents AI entering the steep part of the capability curve for high-level technical work.

Compelling Stories

Available with a Premium subscription

Thought-Provoking Quotes

Available with a Premium subscription

Strategies & Frameworks

Available with a Premium subscription

Similar Strategies

Available with a Plus subscription

Additional Context

Available with a Premium subscription

Key Takeaways Table

Available with a Plus subscription

Critical Analysis

Available with a Plus subscription

Books & Articles Mentioned

Available with a Plus subscription

Products, Tools & Software Mentioned

Available with a Plus subscription

More episodes like this

In Good Company with Nicolai Tangen
January 14, 2026

Figma CEO: From Idea to IPO, Design at Scale and AI’s Impact on Creativity

In Good Company with Nicolai Tangen
We Study Billionaires - The Investor’s Podcast Network
January 14, 2026

BTC257: Bitcoin Mastermind Q1 2026 w/ Jeff Ross, Joe Carlasare, and American HODL (Bitcoin Podcast)

We Study Billionaires - The Investor’s Podcast Network
Uncensored CMO
January 14, 2026

Rory Sutherland on why luck beats logic in marketing

Uncensored CMO
This Week in Startups
January 13, 2026

How to Make Billions from Exposing Fraud | E2234

This Week in Startups
Swipe to navigate