Command Palette

Search for a command to run...

PodMine
Dwarkesh Podcast
Dwarkesh Podcast•October 17, 2025

Andrej Karpathy — AGI is still a decade away

Andrej Karpathy discusses his views on AI progress, the current limitations of language models, the importance of education, and his new project Eureka aimed at creating a "Starfleet Academy" for technical learning.

Summary Sections

  • Podcast Summary
  • Speakers
  • Key Takeaways
  • Statistics & Facts
  • Compelling StoriesPremium
  • Thought-Provoking QuotesPremium
  • Strategies & FrameworksPremium
  • Similar StrategiesPlus
  • Additional ContextPremium
  • Key Takeaways TablePlus
  • Critical AnalysisPlus
  • Books & Articles MentionedPlus
  • Products, Tools & Software MentionedPlus
0:00/0:00

Timestamps are as accurate as they can be but may be slightly off. We encourage you to listen to the full context.

0:00/0:00

Podcast Summary

Former Tesla AI Director and OpenAI Co-founder Andre Karpathy discusses why this will be the "decade of agents" rather than the "year of agents," sharing his perspective on AI timelines based on fifteen years of experience in the field. (00:27) He traces the evolution of AI through major paradigm shifts - from the deep learning revolution with AlexNet to the reinforcement learning era with Atari games, and finally to the current LLM breakthrough. Throughout the conversation, Karpathy emphasizes his practical engineering mindset, explaining why he sees AI as fundamentally an extension of computing rather than a magical leap toward superintelligence.

  • Main themes include realistic AI timelines, the technical challenges still facing current agents, the nature of intelligence and learning, and the gradual automation of society through what Karpathy calls the "autonomy slider."

Speakers

Andre Karpathy

Andre Karpathy is a renowned AI researcher and educator who co-founded OpenAI and served as the Director of AI at Tesla's Autopilot division from 2017-2022, where he led the development of self-driving car technology. (02:42) He has nearly two decades of experience in AI and deep learning, having worked closely with Geoffrey Hinton at University of Toronto during the early days of the deep learning revolution. Currently, he's building Eureka, an educational platform he describes as "Starfleet Academy" for technical knowledge, and recently released NanoChat - a simplified but complete ChatGPT implementation in 1,000 lines of code.

Key Takeaways

Intelligence Requires More Than Pattern Recognition

Current LLMs lack many cognitive components that would make them truly intelligent agents. (19:40) Karpathy draws analogies between AI models and brain regions, suggesting that transformers might represent "cortical tissue" while reasoning traces could be like the "prefrontal cortex." However, many brain regions remain unexplored in current AI systems, including the equivalent of the hippocampus for memory formation, the amygdala for emotions and instincts, and other ancient nuclei that handle different aspects of cognition. This is why these models still feel cognitively deficient despite impressive capabilities.

Reinforcement Learning Is Fundamentally Flawed

Karpathy argues that current reinforcement learning methods are "terrible" and create noisy learning signals. (40:01) He uses the vivid analogy of "sucking supervision through a straw" - where an agent might do hundreds of rollouts solving a problem, but only gets a single reward signal at the end indicating success or failure. This means every action in a successful trajectory gets upweighted, even wrong turns and mistakes along the way. Humans don't learn this way - they have sophisticated review processes and can assign partial credit to different parts of their problem-solving approach.

Pre-training Creates "Ghost-like" Intelligence

Unlike biological intelligence that evolved through natural selection, LLMs are trained through imitation of human data on the internet, creating what Karpathy calls "ghost" or "spirit" entities. (08:42) This process does two things: it accumulates knowledge from internet documents, but more importantly, it develops intelligence by learning algorithmic patterns and developing capabilities like in-context learning. The knowledge component may actually be holding back neural networks by making them too reliant on memorization rather than reasoning from first principles.

The Demo-to-Product Gap Is Enormous

From his experience leading Tesla's self-driving efforts, Karpathy learned that impressive demos can be misleading about timeline to deployment. (105:15) He describes a "march of nines" where each additional nine of reliability (90% to 99% to 99.9%) requires constant amounts of work. This is especially true in domains with high cost of failure, like self-driving cars or production software systems. Many AI applications today are impressive demos but still face years of engineering work to become reliable products that can handle edge cases and real-world complexity.

Education Will Transform Into "Cognitive Gym"

Post-AGI, Karpathy envisions education becoming like physical fitness - something people pursue for personal fulfillment rather than economic necessity. (128:04) Just as people go to the gym despite not needing physical strength for work, people will pursue learning because it's "fun, healthy, and makes you look hot." His vision for Eureka is to build perfect AI tutors that can understand exactly where a student is, serve appropriately challenging material, and make learning feel effortless. This could unlock human potential in ways similar to how modern fitness culture has made physical capabilities that were once rare become commonplace.

Statistics & Facts

  1. Karpathy has nearly fifteen years of experience in AI research and has witnessed multiple major paradigm shifts in the field. (02:42) This gives him perspective on how breakthrough predictions have typically played out over time.
  2. In his experiment recreating Jan LeCun's 1989 convolutional network, Karpathy found that thirty-three years of algorithmic progress alone could only halve the error rate. (25:16) Significant further improvements required more data, longer training, and better computational optimizations, suggesting that progress requires simultaneous advances across multiple dimensions.
  3. Llama 3 was trained on 15 trillion tokens, and a 70B model stores approximately 0.07 bits per token seen during pre-training. (17:17) In contrast, during in-context learning, the KV cache grows by about 320 kilobytes per token, representing a 35-million-fold difference in information density between training and inference.

Compelling Stories

Available with a Premium subscription

Thought-Provoking Quotes

Available with a Premium subscription

Strategies & Frameworks

Available with a Premium subscription

Similar Strategies

Available with a Plus subscription

Additional Context

Available with a Premium subscription

Key Takeaways Table

Available with a Plus subscription

Critical Analysis

Available with a Plus subscription

Books & Articles Mentioned

Available with a Plus subscription

Products, Tools & Software Mentioned

Available with a Plus subscription