Command Palette

Search for a command to run...

PodMine
Core Memory
Core Memory •January 22, 2026

He Left OpenAI To Think Bigger - EP 53 Jerry Tworek

Jerry Tworek, who recently left OpenAI after seven years, discusses his departure, the current state of AI research, and his desire to explore new, less conservative approaches to AI development.
AI & Machine Learning
Indie Hackers & SaaS Builders
Tech Policy & Ethics
Developer Culture
Ilya Sutskever
Ashley Vance
John Carmack
Jerry Tworek

Summary Sections

  • Podcast Summary
  • Speakers
  • Key Takeaways
  • Statistics & Facts
  • Compelling StoriesPremium
  • Thought-Provoking QuotesPremium
  • Strategies & FrameworksPremium
  • Similar StrategiesPlus
  • Additional ContextPremium
  • Key Takeaways TablePlus
  • Critical AnalysisPlus
  • Books & Articles MentionedPlus
  • Products, Tools & Software MentionedPlus
0:00/0:00

Timestamps are as accurate as they can be but may be slightly off. We encourage you to listen to the full context.

0:00/0:00

Podcast Summary

Jerry Tworek, a legendary AI researcher who recently left OpenAI after seven years, joins the Core Memory podcast for his "exit interview." (03:35) Tworek joined OpenAI in 2019 when it was just 30 people and worked on many of the company's most consequential products, including the reasoning technology that evolved from Q* to Strawberry to o1. In this revealing conversation, he discusses why he left OpenAI to pursue research that he felt the company couldn't support, his views on the current state of AI development, and his plans for the future.

  • Main Theme: The tension between commercial pressures and research innovation in AI labs, with Tworek advocating for more diverse approaches rather than all labs pursuing similar transformer scaling strategies.

Speakers

Jerry Tworek

Jerry Tworek is a renowned AI researcher from Poland who spent seven years at OpenAI, joining in 2019 when the company had around 30 employees. He led or contributed to many of OpenAI's breakthrough products, most notably the reasoning technology that began as Q* and eventually became o1 and Strawberry models. Before AI, Tworek worked in high-frequency trading and is known for his high risk tolerance and focus on foundational research breakthroughs.

Ashley Vance

Ashley Vance is the host of the Core Memory podcast and author covering technology and innovation. He has extensive experience covering the AI industry and has been following the development of major AI labs and their key researchers for years.

Kylie Robinson

Kylie Robinson is co-host of the Core Memory podcast, bringing fresh perspective and energy to covering the rapidly evolving AI landscape. She focuses on the human and business dynamics within the AI industry.

Key Takeaways

Research Innovation Requires Risk-Taking and Focus

Tworek emphasizes that meaningful research breakthroughs come from taking significant risks and maintaining deep focus on specific problems. (22:00) He argues that as AI companies grow larger and face commercial pressures, they naturally become more risk-averse, making it harder to pursue the kind of pioneering work that led to major breakthroughs like reinforcement learning scaling. The key is having conviction in your research direction and being willing to bet everything on it, even if it might fail. This approach contrasts sharply with spreading resources across many safe, incremental projects.

The AI Industry Has Become Too Homogeneous

All major AI labs are essentially doing the same thing - scaling transformers - which Tworek finds deeply concerning. (16:36) He argues that while competition is good, having five major companies all following identical approaches limits innovation. Most users can't even distinguish between different models despite teams thinking they're doing meaningfully different work. This lack of diversity in approaches means fewer opportunities for breakthrough discoveries that could fundamentally change the field.

Architectural Innovation Beyond Transformers is Necessary

Despite transformers' success over the past six years, Tworek believes the architecture has limitations and that new approaches are needed. (33:41) He advocates for exploring novel architectures that may look somewhat like transformers or completely different. The field has become too focused on incremental improvements to transformers rather than questioning whether this is the optimal architecture for all AI tasks. This represents one of his primary research interests going forward.

Continual Learning is Critical for True AGI

Current AI models operate with separate learning and inference modes, unlike humans who learn continuously. (33:55) Tworek identifies continual learning - the ability to learn from data in real-time during operation - as one of the final crucial capabilities needed for AGI. Without this ability, models remain fundamentally limited and "dumb" compared to human intelligence. Successfully integrating continual learning with current models would represent a major step toward genuine artificial general intelligence.

Video Games Offer Ideal Training Environments for AI

Tworek sees video games as uniquely valuable training environments because they're designed to be interesting to human intelligence. (26:55) Games incorporate storytelling, problem-solving, resource allocation, and puzzle-solving in ways that are engaging and non-repetitive. Unlike early reinforcement learning approaches that trained from scratch, combining game-based training with strong world knowledge from pre-training could create more capable agents. This approach leverages the fact that games are crafted to challenge and develop human cognitive abilities.

Statistics & Facts

  1. OpenAI grew from approximately 30 employees when Tworek joined in 2019 to around 1,000 employees today. (06:20) This represents more than 30x growth in just five years, illustrating the hyper-growth that Tworek describes as completely transforming the company annually.
  2. Tworek estimates that 99.9% of users cannot distinguish between different AI models despite teams believing they're doing meaningfully different work. (17:32) This statistic underscores his argument about the homogenization of AI development across major labs.
  3. The transformer architecture has been the dominant approach in AI for six years, with companies scaling it consistently. (12:43) This timeframe highlights how the field has converged on a single architectural approach for an extended period.

Compelling Stories

Available with a Premium subscription

Thought-Provoking Quotes

Available with a Premium subscription

Strategies & Frameworks

Available with a Premium subscription

Similar Strategies

Available with a Plus subscription

Additional Context

Available with a Premium subscription

Key Takeaways Table

Available with a Plus subscription

Critical Analysis

Available with a Plus subscription

Books & Articles Mentioned

Available with a Plus subscription

Products, Tools & Software Mentioned

Available with a Plus subscription

More episodes like this

"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis
February 1, 2026

The AI-Powered Biohub: Why Mark Zuckerberg & Priscilla Chan are Investing in Data, from Latent.Space

"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis
Lenny's Podcast: Product | Career | Growth
February 1, 2026

Dr. Becky on the surprising overlap between great parenting and great leadership

Lenny's Podcast: Product | Career | Growth
The Prof G Pod with Scott Galloway
February 1, 2026

First Time Founders: Has Substack Changed Media For Good?

The Prof G Pod with Scott Galloway
David Senra
February 1, 2026

Jimmy Iovine, Interscope Records & Beats by Dre

David Senra
Swipe to navigate