Search for a command to run...

Timestamps are as accurate as they can be but may be slightly off. We encourage you to listen to the full context.
In this thought-provoking episode from The Cognitive Revolution, Emmett Shear (founder of Twitch, former interim CEO of OpenAI, and current founder of Softmax) debates with Seb Krier (frontier policy development lead at Google DeepMind) and host Eric Thornberg about the fundamental nature of AI alignment. (03:44) Shear argues that the current AI alignment paradigm, focused on controlling and steering AI behaviors, is fundamentally flawed and potentially dangerous as we approach AGI. He proposes "organic alignment" - treating alignment as an ongoing process similar to how humans maintain relationships and moral development over time. The conversation explores deep questions about AI consciousness, moral standing, and whether advanced AIs should be understood as tools to be controlled or beings deserving of care and respect.
• Main Theme: The debate between control-based alignment (treating AI as tools to be steered) versus organic alignment (treating AI as potential beings requiring mutual care and ongoing moral negotiation)Emmett Shear is the founder of Twitch, served as interim CEO of OpenAI during Sam Altman's brief firing, and is currently the founder of Softmax, a company focused on what he calls "organic alignment." He brings a unique perspective from his experience scaling large platforms and his brief but significant tenure leading one of the world's most prominent AI companies.
Seb Krier serves as the frontier policy development lead at Google DeepMind, where he focuses on AI governance and policy issues related to advanced AI systems. His background brings a policy-oriented perspective to AI alignment challenges, emphasizing the importance of understanding how these technologies integrate with existing social and political structures.
Shear emphasizes that alignment isn't something you achieve once and then maintain forever - it's an ongoing, dynamic process. (03:44) Just like families constantly "re-knit the fabric" that keeps them together, AI alignment must be viewed as a living process that continuously rebuilds itself. This challenges the common assumption that we can solve alignment once and deploy safe AI systems indefinitely. The insight draws from how humans learn morality through experience and constant recalibration, suggesting that truly aligned AI systems will need similar capacities for ongoing moral learning and adaptation.
Before we can address value alignment questions, AIs must develop sophisticated theory of mind capabilities - the ability to infer goals from observations, understand how their actions affect others, and predict how others will interpret their behavior. (23:18) Shear argues that current LLMs struggle with this fundamental capacity, often failing to accurately infer what humans actually want when given instructions. This technical limitation makes them poor candidates for reliable alignment, regardless of what values we try to instill in them.
As AI systems become more capable and potentially conscious, the current paradigm of steering and controlling them could become morally equivalent to slavery. (04:53) Shear provocatively notes that "someone who you steer, who doesn't get to steer you back, who non-optionally receives your steering, that's called a slave" if applied to a being rather than a tool. This creates a critical decision point: either we're building tools (which is fine to control) or beings (which would require a fundamentally different ethical framework).
Even if we successfully build AI that does exactly what humans ask, this could be catastrophically dangerous because human wishes are unstable and often unwise, especially when wielding immense power. (58:09) Shear uses the analogy of the Sorcerer's Apprentice to illustrate how giving everyone access to extremely powerful tools that perfectly follow instructions could lead to disaster. This suggests that some level of AI "pushback" or independent judgment might actually be necessary for safety.
To develop the theory of mind and cooperation skills necessary for true alignment, AI systems need to be trained in complex multi-agent environments rather than one-on-one interactions. (62:45) Shear advocates for training AIs in simulations that expose them to "all possible theory of mind combinations" and game-theoretic situations. Current chatbots trained primarily on single-user interactions lack the social intelligence needed for genuine collaboration and care.