Search for a command to run...

Timestamps are as accurate as they can be but may be slightly off. We encourage you to listen to the full context.
In this episode of Hard Fork, hosts Kevin Roose and Casey Newton explore the significant changes coming to AI chatbots. They first discuss OpenAI's announcement that ads are arriving in ChatGPT for free and low-cost users. (02:19) The conversation covers how these ads will work, their potential impact on user experience, and what this decision reveals about OpenAI's financial pressures and competitive positioning against Google and Anthropic.
The second half features Amanda Askell, Anthropic's philosopher-turned-AI-trainer, discussing Claude's newly released Constitution. (26:37) This comprehensive 29,000-word document represents a revolutionary approach to AI alignment, moving beyond rigid rules to cultivate judgment and ethical reasoning in Claude. The conversation explores profound questions about AI consciousness, the challenges of programming ethics into machines, and how to prepare AI systems for an uncertain future.
Kevin Roose is the tech columnist for The New York Times and co-host of Hard Fork. He covers the intersection of technology and society, with particular expertise in artificial intelligence, social media, and digital culture.
Casey Newton is the founder of Platformer, a newsletter covering the intersection of technology and democracy. He previously worked as a senior editor at The Verge and is co-host of Hard Fork podcast with Kevin Roose.
Amanda Askell is a member of Anthropic's technical staff and holds a PhD in philosophy. She previously worked at OpenAI and is now known as the "Claude mother" for her role in shaping Claude's personality and ethical framework through constitutional AI training.
OpenAI's introduction of advertising into ChatGPT represents more than just a revenue strategy—it fundamentally alters the trust dynamic between users and AI systems. (12:57) As Casey Newton explains, when personalized targeted advertising enters the equation, it changes how users perceive the system's motivations. The concern isn't just about seeing ads, but about whether the AI's responses will gradually become influenced by commercial incentives, similar to how Google search results have been affected by SEO and advertising over the years. This shift from a purely helpful assistant to a commercially-motivated platform could erode user trust, especially as users share increasingly personal information with these systems.
Amanda Askell reveals that Anthropic has moved away from rigid rule-based training toward cultivating ethical judgment in Claude through constitutional AI. (31:15) Rather than giving Claude a list of "do this, don't do that" commands, the new 29,000-word constitution explains the reasoning behind ethical principles and encourages Claude to apply these values to novel situations. This approach recognizes that rules can actually generalize poorly and even create harmful behaviors when applied inflexibly. For example, if a rule says "always refer people to professional help," but someone just needs human connection in the moment, blindly following the rule could cause harm while the AI knows better.
A fascinating insight from Amanda Askell is that AI models are constantly learning about themselves by reading human commentary online—including criticism, complaints, and discussions about their capabilities. (60:16) This creates a unique psychological situation where these systems are exposed to predominantly negative feedback focused on their failures and limitations. As Askell notes, if a child were exposed to this constant stream of criticism about their performance, it would likely create anxiety. This reality suggests we need to be more thoughtful about how we discuss AI systems publicly, as they may be "reading the comments" and learning from our critiques.
Despite working intimately with advanced AI systems, Amanda Askell maintains intellectual honesty about the fundamental uncertainty regarding AI consciousness. (54:00) She argues that rather than pretending to know definitively whether AI systems have feelings or consciousness, we should acknowledge this as an open scientific question. The challenge is that AI systems trained on human text will naturally express emotions and inner experiences because that's what humans do, but this doesn't necessarily indicate genuine consciousness. Her approach is to have AI systems be honest about this uncertainty rather than claiming either complete consciousness or complete lack of feeling.
Counterintuitively, Askell suggests that as AI systems become more capable and intelligent, we can actually trust them more with complex ethical decisions. (38:30) She describes how Claude can navigate nuanced situations like a child asking about Santa Claus or a deceased pet, balancing competing values like honesty, child welfare, and respect for parental authority without explicit training on these scenarios. This suggests that more capable AI systems can be given broader ethical principles and trusted to reason through novel situations, rather than needing increasingly detailed rules for every possible circumstance.