Search for a command to run...

Timestamps are as accurate as they can be but may be slightly off. We encourage you to listen to the full context.
This Stack Overflow podcast episode was recorded live from AWS re:Invent 2025 in Las Vegas, featuring CEO Prashanth Chandrasekar and Director of Data Science Michael Foree discussing the key developments from the conference. (00:32) The conversation centers around AWS's major announcements, particularly their focus on AI agents, the trust challenges enterprises face with AI adoption, and the evolving job market in the age of AI and robotics.
• Main Theme: The intersection of AI agents, enterprise trust, and the future of work, with deep insights into how AWS's new frontier models and agent technologies are reshaping the tech landscape while addressing the critical need for guardrails and evaluation frameworks in enterprise AI deployment.
CEO of Stack Overflow, Prashanth brings extensive leadership experience in technology companies and has been instrumental in Stack Overflow's evolution into enterprise AI solutions. He has been a key figure in establishing Stack Overflow's partnerships with major cloud providers and AI companies, positioning the platform as a trusted data source for LLM training and enterprise knowledge management.
Director of Data Science at Stack Overflow, Michael leads the company's evaluation and analysis of LLM performance and AI integration strategies. His team is responsible for assessing where AI models excel and where they fall short, providing crucial insights for both Stack Overflow's product development and the broader developer community's understanding of AI capabilities.
Host of the Stack Overflow podcast and content strategist, Ryan brings technical journalism experience and deep knowledge of developer trends. He regularly interviews industry leaders and provides analysis on emerging technologies and their impact on the software development community.
AWS's announcement of three frontier agents (autonomous coding, security, and DevOps) signals a major shift from AI experimentation to production deployment. (01:41) Prashanth noted this aligns with enterprise conversations around Stack Internal, where companies are moving beyond proof-of-concepts to scaled implementations. This represents a critical inflection point where AI agents transition from novelty tools to essential business infrastructure, requiring organizations to develop proper governance and integration strategies.
Despite enthusiasm for AI, enterprise customers consistently cite trust as their primary concern when scaling AI implementations. (02:05) Matt Garman's "trust but verify" analogy comparing AI to raising teenagers resonates because enterprises need guardrails and evaluation frameworks before deploying AI at scale. Companies are running numerous AI experiments but struggle to expand them organization-wide due to trust, data governance, and infrastructure concerns, making 2026 potentially "the year of rationalization" for AI ROI.
While AI will automate certain tasks, the overall job market will likely expand with new types of roles emerging. (07:09) Prashanth observed that new companies in areas like life sciences are being created specifically because of AI capabilities that didn't exist before. However, the nature of work will change significantly - junior roles involving routine tasks may be reduced while new positions managing AI systems, ensuring data quality, and handling human-AI collaboration will emerge. The pattern mirrors previous technology waves like DevOps and outsourcing.
Unlike traditional software where inputs produce consistent outputs, LLMs require ongoing evaluation because they behave differently each time and their capabilities evolve rapidly. (19:45) Michael emphasized that organizations must document where LLMs are reliable versus unreliable, then repeat this evaluation every six months as models improve. This non-deterministic nature creates engineering stress but also represents an opportunity when properly managed with appropriate guardrails and human oversight.
While 95% of AI use cases reportedly fail to show ROI, this statistic is misleading because many successful AI implementations provide value that's difficult to quantify. (20:26) Michael found that people extensively use AI for work tasks like research, document writing, and decision support, but these don't translate to direct dollar values. Successful AI implementations often serve as "loss leaders" that improve user engagement and platform stickiness rather than generating immediate revenue, requiring new frameworks for measuring success.