Search for a command to run...

Timestamps are as accurate as they can be but may be slightly off. We encourage you to listen to the full context.
In this insightful episode of the a16z podcast, Beyang Liu, CTO and co-founder of Sourcegraph, reveals how AI agents are fundamentally transforming software development. (10:00) Liu shares that over 90% of his code now comes from agents, demonstrating the dramatic shift in how developers work. The conversation with Martin Casado and Guido Appenzeller explores the evolution from code search to AI-powered coding agents, the strategic implications of Chinese dominance in open-source AI models, and the unintended consequences of AI safety regulations. (38:10) Liu argues that the "Terminator narrative" around AI has created policy environments that may have inadvertently handed the infrastructure layer of the AI revolution to Chinese labs, one fine-tuned model at a time.
Beyang Liu is the co-founder and CTO of Sourcegraph, where he has spent over a decade building developer tools. He previously worked as a developer at Palantir in its early days, where he met his co-founder Quinn while working on data analysis software for large enterprise codebases. Liu studied machine learning under Daphne Koller at Stanford, focusing on computer vision research, making him an "OG AI" expert who brings both systems thinking and deep AI knowledge to the current coding revolution.
Martin Casado is a General Partner at Andreessen Horowitz (a16z), where he focuses on enterprise software investments. He is known for his deep technical expertise and strategic insights into emerging technology trends, particularly in the intersection of AI and enterprise software development.
Guido Appenzeller is a General Partner at Andreessen Horowitz (a16z), bringing extensive experience in technology and venture capital. He contributes valuable perspectives on the investment and strategic implications of AI technology developments in the software engineering space.
Liu reveals a fundamental shift in computing: for the first time in computer science history, we've abdicated correctness and logic to a third party. (15:03) Unlike traditional computing where inputs produce predictable outputs, AI agents introduce nondeterminism into core software development. The new atomic unit isn't a function call but a "stochastic subroutine" - an agent that might take different paths to solve the same problem. This requires developers to think differently about reliability, moving from deterministic guarantees to probabilistic confidence levels. Liu notes he has 99% confidence his search agent will find what he needs, even though the path it takes varies each time.
Liu shares his personal transformation: over 90% of his code now comes through agents like AMP, fundamentally changing his role from line-by-line coder to orchestrator. (29:52) This shift creates a bittersweet experience for developers - unprecedented productivity coupled with the loss of coding's creative joy. Many developers report feeling like "middle managers of coding," spending most time reviewing agent-generated code rather than creating it. The future interface won't be traditional IDEs or terminals, but tools that help humans orchestrate multiple agents while understanding their outputs. This represents a level-up for developers, moving from implementation details to high-level problem solving and system architecture.
Rather than optimizing for a single model, successful AI applications require understanding that different agents need different trade-offs between intelligence, speed, and cost. (22:20) Sourcegraph uses frontier models like Claude Sonnet for complex reasoning tasks, but switches to smaller, faster models for targeted edits and specific sub-agents. Liu explains that once you reach a quality ceiling for simpler tasks, optimizing for latency becomes more valuable than additional intelligence. This approach allows for both a premium "smart agent" and a fast, ad-supported agent, serving different use cases and user preferences while maximizing the efficiency of each specialized task.
A concerning strategic reality has emerged: while the US leads in chips, frontier intelligence, and overall AI innovation, the most effective open-weight models for agentic workloads are predominantly of Chinese origin. (34:48) Liu notes that in Sourcegraph's extensive testing across the model landscape, Chinese models like QwenCoder, GLM, and others consistently outperform American open-source alternatives for tool use and agentic applications. This creates a dependency problem as application builders worldwide increasingly choose to fine-tune on Chinese base models. The irony is stark: America invented the AI revolution but is losing the open-source infrastructure layer that will power future applications.
The "Terminator narrative" around AI safety - positioning AI as either utopian salvation or existential threat - has created policy paralysis that benefits competitors. (38:34) Liu argues this narrative, while largely debunked among practitioners who use AI daily, has taken on a life of its own in policy circles. The focus on existential risk at the model layer has led to regulatory uncertainty that makes American companies gun-shy about releasing open-source models. Meanwhile, Chinese labs continue developing and releasing competitive open-weight models. This regulatory fear isn't hypothetical - it's already reshaping the competitive landscape, potentially handing the infrastructure advantage to China while American companies remain paralyzed by liability concerns and patchwork state regulations.