Search for a command to run...

Timestamps are as accurate as they can be but may be slightly off. We encourage you to listen to the full context.
In this episode, Alex Kendall, CEO of Wayve, discusses the revolutionary shift from AV 1.0 to AV 2.0 - replacing hand-engineered autonomous vehicle systems with end-to-end neural networks. (02:01) Wayve has pioneered a contrarian approach since 2017, using one massive neural network instead of traditional perception-planning-control stacks that required HD maps and extensive infrastructure. The company has achieved remarkable scalability, launching in over 500 cities worldwide and partnering with major automotive OEMs like Nissan. (10:14) Kendall explains how world models enable sophisticated reasoning in complex driving scenarios, why partnerships with manufacturers create a sustainable path to scale beyond robotaxis, and how language integration opens new product possibilities for the future of embodied AI.
Alex Kendall is the CEO and founder of Wayve, which he established in 2017 with a contrarian vision to replace hand-engineered autonomous vehicle systems with end-to-end deep learning. Under his leadership, Wayve has become a pioneer in AV 2.0 technology, partnering with major automotive manufacturers like Nissan and expanding to over 500 cities globally. Alex has led the company's development of foundation models for embodied AI and world model approaches that enable vehicles to reason through complex driving scenarios.
Pat Grady is a host of the podcast and appears to be involved in venture capital or technology investing, based on his knowledgeable questioning about AI approaches and market dynamics in autonomous vehicles.
Sonya Huang is a co-host of the podcast who demonstrates deep expertise in AI and autonomous vehicle technologies, asking sophisticated questions about technical architectures and market strategies.
The fundamental breakthrough of AV 2.0 is replacing hand-engineered systems with a single neural network that can generalize to new scenarios without extensive infrastructure. (10:47) While AV 1.0 companies need to build HD maps city by city, Wayve's approach enables deployment in new locations within weeks rather than months or years. This generalization capability means the system can reason about novel situations it has never encountered in training data, such as construction workers or unusual road configurations, by understanding the underlying patterns of driving behavior rather than memorizing specific scenarios.
Wayve's development of generative world models, called GAIA, represents a crucial advancement in autonomous driving AI. (13:36) These models simulate multiple camera sensors and diverse environments, allowing the system to predict what will happen next and train through reinforcement learning. This approach enables sophisticated emergent behaviors like nudging forward at blind intersections until the car can see clearly, or slowing down in foggy conditions based on what the system can confidently reason about. The world models essentially give the AI an internal simulator for testing decisions before executing them.
Rather than building vertically integrated robotaxis, Wayve's strategy of partnering with automotive manufacturers provides a path to deploy on millions of vehicles annually. (22:08) With 90 million cars built each year globally, this approach avoids the limitations of city-by-city robotaxi deployments. The key insight is that modern vehicles already have the necessary hardware infrastructure - GPUs, surround cameras, and radar - making software integration possible without expensive hardware retrofits. This strategy enables rapid scale while leveraging manufacturers' existing supply chains and regulatory relationships.
Wayve's Lingo model, released as the first vision-language-action model in autonomous driving, demonstrates how language integration enhances both capability and user experience. (30:25) The system can not only drive but also converse about its decisions, explain risky situations, and even commentate drives. This creates opportunities for personalized driving styles, regulatory transparency, and natural human-robot interaction. Beyond improving representation learning, language alignment opens up a "chauffeur experience" where users can communicate preferences and understand the system's reasoning in natural language.
Success in autonomous driving AI depends more on data diversity than sheer volume, requiring sophisticated curation and filtering techniques. (15:22) Wayve aggregates data across different vehicle types, sensor configurations, countries, and driving scenarios, then uses unsupervised learning to identify unusual experiences and scenarios where the system performs poorly. This approach enables the AI to learn from edge cases and anomalies that would be impossible to encounter through any single data source, creating the generalization needed for global deployment.