Search for a command to run...

Timestamps are as accurate as they can be but may be slightly off. We encourage you to listen to the full context.
In this enlightening episode, Karol Hausman and Tobi Springenberg from Physical Intelligence discuss their revolutionary approach to robotics through foundation models. (00:31) The duo explains why traditional robotics has been bottlenecked by intelligence rather than hardware limitations, and how their end-to-end learning approach is finally making real-world deployment possible. (31:00) Their latest model, π*0.6, demonstrates how reinforcement learning from experience can achieve robust performance, with robots successfully making coffee for thirteen hours straight and folding laundry for four hours without failure. (32:36) The conversation reveals how Physical Intelligence is moving beyond simple imitation learning to create truly generalizable robotic behaviors across different tasks and environments.
Co-founder of Physical Intelligence, Karol brings extensive expertise in robotics and machine learning to the foundation model approach. His background spans academic research and practical robotics applications, with a focus on developing scalable intelligence solutions for robotic systems.
Co-founder of Physical Intelligence, Tobi specializes in machine learning and AI systems. His work centers on bridging the gap between large-scale AI models and real-world robotic applications, particularly in developing training methodologies that enable robots to learn from experience.
The fundamental limitation in robotics has never been hardware capability but rather the intelligence layer. (03:25) As Hausman explains, robots have been capable of incredible tasks for over a decade when teleoperated by humans, demonstrating that the hardware exists but lacks autonomous intelligence. This insight led Physical Intelligence to focus exclusively on building the intelligence layer rather than developing application-specific robots, avoiding the common trap of becoming narrowly focused on single-use cases.
Traditional robotics broke down problems into perception, planning, and control modules, but this segmentation created artificial interfaces that limited performance. (18:00) The breakthrough came from realizing that humans don't think in terms of separate modules when picking up a glass - they simply act. Physical Intelligence's approach trains everything end-to-end, allowing the neural network to determine the optimal internal organization rather than imposing predetermined boundaries.
Successful robotic foundation models require diverse training data across different environments, tasks, and conditions rather than simply collecting more of the same type of data. (24:18) The team discovered that performance plateaus when collecting additional data using the same methods, emphasizing the critical importance of data diversity. This principle enabled their models to operate in completely new environments, like homes they'd never seen before, by learning from varied training scenarios.
The π*0.6 model represents a breakthrough in learning from the robot's own experience through reinforcement learning, moving beyond simple imitation of human demonstrations. (26:00) This approach allows robots to improve performance by learning from both successful actions and failures, with human corrections and reward signals guiding the learning process. The result is dramatically improved reliability and speed, with robots capable of operating for hours without failure.
Unlike locomotion tasks where simulation can be effective, manipulation tasks require real-world reinforcement learning due to the complexity of modeling how objects behave. (30:00) Springenberg illustrates this with examples like cardboard boxes sticking together in ways that simulations wouldn't predict. The long tail of real-world failures and edge cases makes it essential to train and deploy in actual environments rather than relying solely on simulated training.