THE FUTURE IS HERE

Ex-OpenAI Researcher On Why He Left, His Honest AGI Timeline, & The Limits of Scaling RL

This episode features Jerry Tworek, a key architect behind OpenAI’s breakthrough reasoning models (o1, o3) and Codex, discussing the current state and future of AI. Jerry explores the real limits and promise of scaling pre-training and reinforcement learning, arguing that while these paradigms deliver predictable improvements, they’re fundamentally constrained by data availability and struggle with generalization beyond their training objectives. He reveals his updated belief that continual learning—the ability for models to update themselves based on failure and work through problems autonomously—is necessary for AGI, as current models hit walls and become “hopeless” when stuck. Jerry discusses the convergence of major labs toward similar approaches driven by economic forces, the tension between exploration and exploitation in research, and why he left OpenAI to pursue new research directions. He offers candid insights on the competitive dynamics between labs, the focus required to win in specific domains like coding, what makes great AI researchers, and his surprisingly near-term predictions for robotics (2-3 years) while warning about the societal implications of widespread work automation that we’re not adequately preparing for.

0:00 Intro
1:26 Scaling Paradigms in AI
3:36 Challenges in Reinforcement Learning
11:48 AGI Timelines
18:36 Converging Labs and Economic Forces
25:05 Jerry’s Departure from OpenAI
31:18 Pivotal Decisions in OpenAI’s Journey
35:06 Balancing Research and Product Development
38:42 The Future of AI Coding
41:33 Specialization vs. Generalization in AI
48:47 Hiring and Building Research Teams
55:21 Quickfire

With your co-hosts: 
@jacobeffron 
– Partner at Redpoint, Former PM Flatiron Health 
@patrickachase 
– Partner at Redpoint, Former ML Engineer LinkedIn 
@ericabrescia 
– Former COO Github, Founder Bitnami (acq’d by VMWare) 
@jordan_segall 
– Partner at Redpoint