[State of Post-Training] From GPT-4.1 to 5.1: RLVR, Agent & Token Efficiency - Josh McGrath, OpenAI - Latent Space: The AI Engineer Podcast Recap

Podcast: Latent Space: The AI Engineer Podcast

Published: 2025-12-31

Duration: 28 min

Guests: Josh McGrath

Summary

Josh McGrath from OpenAI discusses the evolution from GPT-4.1 to 5.1, focusing on post-training advancements, particularly in reinforcement learning and token efficiency.

What Happened

Josh McGrath, a post-training researcher at OpenAI, dives into the developments from GPT-4.1 to GPT-5.1, highlighting the transition from non-thinking models to more sophisticated systems. He explains that while pre-training remains crucial, post-training offers a more exciting frontier with the potential for significant behavioral changes in AI models.

McGrath describes the complexities of reinforcement learning (RL) compared to pre-training, emphasizing the higher number of moving parts and infrastructure needed for RL tasks. He discusses the challenges of understanding unfamiliar code, whether it's internally developed or from external partners, and how tools like Codex have transformed his workflow by automating routine tasks.

The conversation touches on the release of a shopping model during the Black Friday period, highlighting new interaction paradigms, such as interruptability, that allow users to refine their searches dynamically. McGrath explains why this was launched as a separate model and draws parallels with the deep research model and GPT-5's high reasoning capabilities.

He addresses community debates around the spiritual successor of the deep research model and the introduction of personality toggles in the models, which provide users with more control over the AI's interaction style. McGrath prefers a tool-like interaction, aligning with the 'Anton versus Clippy' divide.

The episode covers the shift from traditional optimization papers to those focusing on input data quality, with McGrath noting that reinforcement learning through human feedback (RLHF) offers a spectrum of signal quality that needs more exploration. He also discusses the ongoing evolution of context windows and token efficiency, which significantly impact the user experience.

Josh McGrath reflects on the long-term future of AI, expressing uncertainty about the ultimate limits of context windows and token usage. He emphasizes the importance of co-designing systems and models and the need for educational systems to produce engineers skilled in both distributed systems and machine learning.

Finally, McGrath highlights the ongoing balance between pre-training and post-training investments, noting that while pre-training is not dead, the industry is witnessing a shift in computational resources towards post-training as both areas evolve.

Key Insights