Search a title or topic

Over 20 million podcasts, powered by 

Player FM logo
Artwork

Content provided by Rahul Singh. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Rahul Singh or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.
Player FM - Podcast App
Go offline with the Player FM app!

Deterministic by Design: Why "Temp=0" Still Drifts and How to Fix It

24:58
 
Share
 

Manage episode 509444938 series 3693167
Content provided by Rahul Singh. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Rahul Singh or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://podcastplayer.com/legal.

Send us a text

Why do LLMs still give different answers even with temperature set to zero? In this episode of The Second Brain AI Podcast, we unpack new research from Thinking Machines Lab on defeating nondeterminism in LLM inference. We cover the surprising role of floating-point math, the real system-level culprit, lack of batch invariance, and how redesigned kernels can finally deliver bit-identical outputs. We also explore the trade-offs, real-world implications for testing and reliability, and how this breakthrough enables reproducible research and true on-policy reinforcement learning.

Sources:

  continue reading

10 episodes

Artwork
iconShare
 
Manage episode 509444938 series 3693167
Content provided by Rahul Singh. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Rahul Singh or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://podcastplayer.com/legal.

Send us a text

Why do LLMs still give different answers even with temperature set to zero? In this episode of The Second Brain AI Podcast, we unpack new research from Thinking Machines Lab on defeating nondeterminism in LLM inference. We cover the surprising role of floating-point math, the real system-level culprit, lack of batch invariance, and how redesigned kernels can finally deliver bit-identical outputs. We also explore the trade-offs, real-world implications for testing and reliability, and how this breakthrough enables reproducible research and true on-policy reinforcement learning.

Sources:

  continue reading

10 episodes

All episodes

×
 
Loading …

Welcome to Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Copyright 2025 | Privacy Policy | Terms of Service | | Copyright
Listen to this show while you explore
Play