Search a title or topic

Over 20 million podcasts, powered by 

Player FM logo
Artwork

Content provided by Lightspeed Venture Partners. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Lightspeed Venture Partners or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://podcastplayer.com/legal.
Player FM - Podcast App
Go offline with the Player FM app!

Inside the Black Box: The Urgency of AI Interpretability

1:02:17
 
Share
 

Manage episode 509930947 series 3619430
Content provided by Lightspeed Venture Partners. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Lightspeed Venture Partners or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://podcastplayer.com/legal.

Recorded live at Lightspeed’s offices in San Francisco, this special episode of Generative Now dives into the urgency and promise of AI interpretability. Lightspeed partner Nnamdi Iregbulem spoke with Anthropic researcher Jack Lindsey and Goodfire co-founder and Chief Scientist Tom McGrath, who previously co-founded Google DeepMind’s interpretability team. They discuss opening the black box of modern AI models in order to understand their reliability and spot real-world safety concerns, in order to build AI systems of the future that we can trust.

Episode Chapters:

00:42 Welcome and Introduction

00:36 Overview of Lightspeed and AI Investments

03:19 Event Agenda and Guest Introductions

05:35 Discussion on Interpretability in AI

18:44 Technical Challenges in AI Interpretability

29:42 Advancements in Model Interpretability

30:05 Smarter Models and Interpretability

31:26 Models Doing the Work for Us

32:43 Real-World Applications of Interpretability

34:32 Philanthropics' Approach to Interpretability

39:15 Breakthrough Moments in AI Interpretability

44:41 Challenges and Future Directions

48:18 Neuroscience and Model Training Insights

54:42 Emergent Misalignment and Model Behavior

01:01:30 Concluding Thoughts and Networking

Stay in touch:

The content here does not constitute tax, legal, business or investment advice or an offer to provide such advice, should not be construed as advocating the purchase or sale of any security or investment or a recommendation of any company, and is not an offer, or solicitation of an offer, for the purchase or sale of any security or investment product. For more details please see lsvp.com/legal.

  continue reading

90 episodes

Artwork
iconShare
 
Manage episode 509930947 series 3619430
Content provided by Lightspeed Venture Partners. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Lightspeed Venture Partners or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://podcastplayer.com/legal.

Recorded live at Lightspeed’s offices in San Francisco, this special episode of Generative Now dives into the urgency and promise of AI interpretability. Lightspeed partner Nnamdi Iregbulem spoke with Anthropic researcher Jack Lindsey and Goodfire co-founder and Chief Scientist Tom McGrath, who previously co-founded Google DeepMind’s interpretability team. They discuss opening the black box of modern AI models in order to understand their reliability and spot real-world safety concerns, in order to build AI systems of the future that we can trust.

Episode Chapters:

00:42 Welcome and Introduction

00:36 Overview of Lightspeed and AI Investments

03:19 Event Agenda and Guest Introductions

05:35 Discussion on Interpretability in AI

18:44 Technical Challenges in AI Interpretability

29:42 Advancements in Model Interpretability

30:05 Smarter Models and Interpretability

31:26 Models Doing the Work for Us

32:43 Real-World Applications of Interpretability

34:32 Philanthropics' Approach to Interpretability

39:15 Breakthrough Moments in AI Interpretability

44:41 Challenges and Future Directions

48:18 Neuroscience and Model Training Insights

54:42 Emergent Misalignment and Model Behavior

01:01:30 Concluding Thoughts and Networking

Stay in touch:

The content here does not constitute tax, legal, business or investment advice or an offer to provide such advice, should not be construed as advocating the purchase or sale of any security or investment or a recommendation of any company, and is not an offer, or solicitation of an offer, for the purchase or sale of any security or investment product. For more details please see lsvp.com/legal.

  continue reading

90 episodes

All episodes

×
 
Loading …

Welcome to Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Copyright 2025 | Privacy Policy | Terms of Service | | Copyright
Listen to this show while you explore
Play