Search a title or topic

Over 20 million podcasts, powered by 

Player FM logo
Artwork

Content provided by NeurIPS 2025 by Basis Set and Basis Set. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by NeurIPS 2025 by Basis Set and Basis Set or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://podcastplayer.com/legal.
Player FM - Podcast App
Go offline with the Player FM app!

Computer Vision's Journey

14:17
 
Share
 

Manage episode 523382910 series 2783843
Content provided by NeurIPS 2025 by Basis Set and Basis Set. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by NeurIPS 2025 by Basis Set and Basis Set or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://podcastplayer.com/legal.
AI vision is solved. AI reasoning is not. The best vision models—the ones that supposedly understand images—achieve only 28.8% accuracy on tasks requiring physics, time, and causality. You'll trace the journey from 2015's Faster R-CNN breakthrough (56,700+ citations) through the evolution from messy multi-step pipelines to elegant end-to-end deep learning, only to discover the humbling reality: AI can classify objects brilliantly but can't reason about what it sees. Worse, there's a "reasoning illusion"—models get right answers through wrong processes. This episode shows you why the gap between perception and understanding matters. Topics Covered - Faster R-CNN: The breakthrough that gave AI eyes - Region Proposal Networks explained simply - The reasoning gap: classification ≠ understanding - RiseBench: Testing temporal, causal, spatial, and logical reasoning - World models for self-driving (Gaia 2) - The "reasoning illusion": right answers, wrong process - Process Verified Accuracy: checking the work, not just the answer
  continue reading

23 episodes

Artwork
iconShare
 
Manage episode 523382910 series 2783843
Content provided by NeurIPS 2025 by Basis Set and Basis Set. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by NeurIPS 2025 by Basis Set and Basis Set or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://podcastplayer.com/legal.
AI vision is solved. AI reasoning is not. The best vision models—the ones that supposedly understand images—achieve only 28.8% accuracy on tasks requiring physics, time, and causality. You'll trace the journey from 2015's Faster R-CNN breakthrough (56,700+ citations) through the evolution from messy multi-step pipelines to elegant end-to-end deep learning, only to discover the humbling reality: AI can classify objects brilliantly but can't reason about what it sees. Worse, there's a "reasoning illusion"—models get right answers through wrong processes. This episode shows you why the gap between perception and understanding matters. Topics Covered - Faster R-CNN: The breakthrough that gave AI eyes - Region Proposal Networks explained simply - The reasoning gap: classification ≠ understanding - RiseBench: Testing temporal, causal, spatial, and logical reasoning - World models for self-driving (Gaia 2) - The "reasoning illusion": right answers, wrong process - Process Verified Accuracy: checking the work, not just the answer
  continue reading

23 episodes

All episodes

×
 
Loading …

Welcome to Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Copyright 2025 | Privacy Policy | Terms of Service | | Copyright
Listen to this show while you explore
Play