A podcast about the fundamentals of safe and resilient modeling systems behind the AI that impacts our lives and our businesses.
…
continue reading
Sid Mangalik Podcasts
1
Why validity beats scale when building multi‑step AI systems
40:16
40:16
Play later
Play later
Lists
Like
Liked
40:16In this episode, Dr. Sebastian (Seb) Benthall joins us to discuss research from his and Andrew's paper entitled “Validity Is What You Need” for agentic AI that actually works in the real world. Our discussion connects systems engineering, mechanism design, and requirements to multi‑step AI that creates enterprise impact to achieve measurable outcom…
…
continue reading
1
2025 AI review: Why LLMs stalled and the outlook for 2026
42:06
42:06
Play later
Play later
Lists
Like
Liked
42:06Here it is! We review the year where scaling large AI models hit its ceiling, Google reclaimed momentum with efficient vertical integration, and the market shifted from hype to viability. Join us as we talk about why human-in-the-loop is failing, why generative AI agents validating other agents compounds errors, and how small expert data quietly be…
…
continue reading
1
Big data, small data, and AI oversight with David Sandberg
49:48
49:48
Play later
Play later
Lists
Like
Liked
49:48In this episode, we look at the actuarial principles that make models safer: parallel modeling, small data with provenance, and real-time human supervision. To help us, long-time insurtech and startup advisor David Sandberg, FSA, MAAA, CERA, joins us to share more about his actuarial expertise in data management and AI. We also challenge the hype a…
…
continue reading
1
Metaphysics and modern AI: What is space and time?
38:04
38:04
Play later
Play later
Lists
Like
Liked
38:04We explore how space and time form a single fabric, testing our daily beliefs through questions about free-fall, black holes, speed, and momentum to reveal what models get right and where they break. To help us, we’re excited to have our friend David Theriault, a science and sci-fi afficionado; and our resident astrophysicist, Rachel Losacco, to ta…
…
continue reading
1
Metaphysics and modern AI: What is reality?
38:32
38:32
Play later
Play later
Lists
Like
Liked
38:32In the first episode of our series on metaphysics, Michael Herman joins us from Episode #14 on “What is consciousness?” to discuss reality. More specifically, the question of objects in reality. The team explores Plato’s forms, Aristotle’s realism, emergence, and embodiment to determine whether AI models can approximate from what humans uniquely ex…
…
continue reading
1
Metaphysics and modern AI: What is thinking? - Series Intro
16:19
16:19
Play later
Play later
Lists
Like
Liked
16:19By Dr. Andrew Clark & Sid Mangalik
…
continue reading
1
AI in practice: Guardrails and security for LLMs
35:11
35:11
Play later
Play later
Lists
Like
Liked
35:11In this episode, we talk about practical guardrails for LLMs with data scientist Nicholas Brathwaite. We focus on how to stop PII leaks, retrieve data, and evaluate safety with real limits. We weigh managed solutions like AWS Bedrock against open-source approaches and discuss when to skip LLMs altogether. • Why guardrails matter for PII, secrets, a…
…
continue reading
1
AI in practice: LLMs, psychology research, and mental health
42:28
42:28
Play later
Play later
Lists
Like
Liked
42:28We’re excited to have Adi Ganesan, a PhD researcher at Stony Brook University, the University of Pennsylvania, and Vanderbilt, on the show. We’ll talk about how large language models LLMs) are being tested and used in psychology, citing examples from mental health research. Fun fact: Adi was Sid's research partner during his Ph.D. program. Discussi…
…
continue reading
1
LLM scaling: Is GPT-5 near the end of exponential growth?
22:42
22:42
Play later
Play later
Lists
Like
Liked
22:42The release of OpenAI GPT-5 marks a significant turning point in AI development, but maybe not the one most enthusiasts had envisioned. The latest version seems to reveal the natural ceiling of current language model capabilities with incremental rather than revolutionary improvements over GPT-4. Sid and Andrew call back to some of the model-buildi…
…
continue reading
1
AI governance: Building smarter AI agents from the fundamentals, part 4
37:25
37:25
Play later
Play later
Lists
Like
Liked
37:25Sid Mangalik and Andrew Clark explore the unique governance challenges of agentic AI systems, highlighting the compounding error rates, security risks, and hidden costs that organizations must address when implementing multi-step AI processes. Show notes: • Agentic AI systems require governance at every step: perception, reasoning, action, and lear…
…
continue reading
1
Linear programming: Building smarter AI agents from the fundamentals, part 3
29:46
29:46
Play later
Play later
Lists
Like
Liked
29:46We continue with our series about building agentic AI systems from the ground up and for desired accuracy. In this episode, we explore linear programming and optimization methods that enable reliable decision-making within constraints. Show notes: Linear programming allows us to solve problems with multiple constraints, like finding optimal flights…
…
continue reading
1
Utility functions: Building smarter AI agents from the fundamentals, part 2
41:36
41:36
Play later
Play later
Lists
Like
Liked
41:36The hosts look at utility functions as the mathematical basis for making AI systems. They use the example of a travel agent that doesn’t get tired and can be increased indefinitely to meet increasing customer demand. They also discuss the difference between this structured, economic-based approach with the problems of using large language models fo…
…
continue reading
1
Mechanism design: Building smarter AI agents from the fundamentals, Part 1
37:06
37:06
Play later
Play later
Lists
Like
Liked
37:06What if we've been approaching AI agents all wrong? While the tech world obsesses over larger language models (LLMs) and prompt engineering, there'a a foundational approach that could revolutionize how we build trustworthy AI systems: mechanism design. This episode kicks off an exciting series where we're building AI agents "the hard way"—using pri…
…
continue reading
1
Principles, agents, and the chain of accountability in AI systems
46:26
46:26
Play later
Play later
Lists
Like
Liked
46:26Dr. Michael Zargham provides a systems engineering perspective on AI agents, emphasizing accountability structures and the relationship between principals who deploy agents and the agents themselves. In this episode, he brings clarity to the often misunderstood concept of agents in AI by grounding them in established engineering principles rather t…
…
continue reading
1
Supervised machine learning for science with Christoph Molnar and Timo Freiesleben, Part 2
41:58
41:58
Play later
Play later
Lists
Like
Liked
41:58Part 2 of this series could have easily been renamed "AI for science: The expert’s guide to practical machine learning.” We continue our discussion with Christoph Molnar and Timo Freiesleben to look at how scientists can apply supervised machine learning techniques from the previous episode into their research. Introduction to supervised ML for sci…
…
continue reading
1
Supervised machine learning for science with Christoph Molnar and Timo Freiesleben, Part 1
27:29
27:29
Play later
Play later
Lists
Like
Liked
27:29Machine learning is transforming scientific research across disciplines, but many scientists remain skeptical about using approaches that focus on prediction over causal understanding. That’s why we are excited to have Christoph Molnar return to the podcast with Timo Freiesleben. They are co-authors of "Supervised Machine Learning for Science: How …
…
continue reading
1
The future of AI: Exploring modeling paradigms
33:42
33:42
Play later
Play later
Lists
Like
Liked
33:42Unlock the secrets to AI's modeling paradigms. We emphasize the importance of modeling practices, how they interact, and how they should be considered in relation to each other before you act. Using the right tool for the right job is key. We hope you enjoy these examples of where the greatest AI and machine learning techniques exist in your routin…
…
continue reading
Agentic AI is the latest foray into big-bet promises for businesses and society at large. While promising autonomy and efficiency, AI agents raise fundamental questions about their accuracy, governance, and the potential pitfalls of over-reliance on automation. Does this story sound vaguely familiar? Hold that thought. This discussion about the ove…
…
continue reading
1
Contextual integrity and differential privacy: Theory vs. application with Sebastian Benthall
32:32
32:32
Play later
Play later
Lists
Like
Liked
32:32What if privacy could be as dynamic and socially aware as the communities it aims to protect? Sebastian Benthall, a senior research fellow from NYU’s Information Law Institute, shows us how privacy is complex. He uses Helen Nissenbaum’s work with contextual integrity and concepts in differential privacy to explain the complexity of privacy. Our tal…
…
continue reading
1
Model documentation: Beyond model cards and system cards in AI governance
27:43
27:43
Play later
Play later
Lists
Like
Liked
27:43What if the secret to successful AI governance lies in understanding the evolution of model documentation? In this episode, our hosts challenge the common belief that model cards marked the start of documentation in AI. We explore model documentation practices, from their crucial beginnings in fields like finance to their adaptation in Silicon Vall…
…
continue reading
1
New paths in AI: Rethinking LLMs and model risk strategies
39:51
39:51
Play later
Play later
Lists
Like
Liked
39:51Are businesses ready for large language models as a path to AI? In this episode, the hosts reflect on the past year of what has changed and what hasn’t changed in the world of LLMs. Join us as we debunk the latest myths and emphasize the importance of robust risk management in AI integration. The good news is that many decisions about adoption have…
…
continue reading
1
Complex systems: What data science can learn from astrophysics with Rachel Losacco
41:02
41:02
Play later
Play later
Lists
Like
Liked
41:02Our special guest, astrophysicist Rachel Losacco, explains the intricacies of galaxies, modeling, and the computational methods that unveil their mysteries. She shares stories about how advanced computational resources enable scientists to decode galaxy interactions over millions of years with true-to-life accuracy. Sid and Andrew discuss transfera…
…
continue reading
1
Preparing AI for the unexpected: Lessons from recent IT incidents
34:13
34:13
Play later
Play later
Lists
Like
Liked
34:13Can your AI models survive a big disaster? While a recent major IT incident with CrowdStrike wasn't AI related, the magnitude and reaction reminded us that no system no matter how proven is immune to failure. AI modeling systems are no different. Neglecting the best practices of building models can lead to unrecoverable failures. Discover how the t…
…
continue reading
1
Exploring the NIST AI Risk Management Framework (RMF) with Patrick Hall
41:24
41:24
Play later
Play later
Lists
Like
Liked
41:24Join us as we chat with Patrick Hall, Principal Scientist at Hallresearch.ai and Assistant Professor at George Washington University. He shares his insights on the current state of AI, its limitations, and the potential risks associated with it. The conversation also touched on the importance of responsible AI, the role of the National Institute of…
…
continue reading
1
Data lineage and AI: Ensuring quality and compliance with Matt Barlin
28:29
28:29
Play later
Play later
Lists
Like
Liked
28:29Ready to uncover the secrets of modern systems engineering and the future of AI? Join us for an enlightening conversation with Matt Barlin, the Chief Science Officer of Valence. Matt's extensive background in systems engineering and data lineage sets the stage for a fascinating discussion. He sheds light on the historical evolution of the field, th…
…
continue reading
1
Differential privacy: Balancing data privacy and utility in AI
28:17
28:17
Play later
Play later
Lists
Like
Liked
28:17Explore the basics of differential privacy and its critical role in protecting individual anonymity. The hosts explain the latest guidelines and best practices in applying differential privacy to data for models such as AI. Learn how this method also makes sure that personal data remains confidential, even when datasets are analyzed or hacked. Show…
…
continue reading
1
Responsible AI: Does it help or hurt innovation? With Anthony Habayeb
45:59
45:59
Play later
Play later
Lists
Like
Liked
45:59Artificial Intelligence (AI) stands at a unique intersection of technology, ethics, and regulation. The complexities of responsible AI are brought into sharp focus in this episode featuring Anthony Habayeb, CEO and co-founder of Monitaur, As responsible AI is scrutinized for its role in profitability and innovation, Anthony and our hosts discuss th…
…
continue reading
1
Baseline modeling and its critical role in AI and business performance
36:23
36:23
Play later
Play later
Lists
Like
Liked
36:23Baseline modeling is a necessary part of model validation. In our expert opinion, it should be required before model deployment. There are many baseline modeling types and in this episode, we're discussing their use cases, strengths, and weaknesses. We're sure you'll appreciate a fresh take on how to improve your modeling practices. Show notes Intr…
…
continue reading
1
Information theory and the complexities of AI model monitoring
21:56
21:56
Play later
Play later
Lists
Like
Liked
21:56In this episode, we explore information theory and the not-so-obvious shortcomings of its popular metrics for model monitoring; and where non-parametric statistical methods can serve as the better option. Introduction and latest news 0:03 Gary Marcus has written an article questioning the hype around generative AI, suggesting it may not be as trans…
…
continue reading
1
The importance of anomaly detection in AI
35:48
35:48
Play later
Play later
Lists
Like
Liked
35:48In this episode, the hosts focus on the basics of anomaly detection in machine learning and AI systems, including its importance, and how it is implemented. They also touch on the topic of large language models, the (in)accuracy of data scraping, and the importance of high-quality data when employing various detection methods. You'll even gain some…
…
continue reading
1
What is consciousness, and does AI have it?
32:55
32:55
Play later
Play later
Lists
Like
Liked
32:55We're taking a slight detour from modeling best practices to explore questions about AI and consciousness. With special guest Michael Herman, co-founder of Monitaur and TestDriven.io, the team discusses different philosophical perspectives on consciousness and how these apply to AI. They also discuss the potential dangers of AI in its current state…
…
continue reading
1
Upskilling for AI: Roles, organizations, and new mindsets
41:23
41:23
Play later
Play later
Lists
Like
Liked
41:23Data scientists, researchers, engineers, marketers, and risk leaders find themselves at a crossroads to expand their skills or risk obsolescence. The hosts discuss how a growth mindset and "the fundamentals" of AI can help. Our episode shines a light on this vital shift, equipping listeners with strategies to elevate their skills and integrate mult…
…
continue reading
Get ready for 2024 and a brand new episode! We discuss non-parametric statistics in data analysis and AI modeling. Learn more about applications in user research methods, as well as the importance of key assumptions in statistics and data modeling that must not be overlooked, After you listen to the episode, be sure to check out the supplement mate…
…
continue reading
1
AI regulation, data privacy, and ethics - 2023 summarized
31:27
31:27
Play later
Play later
Lists
Like
Liked
31:27It's the end of 2023 and our first season. The hosts reflect on what's happened with the fundamentals of AI regulation, data privacy, and ethics. Spoiler alert: a lot! And we're excited to share our outlook for AI in 2024. AI regulation and its impact in 2024. Hosts reflect on AI regulation discussions from their first 10 episodes, discussing what …
…
continue reading
1
Managing bias in the actuarial sciences with Joshua Pyle, FCAS
43:41
43:41
Play later
Play later
Lists
Like
Liked
43:41Joshua Pyle joins us in a discussion about managing bias in the actuarial sciences. Together with Andrew's and Sid's perspectives from both the economic and data science fields, they deliver an interdisciplinary conversation about bias that you'll only find here. OpenAI news plus new developments in language models. 0:03 The hosts get to discuss th…
…
continue reading
Episode 9. Continuing our series run about model validation. In this episode, the hosts focus on aspects of performance, why we need to do statistics correctly, and not use metrics without understanding how they work, to ensure that models are evaluated in a meaningful way. AI regulations, red team testing, and physics-based modeling. 0:03 The host…
…
continue reading
1
Model validation: Robustness and resilience
36:23
36:23
Play later
Play later
Lists
Like
Liked
36:23Episode 8. This is the first in a series of episodes dedicated to model validation. Today, we focus on model robustness and resilience. From complex financial systems to why your gym might be overcrowded at New Year's, you've been directly affected by these aspects of model validation. AI hype and consumer trust (0:03) FTC article highlights consum…
…
continue reading
Episode 7. To use or not to use? That is the question about digital twins that the fundamentalists explore. Many solutions continue to be proposed for making AI systems safer, but can digital twins really deliver for AI what we know they can do for physical systems? Tune in and find out. Show notes Digital twins by definition. 0:03 Digital twins ar…
…
continue reading
Episode 6. What does systems engineering have to do with AI fundamentals? In this episode, the team discusses what data and computer science as professions can learn from systems engineering, and how the methods and mindset of the latter can boost the quality of AI-based innovations. Show notes News and episode commentary 0:03 ChatGPT usage is down…
…
continue reading
Episode 5. This episode about synthetic data is very real. The fundamentalists uncover the pros and cons of synthetic data; as well as reliable use cases and the best techniques for safe and effective use in AI. When even SAG-AFTRA and OpenAI make synthetic data a household word, you know this is an episode you can't miss. Show notes What is synthe…
…
continue reading
Episode 4. The AI Fundamentalists welcome Christoph Molnar to discuss the characteristics of a modeling mindset in a rapidly innovating world. He is the author of multiple data science books including Modeling Mindsets, Interpretable Machine Learning, and his latest book Introduction to Conformal Prediction with Python. We hope you enjoy this enlig…
…
continue reading
1
Why data matters | The right data for the right objective with AI
36:37
36:37
Play later
Play later
Lists
Like
Liked
36:37Episode 3. Get ready because we're bringing stats back! An AI model can only learn from the data it has seen. And business problems can’t be solved without the right data. The Fundamentalists break down the basics of data from collection to regulation to bias to quality in AI. Introduction to this episode Why data matters. How do big tech's LLM mod…
…
continue reading
1
Truth-based AI: LLMs and knowledge graphs - back to basics
30:56
30:56
Play later
Play later
Lists
Like
Liked
30:56Truth-based AI: Large language models (LLMs) and knowledge graphs - The AI Fundamentalists, Episode 2 Show Notes What’s NOT new and what is new in the world of LLMs. 3:10 Getting back to the basics of modeling best practices and rigor. What is AI and subsequently LLM regulation going to look like for tech organizations? 5:55 Recommendations for rea…
…
continue reading
1
Why AI Fundamentals? | AI rigor in engineering | Generative AI isn't new | Data quality matters in machine learning
26:03
26:03
Play later
Play later
Lists
Like
Liked
26:03The AI Fundamentalists - Ep1 Summary Welcome to the first episode. 0:03 Welcome to the first episode of the AI Fundamentalists podcast. Introducing the hosts. Introducing Sid and Andrew. 1:23 Introducing Andrew Clark, co-founder and CTO of Monitaur. Introduction of the podcast topic. What is the proper rigorous process for using AI in manufacturing…
…
continue reading