Search a title or topic

Over 20 million podcasts, powered by 

Player FM logo
Artwork

Content provided by Alexandre Andorra. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Alexandre Andorra or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.
Player FM - Podcast App
Go offline with the Player FM app!

#138 Quantifying Uncertainty in Bayesian Deep Learning, Live from Imperial College London

1:23:10
 
Share
 

Manage episode 498899677 series 2635823
Content provided by Alexandre Andorra. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Alexandre Andorra or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://podcastplayer.com/legal.

Proudly sponsored by PyMC Labs, the Bayesian Consultancy. Book a call, or get in touch!


Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work!

Visit our Patreon page to unlock exclusive Bayesian swag ;)

Takeaways:

  • Bayesian deep learning is a growing field with many challenges.
  • Current research focuses on applying Bayesian methods to neural networks.
  • Diffusion methods are emerging as a new approach for uncertainty quantification.
  • The integration of machine learning tools into Bayesian models is a key area of research.
  • The complexity of Bayesian neural networks poses significant computational challenges.
  • Future research will focus on improving methods for uncertainty quantification. Generalized Bayesian inference offers a more robust approach to uncertainty.
  • Uncertainty quantification is crucial in fields like medicine and epidemiology.
  • Detecting out-of-distribution examples is essential for model reliability.
  • Exploration-exploitation trade-off is vital in reinforcement learning.
  • Marginal likelihood can be misleading for model selection.
  • The integration of Bayesian methods in LLMs presents unique challenges.

Chapters:

00:00 Introduction to Bayesian Deep Learning

03:12 Panelist Introductions and Backgrounds

10:37 Current Research and Challenges in Bayesian Deep Learning

18:04 Contrasting Approaches: Bayesian vs. Machine Learning

26:09 Tools and Techniques for Bayesian Deep Learning

31:18 Innovative Methods in Uncertainty Quantification

36:23 Generalized Bayesian Inference and Its Implications

41:38 Robust Bayesian Inference and Gaussian Processes

44:24 Software Development in Bayesian Statistics

46:51 Understanding Uncertainty in Language Models

50:03 Hallucinations in Language Models

53:48 Bayesian Neural Networks vs Traditional Neural Networks

58:00 Challenges with Likelihood Assumptions

01:01:22 Practical Applications of Uncertainty Quantification

01:04:33 Meta Decision-Making with Uncertainty

01:06:50 Exploring Bayesian Priors in Neural Networks

01:09:17 Model Complexity and Data Signal

01:12:10 Marginal Likelihood and Model Selection

01:15:03 Implementing Bayesian Methods in LLMs

01:19:21 Out-of-Distribution Detection in LLMs

Thank you to my Patrons for making this episode possible!

Yusuke Saito, Avi Bryant, Ero Carrera, Giuliano Cruz, James Wade, Tradd Salvo, William Benton, James Ahloy, Robin Taylor,, Chad Scherrer, Zwelithini Tunyiswa, Bertrand Wilden, James Thompson, Stephen Oates, Gian Luca Di Tanna, Jack Wells, Matthew Maldonado, Ian Costley, Ally Salim, Larry Gill, Ian Moran, Paul Oreto, Colin Caprani, Colin Carroll, Nathaniel Burbank, Michael Osthege, Rémi Louf, Clive Edelsten, Henri Wallen, Hugo Botha, Vinh Nguyen, Marcin Elantkowski, Adam C. Smith, Will Kurt, Andrew Moskowitz, Hector Munoz, Marco Gorelli, Simon Kessell, Bradley Rode, Patrick Kelley, Rick Anderson, Casper de Bruin, Philippe Labonde, Michael Hankin, Cameron Smith, Tomáš Frýda, Ryan Wesslen, Andreas Netti, Riley King, Yoshiyuki Hamajima, Sven De Maeyer, Michael DeCrescenzo, Fergal M, Mason Yahr, Naoya Kanai, Aubrey Clayton, Jeannine Sue, Omri Har Shemesh, Scott Anthony Robson, Robert Yolken, Or Duek, Pavel Dusek, Paul Cox, Andreas Kröpelin, Raphaël R, Nicolas Rode, Gabriel Stechschulte, Arkady, Kurt TeKolste, Marcus Nölke, Maggi Mackintosh, Grant Pezzolesi, Joshua Meehl, Javier Sabio, Kristian Higgins, Matt Rosinski, Bart Trudeau, Luis Fonseca, Dante Gates, Matt Niccolls, Maksim Kuznecov, Michael Thomas, Luke Gorrie, Cory Kiser, Julio, Edvin Saveljev, Frederick Ayala, Jeffrey Powell, Gal Kampel, Adan Romero, Will Geary, Blake Walters, Jonathan Morgan, Francesco Madrisotti, Ivy Huang, Gary Clarke, Robert Flannery, Rasmus Hindström, Stefan, Corey Abshire, Mike Loncaric, David McCormick, Ronald Legere, Sergio Dolia, Michael Cao, Yiğit Aşık, Suyog Chandramouli and Adam Tilmar Jakobsen.

Dr. Mélodie Monod (Imperial College London, School of Public Health)

Mélodie completed her PhD as part of the EPSRC Modern Statistics and Statistical Machine Learning program at Imperial College London, transitioned to Novartis as Principal Biostatistician, and is currently a Postdoctoral Researcher in Machine Learning at Imperial. Her research includes diffusion models, Bayesian deep learning, non-parametric Bayesian statistics and pandemic modelling. For more details, see her Google Scholar Publications page.

Dr. François-Xavier Briol (University College London, Department of Statistical Science)

F-X is Associate Professor in the Department of Statistical Science at University College London, where he leads the Fundamentals of Statistical Machine Learning research group and is co-director of the UCL ELLIS unit. His research focuses on developing statistical and machine learning methods for the sciences and engineering, with his recent work focusing on Bayesian computation and robustness to model misspecification. For more details, see his Google Scholar page.

Dr. Yingzhen Li (Imperial College London, Department of Computing)

Yingzhen is Associate Professor in Machine Learning at the Department of Computing at Imperial College London, following several years at Microsoft Research Cambridge as senior researcher. Her research focuses on building reliable machine learning systems which can generalise to unseen environments, including topics such as (deep) probabilistic graphical model design, fast and accurate (Bayesian) inference/computation techniques, uncertainty quantification for computation and downstream tasks, and robust and adaptive machine learning systems. For more details, see her Google Scholar Publications page.

Transcript

This is an automatic transcript and may therefore contain errors. Please get in touch if you're willing to correct them.

  continue reading

163 episodes

Artwork
iconShare
 
Manage episode 498899677 series 2635823
Content provided by Alexandre Andorra. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Alexandre Andorra or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://podcastplayer.com/legal.

Proudly sponsored by PyMC Labs, the Bayesian Consultancy. Book a call, or get in touch!


Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work!

Visit our Patreon page to unlock exclusive Bayesian swag ;)

Takeaways:

  • Bayesian deep learning is a growing field with many challenges.
  • Current research focuses on applying Bayesian methods to neural networks.
  • Diffusion methods are emerging as a new approach for uncertainty quantification.
  • The integration of machine learning tools into Bayesian models is a key area of research.
  • The complexity of Bayesian neural networks poses significant computational challenges.
  • Future research will focus on improving methods for uncertainty quantification. Generalized Bayesian inference offers a more robust approach to uncertainty.
  • Uncertainty quantification is crucial in fields like medicine and epidemiology.
  • Detecting out-of-distribution examples is essential for model reliability.
  • Exploration-exploitation trade-off is vital in reinforcement learning.
  • Marginal likelihood can be misleading for model selection.
  • The integration of Bayesian methods in LLMs presents unique challenges.

Chapters:

00:00 Introduction to Bayesian Deep Learning

03:12 Panelist Introductions and Backgrounds

10:37 Current Research and Challenges in Bayesian Deep Learning

18:04 Contrasting Approaches: Bayesian vs. Machine Learning

26:09 Tools and Techniques for Bayesian Deep Learning

31:18 Innovative Methods in Uncertainty Quantification

36:23 Generalized Bayesian Inference and Its Implications

41:38 Robust Bayesian Inference and Gaussian Processes

44:24 Software Development in Bayesian Statistics

46:51 Understanding Uncertainty in Language Models

50:03 Hallucinations in Language Models

53:48 Bayesian Neural Networks vs Traditional Neural Networks

58:00 Challenges with Likelihood Assumptions

01:01:22 Practical Applications of Uncertainty Quantification

01:04:33 Meta Decision-Making with Uncertainty

01:06:50 Exploring Bayesian Priors in Neural Networks

01:09:17 Model Complexity and Data Signal

01:12:10 Marginal Likelihood and Model Selection

01:15:03 Implementing Bayesian Methods in LLMs

01:19:21 Out-of-Distribution Detection in LLMs

Thank you to my Patrons for making this episode possible!

Yusuke Saito, Avi Bryant, Ero Carrera, Giuliano Cruz, James Wade, Tradd Salvo, William Benton, James Ahloy, Robin Taylor,, Chad Scherrer, Zwelithini Tunyiswa, Bertrand Wilden, James Thompson, Stephen Oates, Gian Luca Di Tanna, Jack Wells, Matthew Maldonado, Ian Costley, Ally Salim, Larry Gill, Ian Moran, Paul Oreto, Colin Caprani, Colin Carroll, Nathaniel Burbank, Michael Osthege, Rémi Louf, Clive Edelsten, Henri Wallen, Hugo Botha, Vinh Nguyen, Marcin Elantkowski, Adam C. Smith, Will Kurt, Andrew Moskowitz, Hector Munoz, Marco Gorelli, Simon Kessell, Bradley Rode, Patrick Kelley, Rick Anderson, Casper de Bruin, Philippe Labonde, Michael Hankin, Cameron Smith, Tomáš Frýda, Ryan Wesslen, Andreas Netti, Riley King, Yoshiyuki Hamajima, Sven De Maeyer, Michael DeCrescenzo, Fergal M, Mason Yahr, Naoya Kanai, Aubrey Clayton, Jeannine Sue, Omri Har Shemesh, Scott Anthony Robson, Robert Yolken, Or Duek, Pavel Dusek, Paul Cox, Andreas Kröpelin, Raphaël R, Nicolas Rode, Gabriel Stechschulte, Arkady, Kurt TeKolste, Marcus Nölke, Maggi Mackintosh, Grant Pezzolesi, Joshua Meehl, Javier Sabio, Kristian Higgins, Matt Rosinski, Bart Trudeau, Luis Fonseca, Dante Gates, Matt Niccolls, Maksim Kuznecov, Michael Thomas, Luke Gorrie, Cory Kiser, Julio, Edvin Saveljev, Frederick Ayala, Jeffrey Powell, Gal Kampel, Adan Romero, Will Geary, Blake Walters, Jonathan Morgan, Francesco Madrisotti, Ivy Huang, Gary Clarke, Robert Flannery, Rasmus Hindström, Stefan, Corey Abshire, Mike Loncaric, David McCormick, Ronald Legere, Sergio Dolia, Michael Cao, Yiğit Aşık, Suyog Chandramouli and Adam Tilmar Jakobsen.

Dr. Mélodie Monod (Imperial College London, School of Public Health)

Mélodie completed her PhD as part of the EPSRC Modern Statistics and Statistical Machine Learning program at Imperial College London, transitioned to Novartis as Principal Biostatistician, and is currently a Postdoctoral Researcher in Machine Learning at Imperial. Her research includes diffusion models, Bayesian deep learning, non-parametric Bayesian statistics and pandemic modelling. For more details, see her Google Scholar Publications page.

Dr. François-Xavier Briol (University College London, Department of Statistical Science)

F-X is Associate Professor in the Department of Statistical Science at University College London, where he leads the Fundamentals of Statistical Machine Learning research group and is co-director of the UCL ELLIS unit. His research focuses on developing statistical and machine learning methods for the sciences and engineering, with his recent work focusing on Bayesian computation and robustness to model misspecification. For more details, see his Google Scholar page.

Dr. Yingzhen Li (Imperial College London, Department of Computing)

Yingzhen is Associate Professor in Machine Learning at the Department of Computing at Imperial College London, following several years at Microsoft Research Cambridge as senior researcher. Her research focuses on building reliable machine learning systems which can generalise to unseen environments, including topics such as (deep) probabilistic graphical model design, fast and accurate (Bayesian) inference/computation techniques, uncertainty quantification for computation and downstream tasks, and robust and adaptive machine learning systems. For more details, see her Google Scholar Publications page.

Transcript

This is an automatic transcript and may therefore contain errors. Please get in touch if you're willing to correct them.

  continue reading

163 episodes

All episodes

×
 
Loading …

Welcome to Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Copyright 2025 | Privacy Policy | Terms of Service | | Copyright
Listen to this show while you explore
Play