In this episode, I caught up with Nils Lehmann to learn about Uncertainty Quantification for Neural Networks. The conversation begins with a discussion on Bayesian neural networks and their ability to quantify the uncertainty of their predictions. Unlike regular deterministic neural networks, Bayesian neural networks offer a more principled method for providing predictions with a measure of confidence.
Nils then introduces the Pytorch Lightning UQ Box project on GitHub, a tool that enables experimentation with a variety of Uncertainty Quantification (UQ) techniques for neural networks. Model interpretability is a crucial topic, essential for providing transparency to end users of machine learning models. The video of this conversation is also available on YouTube here
Bio: Nils Lehmann is a PhD Student at the Technical University of Munich (TUM), supervised by Jonathan Bamber and Xiaoxiang Zhu, working on uncertainty quantification for sea-level rise. More broadly his interests lie in Bayesian Deep Learning, uncertainty quantification and generative modelling for Earth Observational data. He is also passionate about open-source software contributions and a maintainer of the Torchgeo package.
din spec 92005 already considers uncertainty in machine learning and the development of an international iso iec standard will start soon. Maybe interesting for you.
See also conformal prediction