Encoding the latent posterior of Bayesian Neural Networks for uncertainty quantification by Gianni Franchi (ENSTA)


GdR ISIS Théorie du deep learning - June 28, 2021

Encoding the latent posterior of Bayesian Neural Networks for uncertainty quantification

By Gianni Franchi (ENSTA)

Bayesian Neural Networks (BNNs) have been long considered an ideal, yet unscalable solution for improving the robustness and the predictive uncertainty of deep neural networks. While they could capture more accurately the posterior distribution of the network parameters, most BNN approaches are either limited to small networks or rely on constraining assumptions, e.g., parameter independence. These drawbacks have enabled prominence of simple, but computationally heavy approaches such as Deep Ensembles, whose training and testing costs increase linearly with the number of networks. In this presentation, I will introduce an efficient deep BNN that can manage complex computer vision architectures, e.g., ResNet50 DeepLabV3+, and tasks, e.g., semantic segmentation, with fewer assumptions on the parameters.We achieve this by leveraging variational autoencoders (VAEs) to learn the interaction and the latent distribution of each network layer's parameters.
The approach that I will present, Latent-Posterior BNN (LP-BNN), is compatible with the recent BatchEnsemble method, leading to highly efficient (in terms of computation and memory during both training and testing) ensembles. LP-BNNs attain competitive results across multiple metrics in several challenging benchmarks for image classification, semantic segmentation, and out-of-distribution detection.
Preprint: https://arxiv.org/abs/2012.02818




Social Networks

Check the box to autoplay the video.
Check the box to loop the video.
Check the box to indicate the beginning of playing desired.
 Embed in a web page
 Share the link