@kirk86

DIBS: Diversity inducing Information Bottleneck in Model Ensembles

, , , , , and . (2020)cite arxiv:2003.04514Comment: Samarth Sinha* and Homanga Bharadhwaj* contributed equally to this work. Code will be released at https://github.com/rvl-lab-utoronto/dibs.

Abstract

Although deep learning models have achieved state-of-the-art performance on a number of vision tasks, generalization over high dimensional multi-modal data, and reliable predictive uncertainty estimation are still active areas of research. Bayesian approaches including Bayesian Neural Nets (BNNs) do not scale well to modern computer vision tasks, as they are difficult to train, and have poor generalization under dataset-shift. This motivates the need for effective ensembles which can generalize and give reliable uncertainty estimates. In this paper, we target the problem of generating effective ensembles of neural networks by encouraging diversity in prediction. We explicitly optimize a diversity inducing adversarial loss for learning the stochastic latent variables and thereby obtain diversity in the output predictions necessary for modeling multi-modal data. We evaluate our method on benchmark datasets: MNIST, CIFAR100, TinyImageNet and MIT Places 2, and compared to the most competitive baselines show significant improvements in classification accuracy, under a shift in the data distribution and in out-of-distribution detection. Code will be released in this url https://github.com/rvl-lab-utoronto/dibs

Description

[2003.04514] DIBS: Diversity inducing Information Bottleneck in Model Ensembles

Links and resources

Tags

community

  • @kirk86
  • @dblp
@kirk86's tags highlighted