DEEP VARIATIONAL INFORMATION BOTTLENECK

DEEP VARIATIONAL INFORMATION BOTTLENECK

2017 | Alexander A. Alemi, Ian Fischer, Joshua V. Dillon, Kevin Murphy
This paper introduces a variational approximation to the information bottleneck (IB) method, called Deep Variational Information Bottleneck (Deep VIB). The IB method aims to find a representation Z that is maximally informative about the target Y while being as compressive as possible about the input X. The IB objective is defined as maximizing the mutual information between Z and Y, subject to a constraint on the mutual information between Z and X. However, computing mutual information is computationally challenging, so the authors propose using variational inference to approximate the IB objective. Deep VIB uses a neural network to parameterize the stochastic encoder and leverages the reparameterization trick to enable efficient training with stochastic gradient descent. This allows the method to handle high-dimensional, continuous data, such as images, without being restricted to discrete or Gaussian cases. The method is shown to outperform other regularization techniques in terms of generalization performance and robustness to adversarial attacks. Experiments on the MNIST dataset show that Deep VIB models achieve better performance than deterministic models and are more robust to adversarial examples. The method is also shown to be effective on the ImageNet dataset, where it achieves competitive classification accuracy and improved adversarial robustness compared to deterministic models. The results suggest that Deep VIB can learn representations that are both informative about the target and compressive about the input, leading to better generalization and robustness. The method is also shown to be effective in resisting adversarial attacks, as it requires larger perturbations to fool the classifier. The paper concludes that Deep VIB offers a promising approach to learning robust and generalizable representations in deep learning.This paper introduces a variational approximation to the information bottleneck (IB) method, called Deep Variational Information Bottleneck (Deep VIB). The IB method aims to find a representation Z that is maximally informative about the target Y while being as compressive as possible about the input X. The IB objective is defined as maximizing the mutual information between Z and Y, subject to a constraint on the mutual information between Z and X. However, computing mutual information is computationally challenging, so the authors propose using variational inference to approximate the IB objective. Deep VIB uses a neural network to parameterize the stochastic encoder and leverages the reparameterization trick to enable efficient training with stochastic gradient descent. This allows the method to handle high-dimensional, continuous data, such as images, without being restricted to discrete or Gaussian cases. The method is shown to outperform other regularization techniques in terms of generalization performance and robustness to adversarial attacks. Experiments on the MNIST dataset show that Deep VIB models achieve better performance than deterministic models and are more robust to adversarial examples. The method is also shown to be effective on the ImageNet dataset, where it achieves competitive classification accuracy and improved adversarial robustness compared to deterministic models. The results suggest that Deep VIB can learn representations that are both informative about the target and compressive about the input, leading to better generalization and robustness. The method is also shown to be effective in resisting adversarial attacks, as it requires larger perturbations to fool the classifier. The paper concludes that Deep VIB offers a promising approach to learning robust and generalizable representations in deep learning.
Reach us at info@study.space
[slides and audio] Deep Variational Information Bottleneck