TY - GEN
T1 - Extended Variational Inference for Propagating Uncertainty in Convolutional Neural Networks
AU - Dera, Dimah
AU - Rasool, Ghulam
AU - Bouaynaya, Nidhal
N1 - Publisher Copyright:
© 2019 IEEE.
PY - 2019/10
Y1 - 2019/10
N2 - Model confidence or uncertainty is critical in autonomous systems as they directly tie to the safety and trustworthiness of the system. The quantification of uncertainty in the output decisions of deep neural networks (DNNs) is a challenging problem. The Bayesian framework enables the estimation of the predictive uncertainty by introducing probability distributions over the (unknown) network weights; however, the propagation of these high-dimensional distributions through multiple layers and non-linear transformations is mathematically intractable. In this work, we propose an extended variational inference (eVI) framework for convolutional neural network (CNN) based on tensor Normal distributions (TNDs) defined over convolutional kernels. Our proposed eVI framework propagates the first two moments (mean and covariance) of these TNDs through all layers of the CNN. We employ first-order Taylor series linearization to approximate the mean and covariances passing through the non-linear activations. The uncertainty in the output decision is given by the propagated covariance of the predictive distribution. Furthermore, we show, through extensive simulations on the MNIST and CIFAR-10 datasets, that the CNN becomes more robust to Gaussian noise and adversarial attacks.
AB - Model confidence or uncertainty is critical in autonomous systems as they directly tie to the safety and trustworthiness of the system. The quantification of uncertainty in the output decisions of deep neural networks (DNNs) is a challenging problem. The Bayesian framework enables the estimation of the predictive uncertainty by introducing probability distributions over the (unknown) network weights; however, the propagation of these high-dimensional distributions through multiple layers and non-linear transformations is mathematically intractable. In this work, we propose an extended variational inference (eVI) framework for convolutional neural network (CNN) based on tensor Normal distributions (TNDs) defined over convolutional kernels. Our proposed eVI framework propagates the first two moments (mean and covariance) of these TNDs through all layers of the CNN. We employ first-order Taylor series linearization to approximate the mean and covariances passing through the non-linear activations. The uncertainty in the output decision is given by the propagated covariance of the predictive distribution. Furthermore, we show, through extensive simulations on the MNIST and CIFAR-10 datasets, that the CNN becomes more robust to Gaussian noise and adversarial attacks.
UR - http://www.scopus.com/inward/record.url?scp=85077680448&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85077680448&partnerID=8YFLogxK
U2 - 10.1109/MLSP.2019.8918747
DO - 10.1109/MLSP.2019.8918747
M3 - Conference contribution
AN - SCOPUS:85077680448
T3 - IEEE International Workshop on Machine Learning for Signal Processing, MLSP
BT - 2019 IEEE 29th International Workshop on Machine Learning for Signal Processing, MLSP 2019
PB - IEEE Computer Society
T2 - 29th IEEE International Workshop on Machine Learning for Signal Processing, MLSP 2019
Y2 - 13 October 2019 through 16 October 2019
ER -