论文标题
通过基于节点的贝叶斯神经网络来应对协变量转移
Tackling covariate shift with node-based Bayesian neural networks
论文作者
论文摘要
贝叶斯神经网络(BNN)承诺通过提供认识论不确定性的原则概率表示,在协变量转移下的概括改善了。但是,基于重量的BNN通常会在大规模架构和数据集的高计算复杂性上挣扎。最近将基于节点的BNN作为可扩展的替代方案引入,通过将每个隐藏节点乘以潜在的随机变量繁殖,从而引起认知不确定性,同时学习了权重的点刻度。在本文中,我们将这些潜在的噪声变量解释为训练过程中简单和域 - 不合时宜的数据扰动的隐式表示,从而产生了由于输入损坏而导致的协变量转移的BNN。我们观察到,隐性腐败的多样性取决于潜在变量的熵,并提出了一种直接的方法来增加训练期间这些变量的熵。我们评估了分布外图像分类基准的方法,并显示出由于输入扰动而导致的协变量转移下基于节点的BNN的不确定性估计。作为副作用,该方法还提供了针对嘈杂训练标签的鲁棒性。
Bayesian neural networks (BNNs) promise improved generalization under covariate shift by providing principled probabilistic representations of epistemic uncertainty. However, weight-based BNNs often struggle with high computational complexity of large-scale architectures and datasets. Node-based BNNs have recently been introduced as scalable alternatives, which induce epistemic uncertainty by multiplying each hidden node with latent random variables, while learning a point-estimate of the weights. In this paper, we interpret these latent noise variables as implicit representations of simple and domain-agnostic data perturbations during training, producing BNNs that perform well under covariate shift due to input corruptions. We observe that the diversity of the implicit corruptions depends on the entropy of the latent variables, and propose a straightforward approach to increase the entropy of these variables during training. We evaluate the method on out-of-distribution image classification benchmarks, and show improved uncertainty estimation of node-based BNNs under covariate shift due to input perturbations. As a side effect, the method also provides robustness against noisy training labels.