Intelligence relies on an agent's knowledge of what it does not know. This capability can be assessed based on the quality of joint predictions of labels across multiple inputs. In principle, ensemble-based approaches produce effective joint predictions, but the computational costs of training large ensembles can become prohibitive. We introduce the epinet: an architecture that can supplement any conventional neural network, including large pretrained models, and can be trained with modest incremental computation to estimate uncertainty. With an epinet, conventional neural networks outperform very large ensembles, consisting of hundreds or more particles, with orders of magnitude less computation. The epinet does not fit the traditional framework of Bayesian neural networks. To accommodate development of approaches beyond BNNs, such as the epinet, we introduce the epistemic neural network (ENN) as an interface for models that produce joint predictions.
翻译:这种能力可以基于对多种投入的标签进行联合预测的质量来评估。原则上,基于共同方式的方法可以产生有效的联合预测,但培训大型集合的计算成本会变得令人望而却步。我们引入了肾上腺:一个能够补充任何常规神经网络的建筑,包括大型预先培训模型,并且可以进行适度的渐进计算来估计不确定性。如果一个肾上腺,传统的神经网络能够超越由数百个或更多的粒子组成的非常大的昆虫群,而计算数量则更少。这个顶膜网络并不符合巴伊西亚神经网络的传统框架。为了适应BNNs以外的方法的发展,例如肾上腺网,我们引入缩写神经网络(ENN)作为产生联合预测模型的界面。