Deep neural networks have significantly contributed to the success in predictive accuracy for classification tasks. However, they tend to make over-confident predictions in real-world settings, where domain shifting and out-of-distribution (OOD) examples exist. Most research on uncertainty estimation focuses on computer vision because it provides visual validation on uncertainty quality. However, few have been presented in the natural language process domain. Unlike Bayesian methods that indirectly infer uncertainty through weight uncertainties, current evidential uncertainty-based methods explicitly model the uncertainty of class probabilities through subjective opinions. They further consider inherent uncertainty in data with different root causes, vacuity (i.e., uncertainty due to a lack of evidence) and dissonance (i.e., uncertainty due to conflicting evidence). In our paper, we firstly apply evidential uncertainty in OOD detection for text classification tasks. We propose an inexpensive framework that adopts both auxiliary outliers and pseudo off-manifold samples to train the model with prior knowledge of a certain class, which has high vacuity for OOD samples. Extensive empirical experiments demonstrate that our model based on evidential uncertainty outperforms other counterparts for detecting OOD examples. Our approach can be easily deployed to traditional recurrent neural networks and fine-tuned pre-trained transformers.
翻译:深心神经网络大大促进了分类任务预测准确性的成功,然而,它们往往在现实世界环境中作出过于自信的预测,在现实世界中存在着域转移和分配外(OOD)的例子。关于不确定性估计的大多数研究侧重于计算机视野,因为它提供了对不确定性质量的视觉验证。然而,在自然语言过程域中却很少出现。与通过重量不确定性间接推断不确定性的巴伊西亚方法不同,目前以证据为基础的不确定性为基础的方法通过主观观点明确模拟等级概率的不确定性。它们进一步考虑数据中存在不同根源、空洞(即缺乏证据造成的不确定性)和不协调(即因证据相互冲突造成的不确定性)的内在不确定性。在我们的文件中,我们首先将证据不确定性应用于OOOD检测用于文本分类任务。我们建议一个廉价的框架,既采用辅助外派和假非自成样品,以事先对某类的了解来培训模型,这种模型对OOD样品具有高度的挥发性。广泛的实证实验表明,我们基于显性不确定性的模式可以很容易地超越常规变换网络,我们经常调整的模型可以用来探测。