Anomaly detection (AD) has attracted considerable attention in both academia and industry. Due to the lack of anomalous data in many practical cases, AD is usually solved by first modeling the normal data pattern and then determining if data fit this model. Generative models (GMs) seem a natural tool to achieve this purpose, which learn the normal data distribution and estimate it using a probability density function (PDF). However, some works have observed the ideal performance of such GM-based AD methods. In this paper, we propose a new perspective on the ideal performance of GM-based AD methods. We state that in these methods, the implicit assumption that connects GMs'results to AD's goal is usually implausible due to normal data's multi-peaked distribution characteristic, which is quite common in practical cases. We first qualitatively formulate this perspective, and then focus on the Gaussian mixture model (GMM) to intuitively illustrate the perspective, which is a typical GM and has the natural property to approximate multi-peaked distributions. Based on the proposed perspective, in order to bypass the implicit assumption in the GMM-based AD method, we suggest integrating the Discriminative idea to orient GMM to AD tasks (DiGMM). With DiGMM, we establish a connection of generative and discriminative models, which are two key paradigms for AD and are usually treated separately before. This connection provides a possible direction for future works to jointly consider the two paradigms and incorporate their complementary characteristics for AD.
翻译:异常检测(AD)在学术界和行业都引起了相当大的注意。由于许多实际案例缺乏异常数据,AD通常通过首先建模正常数据模式,然后确定数据是否适合这一模式来解决。生成模型(GMs)似乎是实现这一目的的自然工具,它们学习正常数据分布并使用概率密度函数(PDF)进行估算。然而,有些工作观察了这种基于全球机制的AD方法的理想性能。在本文件中,我们提出了关于基于全球机制的AD方法的理想性能的新观点。我们指出,在这些方法中,将全球机制的结果与AD目标相配合的隐含假设通常难以令人信服,因为正常数据多位分布特征在实际案例中非常常见。我们首先从质量上提出这一视角,然后侧重于高斯混合混合物模型(GMMM),以便直截地说明这种观点,这是一种典型的全球机制模式,其自然属性可以接近多位模式的分布。我们从拟议的观点出发,通常要绕过GMDM的隐含性假设,或者在GMM-D-D-AD上将双个核心任务与DM-ADM-AD-AD方法联系起来。我们建议了一种共同进行。