Practical anomaly detection requires applying numerous approaches due to the inherent difficulty of unsupervised learning. Direct comparison between complex or opaque anomaly detection algorithms is intractable; we instead propose a framework for associating the scores of multiple methods. Our aim is to answer the question: how should one measure the similarity between anomaly scores generated by different methods? The scoring crux is the extremes, which identify the most anomalous observations. A pair of algorithms are defined here to be similar if they assign their highest scores to roughly the same small fraction of observations. To formalize this, we propose a measure based on extremal similarity in scoring distributions through a novel upper quadrant modeling approach, and contrast it with tail and other dependence measures. We illustrate our method with simulated and real experiments, applying spectral methods to cluster multiple anomaly detection methods and to contrast our similarity measure with others. We demonstrate that our method is able to detect the clusters of anomaly detection algorithms to achieve an accurate and robust ensemble algorithm.
翻译:实际异常的检测要求应用多种方法,因为缺乏监督的学习本身存在困难。对复杂或不透明异常的检测算法进行直接比较是难以解决的;我们提议了一个框架,将多种方法的分数联系起来。我们的目的是回答问题:如何衡量不同方法产生的异常分数之间的相似性?评分柱是极端的,它确定了最反常的观测结果。如果一对算法将最高分分配给大致相同一小部分的观察结果,这里就定义相似。为了正式确定这一点,我们建议了一种基于通过新型的上象方模型方法在分数分布中极端相似性的措施,并将它与尾部和其他依赖性措施进行比较。我们用模拟的和真实的实验来说明我们的方法,用光谱方法将多重异常的检测方法组合起来,并将我们与其它方法进行比较。我们证明,我们的方法能够探测异常检测算法的集群,以便实现准确和稳健的混合算法。