We consider the problem of anomaly detection with a small set of partially labeled anomaly examples and a large-scale unlabeled dataset. This is a common scenario in many important applications. Existing related methods either exclusively fit the limited anomaly examples that typically do not span the entire set of anomalies, or proceed with unsupervised learning from the unlabeled data. We propose here instead a deep reinforcement learning-based approach that enables an end-to-end optimization of the detection of both labeled and unlabeled anomalies. This approach learns the known abnormality by automatically interacting with an anomaly-biased simulation environment, while continuously extending the learned abnormality to novel classes of anomaly (i.e., unknown anomalies) by actively exploring possible anomalies in the unlabeled data. This is achieved by jointly optimizing the exploitation of the small labeled anomaly data and the exploration of the rare unlabeled anomalies. Extensive experiments on 48 real-world datasets show that our model significantly outperforms five state-of-the-art competing methods.
翻译:我们考虑异常点检测问题,使用少量部分标签的异常点示例和大规模无标签数据集。这是许多重要应用中常见的情景。现有的相关方法要么完全适合通常不跨越整个异常点的有限异常点实例,要么在没有监督的情况下从未标的数据中学习。我们在这里建议采用深层强化学习法,使标签的异常点和未标的异常点的检测得到端到端的优化。这个方法通过与异常点模拟环境自动互动来了解已知的异常点,同时通过积极探索未标数据中可能出现的异常点,不断将所学的异常点扩大到新型的异常点类别(即未知异常点)。这是通过联合优化利用小标签的异常点数据和探索稀有的未标的异常点来实现的。48个真实世界数据集的广泛实验显示,我们的模型大大超越了5种最先进的竞争方法。