Detecting test data deviating from training data is a central problem for safe and robust machine learning. Likelihoods learned by a generative model, e.g., a normalizing flow via standard log-likelihood training, perform poorly as an anomaly score. We propose to use an unlabelled auxiliary dataset and a probabilistic outlier score for anomaly detection. We use a self-supervised feature extractor trained on the auxiliary dataset and train a normalizing flow on the extracted features by maximizing the likelihood on in-distribution data and minimizing the likelihood on the auxiliary dataset. We show that this is equivalent to learning the normalized positive difference between the in-distribution and the auxiliary feature density. We conduct experiments on benchmark datasets and show a robust improvement compared to likelihood, likelihood ratio methods and state-of-the-art anomaly detection methods.
翻译:检测偏离培训数据的测试数据是安全和稳健的机器学习的中心问题。通过基因模型(例如,通过标准日志相似性培训实现正常流动)所学到的可能性,作为异常分差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差的差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差的差差差差差差差差差差差差差差差差差差差差差差差差差差差差差的差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差的差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差差