Anomaly detection on attributed graphs is a crucial topic for its practical application. Existing methods suffer from semantic mixture and imbalance issue because they mainly focus on anomaly discrimination, ignoring representation learning. It conflicts with the assortativity assumption that anomalous nodes commonly connect with normal nodes directly. Additionally, there are far fewer anomalous nodes than normal nodes, indicating a long-tailed data distribution. To address these challenges, a unique algorithm,Decoupled Self-supervised Learning forAnomalyDetection (DSLAD), is proposed in this paper. DSLAD is a self-supervised method with anomaly discrimination and representation learning decoupled for anomaly detection. DSLAD employs bilinear pooling and masked autoencoder as the anomaly discriminators. By decoupling anomaly discrimination and representation learning, a balanced feature space is constructed, in which nodes are more semantically discriminative, as well as imbalance issue can be resolved. Experiments conducted on various six benchmark datasets reveal the effectiveness of DSLAD.
翻译:在带属性图中进行异常检测是一个具有实用性的重要问题。现有的方法存在语义混合和不平衡问题,因为它们主要侧重于异常判别,忽略了表征学习,这与异常节点通常直接连接正常节点的同伴偏好不符。此外,异常节点要远远少于正常节点,这表明存在长尾数据分布。为了解决这些挑战,本文提出了一种独特的算法,名为Decoupled Self-supervised Learning forAnomalyDetection (DSLAD),它是一种将异常判别和表征学习分离的自监督方法,用于异常检测。DSLAD采用双线性汇聚和掩蔽自编码器作为异常鉴别器。通过分离异常判别和表征学习,构建了一个平衡的特征空间,其中节点更具有语义鉴别力,并且不平衡问题也得到了解决。在各种六个基准数据集上进行的实验表明了DSLAD的有效性。