A recent trend in machine learning has been to enrich learned models with the ability to explain their own predictions. The emerging field of Explainable AI (XAI) has so far mainly focused on supervised learning, in particular, deep neural network classifiers. In many practical problems however, label information is not given and the goal is instead to discover the underlying structure of the data, for example, its clusters. While powerful methods exist for extracting the cluster structure in data, they typically do not answer the question why a certain data point has been assigned to a given cluster. We propose a new framework that can, for the first time, explain cluster assignments in terms of input features in an efficient and reliable manner. It is based on the novel insight that clustering models can be rewritten as neural networks - or 'neuralized'. Cluster predictions of the obtained networks can then be quickly and accurately attributed to the input features. Several showcases demonstrate the ability of our method to assess the quality of learned clusters and to extract novel insights from the analyzed data and representations.
翻译:机器学习的最近趋势是丰富学习的模型,使其有能力解释自己的预测。新兴的可解释的AI(XAI)领域迄今为止主要侧重于监督的学习,特别是深神经网络分类。然而,在许多实际问题中,没有提供标签信息,而目标是发现数据的基本结构,例如数据组。虽然在数据中提取组群结构有强大的方法,但它们通常不回答为什么将某个数据点指定给某个组的问题。我们提出了一个新的框架,首次能够以高效和可靠的方式以输入特征的方式解释集群任务。它基于新颖的洞察,即集群模型可以被改写成神经网络,或者“内化 ” 。然后获得的网络群集预测可以快速和准确地归因于输入特征。一些演示展示了我们评估所学到的集群质量和从分析的数据和表述中提取新洞察到的洞察力的方法的能力。