Dimensionality reduction is a popular preprocessing and a widely used tool in data mining. Transparency, which is usually achieved by means of explanations, is nowadays a widely accepted and crucial requirement of machine learning based systems like classifiers and recommender systems. However, transparency of dimensionality reduction and other data mining tools have not been considered in much depth yet, still it is crucial to understand their behavior -- in particular practitioners might want to understand why a specific sample got mapped to a specific location. In order to (locally) understand the behavior of a given dimensionality reduction method, we introduce the abstract concept of contrasting explanations for dimensionality reduction, and apply a realization of this concept to the specific application of explaining two dimensional data visualization.
翻译:减少维度是一个广受欢迎的预处理工具,也是数据开采中广泛使用的工具。透明度通常是通过解释实现的。透明度现在已成为基于机器学习的系统,例如分类和建议系统的广泛接受和关键要求。然而,尚未深入考虑减少维度和其他数据挖掘工具的透明度,但了解其行为仍然至关重要,特别是从业人员可能想了解为什么具体样本被映射到特定地点。为了(当地)了解特定维度减少方法的行为,我们引入了对减少维度进行对比解释的抽象概念,并将这一概念的实现应用于解释两个维度数据可视化的具体应用。