Developing interpretable machine learning models has become an increasingly important issue. One way in which data scientists have been able to develop interpretable models has been to use dimension reduction techniques. In this paper, we examine several dimension reduction techniques including two recent approaches developed in the network psychometrics literature called exploratory graph analysis (EGA) and unique variable analysis (UVA). We compared EGA and UVA with two other dimension reduction techniques common in the machine learning literature (principal component analysis and independent component analysis) as well as no reduction to the variables real data. We show that EGA and UVA perform as well as the other reduction techniques or no reduction. Consistent with previous literature, we show that dimension reduction can decrease, increase, or provide the same accuracy as no reduction of variables. Our tentative results find that dimension reduction tends to lead to better performance when used for classification tasks.
翻译:开发可解释的机器学习模型已成为一个日益重要的问题。数据科学家能够开发可解释模型的一种方式是使用减少维度的技术。在本文中,我们研究了几个减少维度的技术,包括网络心理计量学文献中最近开发的两种方法,即探索图分析(EGA)和独特变量分析(UVA ) 。我们把EGA和UVA与机器学习文献中常见的另外两种减少维度技术(主要组成部分分析和独立组成部分分析)进行了比较,对变量真实数据没有减少。我们表明EGA和UVA的表现以及其他减少技术或没有减少。与以前的文献一样,我们表明减少维度可以减少、增加或提供与不减少变量相同的精确度。我们的初步结果发现,在用于分类任务时,减少维度往往导致更好的业绩。