Developing interpretable machine learning models has become an increasingly important issue. One way in which data scientists have been able to develop interpretable models has been to use dimension reduction techniques. In this paper, we examine several dimension reduction techniques including two recent approaches developed in the network psychometrics literature called exploratory graph analysis (EGA) and unique variable analysis (UVA). We compared EGA and UVA with two other dimension reduction techniques common in the machine learning literature (principal component analysis and independent component analysis) as well as no reduction to the variables real data. We show that EGA and UVA perform as well as the other reduction techniques or no reduction. Consistent with previous literature, we show that dimension reduction can decrease, increase, or provide the same accuracy as no reduction of variables. Our tentative results find that dimension reduction tends to lead to better performance when used for classification tasks.
翻译:随着发展,开发可解释的机器学习模型变得越来越重要。数据科学家们采用降维技术作为开发可解释模型的一种方法。本文研究了几种降维方法,包括网络心理测量学中最近发展的探索性图分析和独特变量分析。我们将EGA和UVA与机器学习文献中常用的两种降维方法(主成分分析和独立成分分析)以及不减少变量应用于实际数据的情况进行比较。我们表明,EGA和UVA表现同其他降维技术或不减少变量时一样好。与先前的文献相一致,我们表明,降维可以降低,增加或提供与不减少变量相同的准确率。我们初步结果发现,降维技术在分类任务中使用时 tends to lead to better performance。