In high-dimensional classification problems, a commonly used approach is to first project the high-dimensional features into a lower dimensional space, and base the classification on the resulting lower dimensional projections. In this paper, we formulate a latent-variable model with a hidden low-dimensional structure to justify this two-step procedure and to guide which projection to choose. We propose a computationally efficient classifier that takes certain principal components (PCs) of the observed features as projections, with the number of retained PCs selected in a data-driven way. A general theory is established for analyzing such two-step classifiers based on any projections. We derive explicit rates of convergence of the excess risk of the proposed PC-based classifier. The obtained rates are further shown to be optimal up to logarithmic factors in the minimax sense. Our theory allows the lower-dimension to grow with the sample size and is also valid even when the feature dimension (greatly) exceeds the sample size. Extensive simulations corroborate our theoretical findings. The proposed method also performs favorably relative to other existing discriminant methods on three real data examples.
翻译:在高维分类问题中,通常使用的方法是首先将高维特征投射到一个低维空间,然后根据由此得出的低维预测进行分类。在本文中,我们设计了一个隐含低维结构的潜在可变模型,以证明这一两步程序的合理性,并指导投影选择。我们提出了一个计算效率高的分类器,将观测到的特征的某些主要组成部分(PCs)作为预测,以数据驱动的方式选定了所保留的个人计算机的数量。根据任何预测,为分析这种两步分解器建立了一个一般理论。我们得出了拟议的PC基分类器超重风险的明显趋同率。还进一步显示,获得的比率最优于微麦克斯意义上的对数系数。我们的理论允许较低分数随着样本大小的增长,即使特征(大大)超过样本大小,也有效。广泛的模拟也证实了我们的理论结论。在三个真实数据实例中,拟议的方法也比其他现有对齐方法。