Sliced inverse regression (SIR, Li 1991) is a pioneering work and the most recognized method in sufficient dimension reduction. While promising progress has been made in theory and methods of high-dimensional SIR, two remaining challenges are still nagging high-dimensional multivariate applications. First, choosing the number of slices in SIR is a difficult problem, and it depends on the sample size, the distribution of variables, and other practical considerations. Second, the extension of SIR from univariate response to multivariate is not trivial. Targeting at the same dimension reduction subspace as SIR, we propose a new slicing-free method that provides a unified solution to sufficient dimension reduction with high-dimensional covariates and univariate or multivariate response. We achieve this by adopting the recently developed martingale difference divergence matrix (MDDM, Lee & Shao 2018) and penalized eigen-decomposition algorithms. To establish the consistency of our method with a high-dimensional predictor and a multivariate response, we develop a new concentration inequality for sample MDDM around its population counterpart using theories for U-statistics, which may be of independent interest. Simulations and real data analysis demonstrate the favorable finite sample performance of the proposed method.
翻译:切片逆回归(SIR,Li 1991)是充分降维模型中最早和最为知名的方法。尽管在高维SIR理论和方法方面取得了很大进展,但仍有两个挑战困扰着高维多元应用。首先,在SIR中选择切片数是一个棘手的问题,它取决于样本量、变量分布和其他实际考虑因素。其次,SIR从单变量响应扩展到多元响应并不是一件简单的事情。我们针对相同的降维子空间,提出了一种新的无切片方法,可以提供高维协变量和单变量或多元响应的充分维度降低的统一解决方案。我们通过采用最近发展起来的鞅差离散矩阵(MDDM,Lee&Shao 2018)和惩罚特征值分解算法来实现这一点。为了证明我们的方法在高维预测器和多元响应下的一致性,我们利用U-统计量理论开发了一种新的样本MDDM集中不等式,这可能是独立感兴趣的。仿真和实际数据分析展示了所提方法的有利的有限样本表现。