Functional principal component analysis (FPCA) is a fundamental tool and has attracted increasing attention in recent decades, while existing methods are restricted to data with a single or finite number of random functions (much smaller than the sample size $n$). In this work, we focus on high-dimensional functional processes where the number of random functions $p$ is comparable to, or even much larger than $n$. Such data are ubiquitous in various fields such as neuroimaging analysis, and cannot be properly modeled by existing methods. We propose a new algorithm, called sparse FPCA, which is able to model principal eigenfunctions effectively under sensible sparsity regimes. While sparsity assumptions are standard in multivariate statistics, they have not been investigated in the complex context where not only is $p$ large, but also each variable itself is an intrinsically infinite-dimensional process. The sparsity structure motivates a thresholding rule that is easy to compute without nonparametric smoothing by exploiting the relationship between univariate orthonormal basis expansions and multivariate Kahunen-Lo\`eve (K-L) representations. We investigate the theoretical properties of the resulting estimators, and illustrate the performance with simulated and real data examples.
翻译:功能性主要组成部分分析(FCCA)是一个基本工具,近几十年来日益引起人们的关注,尽管现有方法仅限于具有单数或有限数量的随机功能的数据(比抽样规模小得多,美元)。在这项工作中,我们侧重于高维功能过程,随机功能的数量与美元相当,甚至甚至比美元大得多。这类数据在神经成像分析等各个领域普遍存在,无法以现有方法进行适当模拟。我们提议一种新的算法,称为稀疏的FPCA,能够在明智的偏移制度下有效地模拟主机能。虽然在多变量统计中,偏移假设是标准的,但它们没有在复杂的环境下被调查,其中随机函数的数量不仅为美元大,而且每个变量本身也是一个内在的无限的维度过程。这种偏狭性结构促使一种门槛规则,很容易在不以现有方法进行适当平衡的情况下进行计算。我们通过利用非ivariate ortal基础扩展和多变量Kahun-Loñéeve(K-Lesti) 和多变式数据模拟模型,我们用模拟和模拟模型来分析结果的特性。