Modern datasets, from areas such as neuroimaging and geostatistics, often come in the form of a random sample of tensor-valued data which can be understood as noisy observations of an underlying smooth multidimensional random function. Many of the traditional techniques from functional data analysis are plagued by the curse of dimensionality and quickly become intractable as the dimension of the domain increases. In this paper, we propose a framework for learning multidimensional continuous representations from a random sample of tensors that is immune to several manifestations of the curse. These representations are defined to be multiplicatively separable and adapted to the data according to an $L^{2}$ optimality criteria, analogous to a multidimensional functional principal components analysis. We show that the resulting estimation problem can be solved efficiently by the tensor decomposition of a carefully defined reduction transformation of the observed data. The incorporation of both regularization and dimensionality reduction is discussed. The advantages of the proposed method over competing methods are demonstrated in a simulation study. We conclude with a real data application in neuroimaging.
翻译:现代数据集来自神经成像学和地理统计学等领域,往往以随机抽样形式提供可被理解为对基本光滑的多功能随机功能进行噪音观测的有价数据值数据。功能数据分析中的许多传统技术都受到维度诅咒的困扰,随着领域范围的扩大,这些传统技术很快就变得难以使用。我们在本文件中提议了一个框架,从随机抽样的可不受多种诅咒表现影响的色粒中学习多层面连续表达方式。这些表达方式的定义是,可重复地相互分离,并适应数据的最佳性标准,类似于多功能主要组成部分分析。我们表明,由此产生的估计问题可以通过审慎界定的减少数据变形的多角度分解而得到有效解决。讨论将规范化和减少维度两者结合的问题。模拟研究显示了拟议方法相对于相互竞争的方法的优势。我们最后在神经成形中以真正的数据应用结束。