The merit of projecting data onto linear subspaces is well known from, e.g., dimension reduction. One key aspect of subspace projections, the maximum preservation of variance (principal component analysis), has been thoroughly researched and the effect of random linear projections on measures such as intrinsic dimensionality still is an ongoing effort. In this paper, we investigate the less explored depths of linear projections onto explicit subspaces of varying dimensionality and the expectations of variance that ensue. The result is a new family of bounds for Euclidean distances and inner products. We showcase the quality of these bounds as well as investigate the intimate relation to intrinsic dimensionality estimation.
翻译:将数据投射到线性子空间的优点众所周知,例如,从维度的减少方面。子空间预测的一个关键方面,即最大保存差异(主要组成部分分析)已经进行了彻底研究,随机线性预测对诸如内在维度等测量措施的影响仍然是一项持续的努力。我们在本文件中调查线性预测在清晰的、不同维度的子空间上探索较少的深度,以及随之而来的差异预期。结果为Euclidean距离和内产物的新的界限组合。我们展示了这些界限的质量,并调查了这些界限与内在维度估计之间的密切关系。