Constrained least squares problems arise in many applications. Their memory and computation costs are expensive in practice involving high-dimensional input data. We employ the so-called "sketching" strategy to project the least squares problem onto a space of a much lower "sketching dimension" via a random sketching matrix. The key idea of sketching is to reduce the dimension of the problem as much as possible while maintaining the approximation accuracy. Tensor structure is often present in the data matrices of least squares, including linearized inverse problems and tensor decompositions. In this work, we utilize a general class of row-wise tensorized sub-Gaussian matrices as sketching matrices in constrained optimizations for the sketching design's compatibility with tensor structures. We provide theoretical guarantees on the sketching dimension in terms of error criterion and probability failure rate. In the context of unconstrained linear regressions, we obtain an optimal estimate for the sketching dimension. For optimization problems with general constraint sets, we show that the sketching dimension depends on a statistical complexity that characterizes the geometry of the underlying problems. Our theories are demonstrated in a few concrete examples, including unconstrained linear regression and sparse recovery problems.
翻译:在许多应用中, 最小的折叠方块问题会出现于许多应用程序中。 它们的内存和计算成本在实践上涉及高维输入数据, 费用昂贵。 我们使用所谓的“ 预置” 战略, 通过随机草图矩阵将最小方块问题投射到一个低得多的“ 预置维度” 空间中。 草图的关键想法是尽可能减少问题的维度, 并同时保持近似准确性。 最小方块的数据矩阵中往往存在感应结构, 包括线性反向问题和高温分解问题。 在这项工作中, 我们使用一整类行的单向偏振荡亚Gaussian 矩阵作为草图矩阵, 用于将最小方块问题投射到草图设计与强力结构兼容性较弱的空间中。 我们从理论上保证草图维度的维度在错误标准和概率失常率率率率率方面。 在未受限制的线性回归中, 我们获得一个最优估度的图度估计值。 关于优化问题集问题, 我们表明, 草图维度的维度取决于统计复杂性, 其特征是少数的回归, 。