We propose novel randomized optimization methods for high-dimensional convex problems based on restrictions of variables to random subspaces. We consider oblivious and data-adaptive subspaces and study their approximation properties via convex duality and Fenchel conjugates. A suitable adaptive subspace can be generated by sampling a correlated random matrix whose second order statistics mirror the input data. We illustrate that the adaptive strategy can significantly outperform the standard oblivious sampling method, which is widely used in the recent literature. We show that the relative error of the randomized approximations can be tightly characterized in terms of the spectrum of the data matrix and Gaussian width of the dual tangent cone at optimum. We develop lower bounds for both optimization and statistical error measures based on concentration of measure and Fano's inequality. We then present the consequences of our theory with data matrices of varying spectral decay profiles. Experimental results show that the proposed approach enables significant speed ups in a wide variety of machine learning and optimization problems including logistic regression, kernel classification with random convolution layers and shallow neural networks with rectified linear units.
翻译:我们根据对随机子空间变量的限制,为高维二次曲线问题提出新的随机优化方法。我们考虑模糊和数据适应的子空间,并通过二次曲线和Fenchel二次曲线研究其近似特性。通过取样,可以产生一个合适的适应性子空间。一个相关随机矩阵,该矩阵的第二顺序统计反映了输入数据。我们说明,适应战略可以大大超过标准、不透明取样方法,在最近的文献中广泛使用。我们表明,随机近似的相对差错可以在数据矩阵和双色锥体高斯宽度的频谱方面进行严格定性。我们根据测量浓度和法诺的不平等性,为优化和统计误差措施制定了较低的界限。然后,我们用不同谱谱衰变图的数据矩阵介绍我们理论的后果。实验结果显示,拟议的方法能够在一系列广泛的机器学习和优化问题中高速上升,包括逻辑回归、内核分解随机相向上层和直线单元的浅神经网络。