We consider sketching algorithms which first quickly compress data by multiplication with a random sketch matrix, and then apply the sketch to quickly solve an optimization problem, e.g., low rank approximation. In the learning-based sketching paradigm proposed by Indyk et al. [2019], the sketch matrix is found by choosing a random sparse matrix, e.g., the CountSketch, and then updating the values of the non-zero entries by running gradient descent on a training data set. Despite the growing body of work on this paradigm, a noticeable omission is that the locations of the non-zero entries of previous algorithms were fixed, and only their values were learned. In this work we propose the first learning algorithm that also optimizes the locations of the non-zero entries. We show this algorithm gives better accuracy for low rank approximation than previous work, and apply it to other problems such as $k$-means clustering for the first time. We show that our algorithm is provably better in the spiked covariance model and for Zipfian matrices. We also show the importance of the sketch monotonicity property for combining learned sketches. Our empirical results show the importance of optimizing not only the values of the non-zero entries but also their positions.
翻译:我们考虑先通过随机草图矩阵乘以随机草图矩阵快速压缩数据的素描算法,然后应用素描法快速解决优化问题,例如低级近似。在Indyk等人([2019])提出的基于学习的素描范范式中,我们发现素描矩阵的方法是选择随机的稀释矩阵,例如CootSketch,然后通过在培训数据集上运行梯度下降来更新非零条目的值。尽管在这一范式上的工作越来越多,但一个明显的遗漏是,以前算法的非零条目的位置已经固定,只有它们的价值才得到学习。在这个工作中,我们提出了第一个也优化非零条目位置的基于学习算法。我们展示了这种算法比以前的工作更准确的低级近似性,并首次将其应用到诸如 $k$- moint- moints broups 等其它问题。我们发现,我们的算法在急剧变换的调模型和Zipfian矩阵中是更好的。我们还展示了原始的单数单项性单项属性属性的重要性,而不是它们所学的同步输入的重要性。