We present a framework for the theoretical analysis of ensembles of low-complexity empirical risk minimisers trained on independent random compressions of high-dimensional data. First we introduce a general distribution-dependent upper-bound on the excess risk, framed in terms of a natural notion of compressibility. This bound is independent of the dimension of the original data representation, and explains the in-built regularisation effect of the compressive approach. We then instantiate this general bound to classification and regression tasks, considering Johnson-Lindenstrauss mappings as the compression scheme. For each of these tasks, our strategy is to develop a tight upper bound on the compressibility function, and by doing so we discover distributional conditions of geometric nature under which the compressive algorithm attains minimax-optimal rates up to at most poly-logarithmic factors. In the case of compressive classification, this is achieved with a mild geometric margin condition along with a flexible moment condition that is significantly more general than the assumption of bounded domain. In the case of regression with strongly convex smooth loss functions we find that compressive regression is capable of exploiting spectral decay with near-optimal guarantees. In addition, a key ingredient for our central upper bound is a high probability uniform upper bound on the integrated deviation of dependent empirical processes, which may be of independent interest.
翻译:我们提出了一个理论分析框架,用于对低复杂度实证风险最低风险的集合进行理论分析,这些风险是受过独立随机压缩高维数据培训的。首先,我们引入了基于自然压缩概念的超风险一般分布依赖的上限,这一界限独立于原始数据代表面的维度,并解释了压缩方法的内在常规效应。然后,我们将这一总体与分类和回归任务连接起来,将约翰逊-伦登斯特拉斯映射作为压缩计划。对于其中每一项任务,我们的战略是发展一个紧凑的压缩功能的上限,通过这样做,我们发现一个基于地理测量性质的分布条件,在这个条件下,压缩算法达到最微缩和最佳的速率,直至大多数多度调节因素。在压缩分类方面,我们实现了这一平衡,同时有一个比约束域假设更普遍的灵活时间条件。对于这些任务来说,我们的战略是发展一个紧凑平稳损失功能的回归,而我们通过这样做,我们发现了一个地理测量特性的分布环境条件,在这个条件下,压缩算出一个最接近平稳的缩缩缩缩度,我们发现一个核心的缩缩缩缩缩成成比例的上,可以利用一个高的缩缩成等的缩缩成等的缩成等的缩成成等的缩成等。