Motivated by the high-frequency data streams continuously generated, real-time learning is becoming increasingly important. These data streams should be processed sequentially with the property that the stream may change over time. In this streaming setting, we propose techniques for minimizing a convex objective through unbiased estimates of its gradients, commonly referred to as stochastic approximation problems. Our methods rely on stochastic approximation algorithms due to their computationally advantage as they only use the previous iterate as a parameter estimate. The reasoning includes iterate averaging that guarantees optimal statistical efficiency under classical conditions. Our non-asymptotic analysis shows accelerated convergence by selecting the learning rate according to the expected data streams. We show that the average estimate converges optimally and robustly to any data stream rate. In addition, noise reduction can be achieved by processing the data in a specific pattern, which is advantageous for large-scale machine learning. These theoretical results are illustrated for various data streams, showing the effectiveness of the proposed algorithms.
翻译:以不断生成的高频数据流为动力,实时学习正在变得日益重要。这些数据流应该与流随时间变化而变化的属性相继处理。在这个流环境中,我们提出通过对其梯度的不偏袒估计(通常称为随机近似问题)最大限度地减少曲线目标的方法。我们的方法依靠随机近似算法,因为它们在计算上只使用先前的迭代作为参数估计。推理包括循环平均率,保证在古典条件下统计效率最佳。我们的非抽取分析显示,通过根据预期的数据流选择学习率加快了趋同速度。我们显示,平均估计值与任何数据流速率的最佳和稳健的汇合。此外,通过以特定模式处理数据可以减少噪音,这有利于大型机器学习。这些理论结果为各种数据流展示,显示了提议的算法的有效性。