We study online convex optimization with switching costs, a practically important but also extremely challenging problem due to the lack of complete offline information. By tapping into the power of machine learning (ML) based optimizers, ML-augmented online algorithms (also referred to as expert calibration in this paper) have been emerging as state of the art, with provable worst-case performance guarantees. Nonetheless, by using the standard practice of training an ML model as a standalone optimizer and plugging it into an ML-augmented algorithm, the average cost performance can be even worse than purely using ML predictions. In order to address the "how to learn" challenge, we propose EC-L2O (expert-calibrated learning to optimize), which trains an ML-based optimizer by explicitly taking into account the downstream expert calibrator. To accomplish this, we propose a new differentiable expert calibrator that generalizes regularized online balanced descent and offers a provably better competitive ratio than pure ML predictions when the prediction error is large. For training, our loss function is a weighted sum of two different losses -- one minimizing the average ML prediction error for better robustness, and the other one minimizing the post-calibration average cost. We also provide theoretical analysis for EC-L2O, highlighting that expert calibration can be even beneficial for the average cost performance and that the high-percentile tail ratio of the cost achieved by EC-L2O to that of the offline optimal oracle (i.e., tail cost ratio) can be bounded. Finally, we test EC-L2O by running simulations for sustainable datacenter demand response. Our results demonstrate that EC-L2O can empirically achieve a lower average cost as well as a lower competitive ratio than the existing baseline algorithms.
翻译:我们用转换成本来研究在线 Convex优化,这是一个实际重要但也是极具挑战性的问题,因为缺乏完整的离线信息。通过利用机器学习(ML)基于优化的优化器的力量, ML 启动在线算法(本文中也称为专家校准)已经成为最新水平,并明确考虑到下游专家校准器,我们提出了一个新的差异化专家校准器,该校准器一般化的在线平衡下行率,并将它插入到一个 ML 放大的算法中,平均成本性能可能比纯粹使用 ML 预测更差。为了应对“如何学习”的优化优化优化优化优化优化优化优化优化,我们建议ECL2O(专家校准学习优化优化)的在线算法,通过明确考虑下游专家校准器来培训一个基于ML的基于ML优化的优化的优化。我们提出了一个新的差异化专家校准器,该校准器可以将正常的在线平衡下降,并且提供比纯的 ML 基准值更具有竞争力的比例,当预测错误很大的时候,我们的损失功能甚至比精准了EL2的精确的计算出一个成本,我们的平均测算算法,我们的平均成本分析可以实现一个更精确的计算。