We study online convex optimization with switching costs, a practically important but also extremely challenging problem due to the lack of complete offline information. By tapping into the power of machine learning (ML) based optimizers, ML-augmented online algorithms (also referred to as expert calibration in this paper) have been emerging as state of the art, with provable worst-case performance guarantees. Nonetheless, by using the standard practice of training an ML model as a standalone optimizer and plugging it into an ML-augmented algorithm, the average cost performance can be highly unsatisfactory. In order to address the "how to learn" challenge, we propose EC-L2O (expert-calibrated learning to optimize), which trains an ML-based optimizer by explicitly taking into account the downstream expert calibrator. To accomplish this, we propose a new differentiable expert calibrator that generalizes regularized online balanced descent and offers a provably better competitive ratio than pure ML predictions when the prediction error is large. For training, our loss function is a weighted sum of two different losses -- one minimizing the average ML prediction error for better robustness, and the other one minimizing the post-calibration average cost. We also provide theoretical analysis for EC-L2O, highlighting that expert calibration can be even beneficial for the average cost performance and that the high-percentile tail ratio of the cost achieved by EC-L2O to that of the offline optimal oracle (i.e., tail cost ratio) can be bounded. Finally, we test EC-L2O by running simulations for sustainable datacenter demand response. Our results demonstrate that EC-L2O can empirically achieve a lower average cost as well as a lower competitive ratio than the existing baseline algorithms.
翻译:我们用转换成本(一个实际重要但也是极具挑战性的问题)来研究在线调值优化,因为转换成本是一个实际重要但也是极具挑战性的问题,因为缺乏完整的离线信息。通过利用机器学习(ML)基础优化器的力量, ML 推荐的在线算法(本文中也称为专家校准)已经成为最新水平,并有最坏的性能保证。然而,通过使用标准做法将ML模型作为独立优化器加以培训,并将它插入一个纯 ML2 缩略法,平均成本表现可能非常低。为了应对“学习”的挑战,我们建议ECL2O(专家校准学习优化优化优化)将基于ML的优化算法算法作为最新水平,明确考虑下游专家校准师。为了实现这一点,我们建议一个新的可区别专家校准标准,将在线平衡的下降率普遍化,在预测错误很大时比纯 ML2 更低。关于我们损失的计算功能是两种不同损失的加权和加权的加分数 -- 降低平均成本,甚至将O的测试测算算算算出一个成本,我们平均成本,我们的平均成本分析能提供更高的成本。