Online optimization with memory costs has many real-world applications, where sequential actions are made without knowing the future input. Nonetheless, the memory cost couples the actions over time, adding substantial challenges. Conventionally, this problem has been approached by various expert-designed online algorithms with the goal of achieving bounded worst-case competitive ratios, but the resulting average performance is often unsatisfactory. On the other hand, emerging machine learning (ML) based optimizers can improve the average performance, but suffer from the lack of worst-case performance robustness. In this paper, we propose a novel expert-robustified learning (ERL) approach, achieving {both} good average performance and robustness. More concretely, for robustness, ERL introduces a novel projection operator that robustifies ML actions by utilizing an expert online algorithm; for average performance, ERL trains the ML optimizer based on a recurrent architecture by explicitly considering downstream expert robustification. We prove that, for any $\lambda\geq1$, ERL can achieve $\lambda$-competitive against the expert algorithm and $\lambda\cdot C$-competitive against the optimal offline algorithm (where $C$ is the expert's competitive ratio). Additionally, we extend our analysis to a novel setting of multi-step memory costs. Finally, our analysis is supported by empirical experiments for an energy scheduling application.
翻译:在线优化中的记忆开销具有许多实际应用,其中在不知道未来输入的情况下进行连续行动。尽管如此,记忆成本会将行动耦合在时间上,增加了重大的挑战。传统上,这个问题已经通过各种专家设计的在线算法来解决,目标是实现有界的最坏竞争比,但是由此产生的平均性能通常不令人满意。另一方面,新兴的基于机器学习(ML)的优化器可以提高平均性能,但是缺乏最坏情况下的性能鲁棒性。在本文中,我们提出了一种新颖的专家鲁棒学习(ERL)方法,实现良好的平均性能和鲁棒性。更具体地说,为了提高鲁棒性,ERL引入了一种新颖的投影操作符,通过利用专家在线算法来增强ML行动的鲁棒性;为了提高平均性能,ERL基于具有回归架构的训练ML优化器,明确考虑下游专家的鲁棒化。我们证明,对于任意的$\lambda\geq1$,ERL可以实现对专家算法的$\lambda$-竞争,对最优离线算法的$\lambda\cdot C$-竞争(其中$C$是专家的竞争比)。此外,我们将分析扩展到多步记忆成本的新领域。最后,我们的分析得到了对能源调度应用的经验实验支持。