In a typical optimization problem, the task is to pick one of a number of options with the lowest cost or the highest value. In practice, these cost/value quantities often come through processes such as measurement or machine learning, which are noisy, with quantifiable noise distributions. To take these noise distributions into account, one approach is to assume a prior for the values, use it to build a posterior, and then apply standard stochastic optimization to pick a solution. However, in many practical applications, such prior distributions may not be available. In this paper, we study such scenarios using a regret minimization model. In our model, the task is to pick the highest one out of $n$ values. The values are unknown and chosen by an adversary, but can be observed through noisy channels, where additive noises are stochastically drawn from known distributions. The goal is to minimize the regret of our selection, defined as the expected difference between the highest and the selected value on the worst-case choices of values. We show that the na\"ive algorithm of picking the highest observed value has regret arbitrarily worse than the optimum, even when $n = 2$ and the noises are unbiased in expectation. On the other hand, we propose an algorithm which gives a constant-approximation to the optimal regret for any $n$. Our algorithm is conceptually simple, computationally efficient, and requires only minimal knowledge of the noise distributions.
翻译:在典型的优化问题中,我们的任务是选择成本最低或价值最高的若干选项之一。在实践中,这些成本/价值数量往往通过测量或机器学习等流程,这些流程十分吵闹,且有可量化的噪音分布。考虑到这些噪音分布,一种办法是假设这些值的先验,利用它来构建后继器,然后应用标准的随机优化来选择一个解决方案。然而,在许多实际应用中,可能无法提供这种先前的分布。在本文中,我们使用一个最小遗憾模式来研究此类情景。在我们的模型中,这些成本/价值往往通过从美元值中挑选出最高的一个,这些数值被一个对手所熟知和选择的,但可以通过吵闹的渠道来观察这些数值,其中添加的噪音是从已知分布中随机提取出来的。我们选择的遗憾是,在最坏选择的值中,最高值和选定值之间的预期差异可能不存在。我们发现的最高值的算算算算法比最佳值要差得多。即使当一个对手选择时,最坏的数值为美元,但可以通过一个不固定的运算法,我们最起码的算算算法要求我们最起码的排序。