We study the problem of directly optimizing arbitrary non-differentiable task evaluation metrics such as misclassification rate and recall. Our method, named MetricOpt, operates in a black-box setting where the computational details of the target metric are unknown. We achieve this by learning a differentiable value function, which maps compact task-specific model parameters to metric observations. The learned value function is easily pluggable into existing optimizers like SGD and Adam, and is effective for rapidly finetuning a pre-trained model. This leads to consistent improvements since the value function provides effective metric supervision during finetuning, and helps to correct the potential bias of loss-only supervision. MetricOpt achieves state-of-the-art performance on a variety of metrics for (image) classification, image retrieval and object detection. Solid benefits are found over competing methods, which often involve complex loss design or adaptation. MetricOpt also generalizes well to new tasks and model architectures.
翻译:我们研究直接优化任意的、不可区分的任务评价指标的问题,如分类率和召回错误。我们的方法叫做MetricOpt,在黑盒中运行,其中目标指标的计算细节未知。我们通过学习一个不同的价值函数来实现这一点,该功能将具体任务的具体模型参数映射成指标观测。学到的值函数很容易插入SGD和Adam等现有优化器,并且对快速微调预先培训的模式十分有效。这导致不断的改进,因为价值函数在微调过程中提供了有效的指标监督,并有助于纠正只损失监督的潜在偏差。MetricOpt在各种(模拟)分类、图像检索和对象探测的衡量标准上取得了最先进的业绩。在竞争方法上发现了坚实的好处,这些方法往往涉及复杂的损失设计或适应。MetricOpt还把新的任务和模型结构概括化为好。