With the maturity of Artificial Intelligence (AI) technology, Large Scale Visual Geo-Localization (LSVGL) is increasingly important in urban computing, where the task is to accurately and efficiently recognize the geo-location of a given query image. The main challenge of LSVGL faced by many experiments due to the appearance of real-word places may differ in various ways. While perspective deviation almost inevitably exists between training images and query images because of the arbitrary perspective. To cope with this situation, in this paper, we in-depth analyze the limitation of triplet loss which is the most commonly used metric learning loss in state-of-the-art LSVGL framework, and propose a new QUInTuplet Loss (QUITLoss) by embedding all the potential positive samples to the primitive triplet loss. Extensive experiments have been conducted to verify the effectiveness of the proposed approach and the results demonstrate that our new loss can enhance various LSVGL methods.
翻译:随着人工智能(AI)技术的成熟,大型视觉地球定位(LSVGL)在城市计算中越来越重要,在城市计算中,任务是准确和有效地确认某一查询图像的地理位置。LSVGL面临的主要挑战可能因各种方式的不同而不同。虽然由于武断的视角,培训图像和查询图像之间几乎不可避免地存在观点偏差。为了应对这种情况,我们在本文件中深入分析了三重损失的限制,这是最常用的LSVGL框架中最常用的计量学习损失,并通过将所有潜在正面样本嵌入原始三重损失中,提出了一个新的QUITUTUPULos损失(QUITLos)建议。已经进行了广泛的实验,以核实拟议方法的有效性,结果表明我们的新损失可以加强LSVGL的各种方法。