Object detection networks have reached an impressive performance level, yet a lack of suitable data in specific applications often limits it in practice. Typically, additional data sources are utilized to support the training task. In these, however, domain gaps between different data sources pose a challenge in deep learning. GAN-based image-to-image style-transfer is commonly applied to shrink the domain gap, but is unstable and decoupled from the object detection task. We propose AWADA, an Attention-Weighted Adversarial Domain Adaptation framework for creating a feedback loop between style-transformation and detection task. By constructing foreground object attention maps from object detector proposals, we focus the transformation on foreground object regions and stabilize style-transfer training. In extensive experiments and ablation studies, we show that AWADA reaches state-of-the-art unsupervised domain adaptation object detection performance in the commonly used benchmarks for tasks such as synthetic-to-real, adverse weather and cross-camera adaptation.
翻译:物体探测网络已经达到令人印象深刻的性能水平,但具体应用中缺乏适当数据往往在实践中限制了数据,通常利用更多数据来源来支持培训任务,但是,在这些方面,不同数据来源之间的领域差距对深层学习构成挑战。基于GAN的图像到图像样式传输通常用于缩小域间差距,但不稳定,与物体探测任务脱钩。我们提议AWADA,即“注意的对地域适应框架”,用于在样式转换和探测任务之间建立反馈循环。我们从物体探测器的建议中绘制地表层物体注意地图,我们把工作重点放在浅地物体区域,并稳定样式转移培训。在广泛的实验和模拟研究中,我们显示AWAWADA在合成到现实、恶劣天气和跨镜头适应等任务通常使用的基准中达到最先进的、不受监督的域内适应物体探测性能。