Every autonomous driving dataset has a different configuration of sensors, originating from distinct geographic regions and covering various scenarios. As a result, 3D detectors tend to overfit the datasets they are trained on. This causes a drastic decrease in accuracy when the detectors are trained on one dataset and tested on another. We observe that lidar scan pattern differences form a large component of this reduction in performance. We address this in our approach, SEE-VCN, by designing a novel viewer-centred surface completion network (VCN) to complete the surfaces of objects of interest within an unsupervised domain adaptation framework, SEE. With SEE-VCN, we obtain a unified representation of objects across datasets, allowing the network to focus on learning geometry, rather than overfitting on scan patterns. By adopting a domain-invariant representation, SEE-VCN can be classed as a multi-target domain adaptation approach where no annotations or re-training is required to obtain 3D detections for new scan patterns. Through extensive experiments, we show that our approach outperforms previous domain adaptation methods in multiple domain adaptation settings. Our code and data are available at https://github.com/darrenjkt/SEE-VCN.
翻译:每个自主驱动数据集都有不同的传感器配置,这些传感器来自不同的地理区域,覆盖各种假设情景。因此,3D探测器往往在它们所培训的数据集上过大。当探测器在一个数据集上接受培训并测试另一个数据集时,这导致精确度急剧下降。我们观察到,Lidar扫描模式差异是性能下降的一大部分。我们的方法SEE-VCN解决这一问题的方法是设计一个新的以观察者为中心的表面完成网络,以便在一个不受监督的域适应框架内完成受关注对象的表面。我们借助SEE-VCN,我们获得了一个跨数据集的统一对象表示,使网络能够侧重于学习几何方法,而不是过于适应扫描模式。我们采用域-变量表示法,SEEN-VCN可以归类为多目标域适应方法,无需说明或再培训即可获得3D探测新扫描模式。我们通过广泛的实验,显示我们的方法超越了多个域适应环境中先前的域适应方法。我们的代码和数据可在 https://girn/EDAR/DARV.