Object recognition is a key enabler across industry and defense. As technology changes, algorithms must keep pace with new requirements and data. New modalities and higher resolution sensors should allow for increased algorithm robustness. Unfortunately, algorithms trained on existing labeled datasets do not directly generalize to new data because the data distributions do not match. Transfer learning (TL) or domain adaptation (DA) methods have established the groundwork for transferring knowledge from existing labeled source data to new unlabeled target datasets. However, current DA approaches assume similar source and target feature spaces and suffer in the case of massive domain shifts or changes in the feature space. Existing methods assume the data are either the same modality, or can be aligned to a common feature space. Therefore, most methods are not designed to support a fundamental domain change such as visual to auditory data. We propose a novel deep learning framework that overcomes this limitation by learning separate feature extractions for each domain while minimizing the distance between the domains in a latent lower-dimensional space. The alignment is achieved by considering the data manifold along with an adversarial training procedure. We demonstrate the effectiveness of the approach versus traditional methods with several ablation experiments on synthetic, measured, and satellite image datasets. We also provide practical guidelines for training the network while overcoming vanishing gradients which inhibit learning in some adversarial training settings.
翻译:作为技术变革,算法必须与新的要求和数据同步。新的模式和更高分辨率传感器应允许提高算法的稳健性。 不幸的是,在现有的标签数据集上培训的算法并不直接概括于新数据,因为数据分布不匹配。 传输学习(TL)或域适应(DA)方法为将现有标签源数据的知识从现有的标签源数据传输到新的未标签目标数据集奠定了基础。 但是,作为技术变革,目前的DA方法具有类似的源和目标特征空间,在大规模域变换或功能空间变化的情况下会受到影响。现有方法假定数据要么是同一模式,要么可以与共同的特性空间相匹配。因此,大多数方法都不是为了支持基本领域的变化,如视觉到听力数据。我们提出了一个新的深层次学习框架,通过为每个域学习单独的特性提取而将区域之间的距离最小化。通过将数据与对称培训程序同时考虑数据多重和对称空间空间变化。我们展示了方法相对于传统方法的有效性,而传统方法则以一些实用的对称性模型的学习模式来抑制对等的学习。