Invariance to diverse types of image corruption, such as noise, blurring, or colour shifts, is essential to establish robust models in computer vision. Data augmentation has been the major approach in improving the robustness against common corruptions. However, the samples produced by popular augmentation strategies deviate significantly from the underlying data manifold. As a result, performance is skewed toward certain types of corruption. To address this issue, we propose a multi-source vicinal transfer augmentation (VITA) method for generating diverse on-manifold samples. The proposed VITA consists of two complementary parts: tangent transfer and integration of multi-source vicinal samples. The tangent transfer creates initial augmented samples for improving corruption robustness. The integration employs a generative model to characterize the underlying manifold built by vicinal samples, facilitating the generation of on-manifold samples. Our proposed VITA significantly outperforms the current state-of-the-art augmentation methods, demonstrated in extensive experiments on corruption benchmarks.
翻译:为了在计算机视野中建立稳健的模型,必须采用多种类型的图像腐败,如噪音、模糊或颜色变化等。数据增强是提高常见腐败稳健性的主要方法。但是,大众增长战略产生的样本与基本数据大不相同。因此,绩效偏向于某些类型的腐败。为解决这一问题,我们提议了一种多源侧点转移增强(VITA)方法,用于生成多种多样的单点样本。拟议的VITA方法由两个互补部分组成:多源昆虫样本的切切转移和集成。相干转移创造了初步增强样本,以提高腐败稳健性。整合采用了一种基因化模型来描述维点样品所构建的基本多维点,便利生成自制样本。我们拟议的VITA方法大大超越了当前最先进的增强方法,在腐败基准的广泛实验中证明了这一点。