Models trained on datasets with texture bias usually perform poorly on out-of-distribution samples since biased representations are embedded into the model. Recently, various image translation and debiasing methods have attempted to disentangle texture biased representations for downstream tasks, but accurately discarding biased features without altering other relevant information is still challenging. In this paper, we propose a novel framework that leverages image translation to generate additional training images using the content of a source image and the texture of a target image with a different bias property to explicitly mitigate texture bias when training a model on a target task. Our model ensures texture similarity between the target and generated images via a texture co-occurrence loss while preserving content details from source images with a spatial self-similarity loss. Both the generated and original training images are combined to train improved classification or segmentation models robust to inconsistent texture bias. Evaluation on five classification- and two segmentation-datasets with known texture biases demonstrates the utility of our method, and reports significant improvements over recent state-of-the-art methods in all cases.
翻译:由于模型中嵌入了偏差表征,因此在分配外的样本中,受过关于具有纹理偏差的数据集培训的模型通常表现不佳。最近,各种图像翻译和贬低方法试图分解下游任务的纹理偏差代表,但准确丢弃偏差特征,而又不改变其他相关信息,仍然具有挑战性。在本文件中,我们提出一个新的框架,利用图像翻译,利用源图像的内容和具有不同偏差属性的目标图像的纹理来生成更多的培训图像,以在培训目标任务模型时明确减少纹理偏差。我们的模型确保目标之间的纹理相似性,并通过纹理共损耗生成图像生成图像,同时保存来源图像中的内容细节,同时保持空间自相异性损失。生成和原始培训图像结合在一起,以训练更完善的分类或分解模型,使其具有不一致的文字偏差性。对五种分类和两套带有已知纹理偏差的分解数据集进行了评价,显示了我们的方法的效用,并报告了在所有案件中对最新状态方法的重大改进。