The breakthrough of contrastive learning (CL) has fueled the recent success of self-supervised learning (SSL) in high-level vision tasks on RGB images. However, CL is still ill-defined for low-level vision tasks, such as joint demosaicking and denoising (JDD), in the RAW domain. To bridge this methodological gap, we present a novel CL approach on RAW images, residual contrastive learning (RCL), which aims to learn meaningful representations for JDD. Our work is built on the assumption that noise contained in each RAW image is signal-dependent, thus two crops from the same RAW image should have more similar noise distribution than two crops from different RAW images. We use residuals as a discriminative feature and the earth mover's distance to measure the distribution divergence for the contrastive loss. To evaluate the proposed CL strategy, we simulate a series of unsupervised JDD experiments with large-scale data corrupted by synthetic signal-dependent noise, where we set a new benchmark for unsupervised JDD tasks with unknown (random) noise variance. Our empirical study not only validates that CL can be applied on distributions (c.f. features), but also exposes the lack of robustness of previous non-ML and SSL JDD methods when the statistics of the noise are unknown, thus providing some further insight into signal-dependent noise problems.
翻译:对比学习(CL)的突破为最近在RGB图像的高级视觉任务中自我监督学习(SSL)的成功提供了动力。然而,在RAW域,CL对于低层次的视觉任务,例如联合演示和分解(JDD),仍然定义不妥。为了缩小这一方法上的差距,我们提出了关于RAW图像的新CL方法,残余对比学习(RCL),目的是了解JDD的有意义表达方式。我们的工作建立在以下假设之上:每张RAW图像中包含的噪音依赖信号,因此同一RAW图像中的两个作物的噪音分布应该比不同RAW图像中的两个作物更为相似。我们用残余物作为歧视特征和地球移动器距离来衡量差异损失的分布差异。为了评估拟议的CL战略,我们模拟了一系列不统一JDD实验,大规模数据因合成信号依赖噪音而腐蚀。我们为未校准的JDDD任务设定了新的基准,而未加校准的JDD(random)噪音差异,我们的经验性研究不仅验证了SL先前的准确的分布方法,而且只是证明CL在SDR的不甚清晰的分布上缺乏。