Guided depth super-resolution (GDSR) is an essential topic in multi-modal image processing, which reconstructs high-resolution (HR) depth maps from low-resolution ones collected with suboptimal conditions with the help of HR RGB images of the same scene. To solve the challenges in interpreting the working mechanism, extracting cross-modal features and RGB texture over-transferred, we propose a novel Discrete Cosine Transform Network (DCTNet) to alleviate the problems from three aspects. First, the Discrete Cosine Transform (DCT) module reconstructs the multi-channel HR depth features by using DCT to solve the channel-wise optimization problem derived from the image domain. Second, we introduce a semi-coupled feature extraction module that uses shared convolutional kernels to extract common information and private kernels to extract modality-specific information. Third, we employ an edge attention mechanism to highlight the contours informative for guided upsampling. Extensive quantitative and qualitative evaluations demonstrate the effectiveness of our DCTNet, which outperforms previous state-of-the-art methods with a relatively small number of parameters. The code is available at \url{https://github.com/Zhaozixiang1228/GDSR-DCTNet}.
翻译:在多式图像处理中,通过利用同一场景的HR RGB 图像,从以亚最佳条件收集的低分辨率低分辨率图像中重建高分辨率(HR)深度地图;为了解决在解释工作机制、提取跨模式特征和超传输RGB纹理方面的挑战,我们提议建立一个新颖的分解科辛变换网络(DCTNet),从三个方面缓解问题。第一,分解科辛变换模块(DCT),利用DCT,解决源自图像域的频道优化问题,重建多频道HR深度特征。第二,我们采用半混合的特征提取模块,利用共同的革命内核提取共同信息和私人内核提取特定模式信息。第三,我们采用一个尖锐的注意机制,突出用于指导更新的资讯型。广泛的定量和定性评估表明了我们的DCTNet 参数的有效性,该参数比先前的州/州/州/州/州/州/州/州/州代码要小得多。