Several coded exposure techniques have been proposed for acquiring high frame rate videos at low bandwidth. Most recently, a Coded-2-Bucket camera has been proposed that can acquire two compressed measurements in a single exposure, unlike previously proposed coded exposure techniques, which can acquire only a single measurement. Although two measurements are better than one for an effective video recovery, we are yet unaware of the clear advantage of two measurements, either quantitatively or qualitatively. Here, we propose a unified learning-based framework to make such a qualitative and quantitative comparison between those which capture only a single coded image (Flutter Shutter, Pixel-wise coded exposure) and those that capture two measurements per exposure (C2B). Our learning-based framework consists of a shift-variant convolutional layer followed by a fully convolutional deep neural network. Our proposed unified framework achieves the state of the art reconstructions in all three sensing techniques. Further analysis shows that when most scene points are static, the C2B sensor has a significant advantage over acquiring a single pixel-wise coded measurement. However, when most scene points undergo motion, the C2B sensor has only a marginal benefit over the single pixel-wise coded exposure measurement.
翻译:为了在低带宽下获取高框架率视频,已经提议了几种编码接触技术。最近,提议了一个编码-2-Bucket相机,可以在一次接触中获得两种压缩测量,这与以前提议的编码接触技术不同,而以前提议的编码接触技术只能获得一种单一测量。虽然两种测量在有效视频恢复方面优于一种,但我们还不知道两种测量在数量上或质量上都有明显的优势。在这里,我们提议了一个统一的基于学习的框架,以便在仅捕捉单一编码图像(Flautter Shutter, Pixel Wise-代码接触)和每一次摄取两种测量的图像(C2B)之间进行这种定性和定量的比较。然而,我们基于学习的框架包括一个变化-变异性共振动层,随后有一个完全革命性深层神经网络。我们提议的统一框架在所有三种感测技术中都实现了艺术重建状态。我们进一步的分析表明,当大多数景点为静态时,C2B传感器在获得单一的像分码测量方面有很大的优势。然而,当大多数场点进行运动动时,C2B传感器仅对单一编码接触进行边测测。