Modeling cameras for the simulation of autonomous robotics is critical for generating synthetic images with appropriate realism to effectively evaluate a perception algorithm in simulation. In many cases though, simulated images are produced by traditional rendering techniques that exclude or superficially handle processing steps and aspects encountered in the actual camera pipeline. The purpose of this contribution is to quantify the degree to which the exclusion from the camera model of various image generation steps or aspects affect the sim-to-real gap in robotics. We investigate what happens if one ignores aspects tied to processes from within the physical camera, e.g., lens distortion, noise, and signal processing; scene effects, e.g., lighting and reflection; and rendering quality. The results of the study demonstrate, quantitatively, that large-scale changes to color, scene, and location have far greater impact than model aspects concerned with local, feature-level artifacts. Moreover, we show that these scene-level aspects can stem from lens distortion and signal processing, particularly when considering white-balance and auto-exposure modeling.
翻译:模拟自动机器人模拟的模拟相机对于制作具有适当现实性的合成图像以有效评价模拟中感知算法至关重要,但在许多情况下,模拟图像是通过排除或表面处理实际照相机管道中遇到的处理步骤和方面的传统制作技术制作的,这一贡献的目的是量化从照相机模型中排除各种图像生成步骤或方面对机器人的模拟到现实差距的影响程度。我们调查如果忽视与物理照相机中过程相关联的方面,例如镜头扭曲、噪音和信号处理;现场效应,例如照明和反射;以及制作质量。研究结果从数量上表明,对颜色、场景和位置的大规模变化的影响远远超过与本地、地貌级文物有关的模型方面。此外,我们表明,这些场面层面的方面可以产生于镜头扭曲和信号处理,特别是在考虑白平衡和自动曝光模型时。