High-quality structured data with rich annotations are critical components in intelligent vehicle systems dealing with road scenes. However, data curation and annotation require intensive investments and yield low-diversity scenarios. The recently growing interest in synthetic data raises questions about the scope of improvement in such systems and the amount of manual work still required to produce high volumes and variations of simulated data. This work proposes a synthetic data generation pipeline that utilizes existing datasets, like nuScenes, to address the difficulties and domain-gaps present in simulated datasets. We show that using annotations and visual cues from existing datasets, we can facilitate automated multi-modal data generation, mimicking real scene properties with high-fidelity, along with mechanisms to diversify samples in a physically meaningful way. We demonstrate improvements in mIoU metrics by presenting qualitative and quantitative experiments with real and synthetic data for semantic segmentation on the Cityscapes and KITTI-STEP datasets. All relevant code and data is released on github (https://github.com/shubham1810/trove_toolkit).
翻译:具有丰富说明的高质量结构化数据是处理道路场景的智能车辆系统的关键组成部分,然而,数据整理和批注需要大量投资和低多样性情景。最近对合成数据的兴趣日益增长,令人质疑这些系统改进的范围以及生成大量和变异模拟数据仍需要多少人工工作。这项工作提议合成数据生成管道,利用模拟数据集中的现有数据集,如nuScenses,解决困难和域隔。我们显示,使用现有数据集的注释和直观提示,我们可以促进自动化多式数据生成,模拟高纤维真实场景特性,同时建立机制,使样本多样化,具有实际意义。我们通过提供定性和定量实验,用真实和合成数据对市景和KITTI-STEP数据集进行语系分解,展示了MIOU测量的改进。所有相关代码和数据都在Github上发布(https://github.com/shubham10_toolkit)。