Environmental perception is an important aspect within the field of autonomous vehicles that provides crucial information about the driving domain, including but not limited to identifying clear driving areas and surrounding obstacles. Semantic segmentation is a widely used perception method for self-driving cars that associates each pixel of an image with a predefined class. In this context, several segmentation models are evaluated regarding accuracy and efficiency. Experimental results on the generated dataset confirm that the segmentation model FasterSeg is fast enough to be used in realtime on lowpower computational (embedded) devices in self-driving cars. A simple method is also introduced to generate synthetic training data for the model. Moreover, the accuracy of the first-person perspective and the bird's eye view perspective are compared. For a $320 \times 256$ input in the first-person perspective, FasterSeg achieves $65.44\,\%$ mean Intersection over Union (mIoU), and for a $320 \times 256$ input from the bird's eye view perspective, FasterSeg achieves $64.08\,\%$ mIoU. Both perspectives achieve a frame rate of $247.11$ Frames per Second (FPS) on the NVIDIA Jetson AGX Xavier. Lastly, the frame rate and the accuracy with respect to the arithmetic 16-bit Floating Point (FP16) and 32-bit Floating Point (FP32) of both perspectives are measured and compared on the target hardware.
翻译:自动车辆领域环境认知是一个重要方面,它提供了驾驶领域的重要信息,包括但不局限于确定明确的驾驶区和周围障碍。语义分割是自驾驶汽车广泛使用的一种认知方法,将图像的每个像素与预定义的类连接起来。在这方面,对若干分解模型进行了准确性和效率方面的评估。生成的数据集的实验结果证实,在自驾驶汽车的低功率计算(封装)设备上,快速可实时用于低功率计算(封装)设备。还采用了一种简单的方法为该模型生成合成培训数据。此外,对第一人视角和鸟眼视图视角的准确性进行了比较。在第一人的角度,320美元乘256美元的投入,PappleSeg达到65.44\\, ⁇ $平均值表示在联盟(米洛U)上,320美元乘以256美元计算,在自驾驶汽车的低功率计算(封装)设备上,快速Segle Seg达到64.08\ ⁇ mIo-lbeximal 角度,而FIA-FAxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx的测量度的测量的硬成本框架框架框架的精确度。