Object detection plays an important role in self-driving cars for security development. However, mobile systems on self-driving cars with limited computation resources lead to difficulties for object detection. To facilitate this, we propose a compiler-aware neural pruning search framework to achieve high-speed inference on autonomous vehicles for 2D and 3D object detection. The framework automatically searches the pruning scheme and rate for each layer to find a best-suited pruning for optimizing detection accuracy and speed performance under compiler optimization. Our experiments demonstrate that for the first time, the proposed method achieves (close-to) real-time, 55ms and 99ms inference times for YOLOv4 based 2D object detection and PointPillars based 3D detection, respectively, on an off-the-shelf mobile phone with minor (or no) accuracy loss.
翻译:物体探测在安全开发自驾驶汽车中起着重要作用。然而,在计算资源有限的自驾驶汽车上,移动系统在探测物体方面造成困难。为了促进这一点,我们提议建立一个编译器神经系统运行搜索框架,以便在自驾驶车辆上实现2D和3D物体探测的高速推断。框架自动搜索每个层的修剪机和速率,以找到最适合在编译器优化下优化探测精确度和速度性能的最佳裁剪。我们的实验表明,拟议的方法首次实现了(近距离)实时、55米和99米的5D天天天天天天线探测和基于3D探测的点PillargPillars5天线探测,分别以小型(或无)准确性损失的现成移动电话为基础。