Training vision-based Urban Autonomous driving models is a challenging problem, which is highly researched in recent times. Training such models is a data-intensive task requiring the storage and processing of vast volumes of (possibly redundant) driving video data. In this paper, we study the problem of developing data-efficient autonomous driving systems. In this context, we study the problem of multi-criteria online video frame subset selection. We study convex optimization-based solutions and show that they are unable to provide solutions with high weightage to the loss of selected video frames. We design a novel convex optimization-based multi-criteria online subset selection algorithm that uses a thresholded concave function of selection variables. We also propose and study a submodular optimization-based algorithm. Extensive experiments using the driving simulator CARLA show that we are able to drop 80% of the frames while succeeding to complete 100% of the episodes w.r.t. the model trained on 100% data, in the most difficult task of taking turns. This results in a training time of less than 30% compared to training on the whole dataset. We also perform detailed experiments on prediction performances of various affordances used by the Conditional Affordance Learning (CAL) model and show that our subset selection improves performance on the crucial affordance "Relative Angle" during turns.
翻译:培训城市自主驱动模型是一个具有挑战性的问题,最近曾对此进行过大量研究。培训这些模型是一个数据密集型任务,需要存储和处理大量(可能多余的)驱动视频数据。在本文中,我们研究了开发数据高效自主驱动系统的问题。在这方面,我们研究了多标准在线视频框架子集选择问题。我们研究了基于配置优化的优化解决方案,并表明它们无法为失去选定视频框架提供高重量的解决方案。我们设计了一个新型的康韦克斯优化型多标准在线子集子选择算法,该算法使用选择变量的临界组合功能。我们还提出并研究一个亚模式优化型算法。使用驱动模拟器CARLA进行的广泛实验表明,在完成100%的视频框架选择过程中,我们能够降低80%的框。我们研究了百分率数据培训的模型,这是最困难的任务。我们设计了一个比整个数据集培训低30%的培训时间。我们还进行了详细测试,还利用了一种以亚值为基础的模型,用于预测“在模型中改进了各种业绩”。