The pedestrian crossing intention prediction problem is to estimate whether or not the target pedestrian will cross the street. State-of-the-art techniques heavily depend on visual data acquired through the front camera of the ego-vehicle to make a prediction of the pedestrian's crossing intention. Hence, the efficiency of current methodologies tends to decrease notably in situations where visual input is imprecise, for instance, when the distance between the pedestrian and ego-vehicle is considerable or the illumination levels are inadequate. To address the limitation, in this paper, we present the design, implementation, and evaluation of the first-of-its-kind pedestrian crossing intention prediction model based on integration of motion sensor data gathered through the smartwatch (or smartphone) of the pedestrian. We propose an innovative machine learning framework that effectively integrates motion sensor data with visual input to enhance the predictive accuracy significantly, particularly in scenarios where visual data may be unreliable. Moreover, we perform an extensive data collection process and introduce the first pedestrian intention prediction dataset that features synchronized motion sensor data. The dataset comprises 255 video clips that encompass diverse distances and lighting conditions. We trained our model using the widely-used JAAD and our own datasets and compare the performance with a state-of-the-art model. The results demonstrate that our model outperforms the current state-of-the-art method, particularly in cases where the distance between the pedestrian and the observer is considerable (more than 70 meters) and the lighting conditions are inadequate.
翻译:行人过境意图预测问题在于估计目标行人是否会过马路。最先进的技术在很大程度上取决于通过自我车前摄像头(或智能手机)获得的视觉数据,以预测行人过境的意图。因此,在视觉输入不准确的情况下,特别是在视觉数据可能不可靠的情况下,当前方法的效率往往明显下降。此外,我们开展了广泛的数据收集工作,并引入了首个行人意图预测数据集,该数据集包括255个包含不同距离和照明条件的视频剪辑。我们用广泛使用的JAAAAD和我们自己的路程模型(特别是路程模型和路程模型)对模型进行了培训,在模型之间展示了我们现有的路程和路程模型之间的大量情况。</s>