Lane detection is one of the most important tasks in self-driving. Due to various complex scenarios (e.g., severe occlusion, ambiguous lanes, etc.) and the sparse supervisory signals inherent in lane annotations, lane detection task is still challenging. Thus, it is difficult for the ordinary convolutional neural network (CNN) to train in general scenes to catch subtle lane feature from the raw image. In this paper, we present a novel module named REcurrent Feature-Shift Aggregator (RESA) to enrich lane feature after preliminary feature extraction with an ordinary CNN. RESA takes advantage of strong shape priors of lanes and captures spatial relationships of pixels across rows and columns. It shifts sliced feature map recurrently in vertical and horizontal directions and enables each pixel to gather global information. RESA can conjecture lanes accurately in challenging scenarios with weak appearance clues by aggregating sliced feature map. Moreover, we propose a Bilateral Up-Sampling Decoder that combines coarse-grained and fine-detailed features in the up-sampling stage. It can recover the low-resolution feature map into pixel-wise prediction meticulously. Our method achieves state-of-the-art results on two popular lane detection benchmarks (CULane and Tusimple). Code has been made available at: https://github.com/ZJULearning/resa.
翻译:由于各种复杂的情况(如严重封闭、模糊的航道等)以及车道说明中固有的监督信号稀少,车道探测任务仍具有挑战性。因此,普通革命神经网络(CNN)很难在一般场景中训练从原始图像中捕捉微妙的航道特征。在本文中,我们展示了一个名为REsuty Feature-Shift Aggragator(RESA)的新颖模块,以在与普通CNN进行初步地貌提取后丰富车道特征。RASA利用各行和列的强型前端前方,捕捉像素的空间关系。它经常在垂直和水平方向移动切片特征图,使每个像素网络能够收集全球信息。RESA可以通过将切片地图图图图的线索精确地描述具有挑战性的情况。此外,我们提议建立一个双边的Uprodubation Decoder,将高精度和精细的相径直径的地段特征结合到上。在上标阶段,它可以将低分辨率的地段测量结果恢复到精确的轨道基准。