The success of deep learning in recent years has lead to a rising demand for neural network architecture engineering. As a consequence, neural architecture search (NAS), which aims at automatically designing neural network architectures in a data-driven manner rather than manually, has evolved as a popular field of research. With the advent of weight sharing strategies across architectures, NAS has become applicable to a much wider range of problems. In particular, there are now many publications for dense prediction tasks in computer vision that require pixel-level predictions, such as semantic segmentation or object detection. These tasks come with novel challenges, such as higher memory footprints due to high-resolution data, learning multi-scale representations, longer training times, and more complex and larger neural architectures. In this manuscript, we provide an overview of NAS for dense prediction tasks by elaborating on these novel challenges and surveying ways to address them to ease future research and application of existing methods to novel problems.
翻译:近年来深层学习的成功导致对神经网络结构工程的需求不断增长,因此,旨在以数据驱动而不是人工方式自动设计神经网络结构的神经结构搜索(NAS)已发展成为一个广受欢迎的研究领域,随着跨结构的权重共享战略的出现,NAS已经适用于更广泛的问题。特别是,目前有许多关于计算机愿景中密集的预测任务的出版物需要像素水平的预测,例如语义分解或物体探测。这些任务带来了新的挑战,例如高分辨率数据导致的更高的记忆足迹、学习多尺度的演示、更长的培训时间以及更复杂和更大的神经结构。在这个手稿中,我们概述了NAS,通过阐述这些新的挑战并探索解决这些问题的方法,以方便未来研究和应用现有方法解决新问题。