Designing convolutional neural networks (CNN) models for mobile devices is challenging because mobile models need to be small and fast, yet still accurate. Although significant effort has been dedicated to design and improve mobile models on all three dimensions, it is challenging to manually balance these trade-offs when there are so many architectural possibilities to consider. In this paper, we propose an automated neural architecture search approach for designing resource-constrained mobile CNN models. We propose to explicitly incorporate latency information into the main objective so that the search can identify a model that achieves a good trade-off between accuracy and latency. Unlike in previous work, where mobile latency is considered via another, often inaccurate proxy (e.g., FLOPS), in our experiments, we directly measure real-world inference latency by executing the model on a particular platform, e.g., Pixel phones. To further strike the right balance between flexibility and search space size, we propose a novel factorized hierarchical search space that permits layer diversity throughout the network. Experimental results show that our approach consistently outperforms state-of-the-art mobile CNN models across multiple vision tasks. On the ImageNet classification task, our model achieves 74.0% top-1 accuracy with 76ms latency on a Pixel phone, which is 1.5x faster than MobileNetV2 (Sandler et al. 2018) and 2.4x faster than NASNet (Zoph et al. 2018) with the same top-1 accuracy. On the COCO object detection task, our model family achieves both higher mAP quality and lower latency than MobileNets.
翻译:设计移动设备同生神经网络模型(CNN)具有挑战性,因为移动模型需要小而快速,但仍然准确。尽管在设计和改进所有三个层面的移动模型方面已经付出了巨大的努力,但当有如此众多的建筑可能性需要考虑时,手工平衡这些权衡是具有挑战性的。在本文件中,我们提议为设计资源限制的移动CNN模型采用自动神经结构搜索方法。我们提议在主要目标中明确纳入潜伏信息,以便搜索能够找到一个在精确度和延缓度之间实现良好权衡的模型。与以往的工作不同,在以往的工作中,通过另一个往往不准确的代理(如FLOPS)来考虑移动内嵌模型,我们在实验中直接测量真实世界的推导力,方法是在特定平台上(如Pixel电话)执行模型。为了进一步在灵活性和搜索空间大小之间取得正确的平衡,我们提议建立一个新的因数级搜索空间模型空间,允许在整个网络中实现分层多样性。实验结果显示,我们的方法始终优于S-Art-al-ality liental liveral listal listal lifalalal ex lax salal ex lax sal lax) lax lax sal lax sal lax sal lax lax lax lax lax lax lax lax lax lax lax lax lax lax lax lax lax lax lax lax lax