To preserve user privacy while enabling mobile intelligence, techniques have been proposed to train deep neural networks on decentralized data. However, training over decentralized data makes the design of neural architecture quite difficult as it already was. Such difficulty is further amplified when designing and deploying different neural architectures for heterogeneous mobile platforms. In this work, we propose an automatic neural architecture search into the decentralized training, as a new DNN training paradigm called Federated Neural Architecture Search, namely federated NAS. To deal with the primary challenge of limited on-client computational and communication resources, we present FedNAS, a highly optimized framework for efficient federated NAS. FedNAS fully exploits the key opportunity of insufficient model candidate re-training during the architecture search process, and incorporates three key optimizations: parallel candidates training on partial clients, early dropping candidates with inferior performance, and dynamic round numbers. Tested on large-scale datasets and typical CNN architectures, FedNAS achieves comparable model accuracy as state-of-the-art NAS algorithm that trains models with centralized data, and also reduces the client cost by up to two orders of magnitude compared to a straightforward design of federated NAS.
翻译:为了保护用户隐私,同时允许移动情报,已提议采用各种技术来培训关于分散数据的深神经网络;然而,关于分散数据的训练使神经结构的设计变得相当困难,在设计和部署不同神经结构以建立不同移动平台时,这种困难进一步加大;在这项工作中,我们提议对分散培训进行自动神经结构搜索,作为称为联邦神经结构搜索的新的DNN培训模式,称为联邦神经结构搜索,即联合NAS。为了应对客户计算和通信资源有限这一主要挑战,我们介绍了FedNAS,这是高效联合NAS的一个高度优化的框架。FedNAS充分利用了在建筑搜索过程中示范候选人培训不足的关键机会,并纳入了三个关键的优化:对部分客户的平行候选人培训,业绩差的提前丢弃候选人和动态的圆数。在大型数据集和典型CNN架构上进行了测试,FedNAS实现了模型的可比模型精确度,即用集中数据培训模型的状态的NAS算法,并且将客户成本降低到两个数量级,比NAS的直截型设计要低两个级。