Along with the rapid development in the field of artificial intelligence, especially deep learning, deep neural network applications are becoming more and more popular in reality. To be able to withstand the heavy load from mainstream users, deployment techniques are essential in bringing neural network models from research to production. Among the two popular computing topologies for deploying neural network models in production are cloud-computing and edge-computing. Recent advances in communication technologies, along with the great increase in the number of mobile devices, has made edge-computing gradually become an inevitable trend. In this paper, we propose an architecture to solve deploying and processing deep neural networks on edge-devices by leveraging their synergy with the cloud and the access-control mechanisms of the database. Adopting this architecture allows low-latency DNN model updates on devices. At the same time, with only one model deployed, we can easily make different versions of it by setting access permissions on the model weights. This method allows for dynamic model licensing, which benefits commercial applications.
翻译:随着人工智能领域的迅速发展,特别是深层学习,深神经网络应用在现实中越来越受欢迎。为了能够承受主流用户的沉重负担,应用技术对于将神经网络模型从研究到生产至关重要。在生产中部署神经网络模型的两个流行计算型号中,有两个是云计算和边位计算。通信技术的最近发展,加上移动设备数量的大量增加,使边缘计算逐渐成为一个不可避免的趋势。在本文中,我们提出了一个结构,通过利用它们与云和数据库的存取控制机制的协同作用来解决边缘装置的深神经网络的部署和处理。采用这一结构,可以使低延迟的DNNN模型在设备上更新。同时,只要使用一个模型,我们就可以很容易地通过在模型重量上设置访问许可来制造不同版本。这种方法允许动态模型许可,这有利于商业应用。