Though deep neural network models exhibit outstanding performance for various applications, their large model size and extensive floating-point operations render deployment on mobile computing platforms a major challenge, and, in particular, on Internet of Things devices. One appealing solution is model quantization that reduces the model size and uses integer operations commonly supported by microcontrollers . To this end, a 1-bit quantized DNN model or deep binary neural network maximizes the memory efficiency, where each parameter in a BNN model has only 1-bit. In this paper, we propose a reconfigurable BNN (RBNN) to further amplify the memory efficiency for resource-constrained IoT devices. Generally, the RBNN can be reconfigured on demand to achieve any one of M (M>1) distinct tasks with the same parameter set, thus only a single task determines the memory requirements. In other words, the memory utilization is improved by times M. Our extensive experiments corroborate that up to seven commonly used tasks can co-exist (the value of M can be larger). These tasks with a varying number of classes have no or negligible accuracy drop-off on three binarized popular DNN architectures including VGG, ResNet, and ReActNet. The tasks span across different domains, e.g., computer vision and audio domains validated herein, with the prerequisite that the model architecture can serve those cross-domain tasks. To protect the intellectual property of an RBNN model, the reconfiguration can be controlled by both a user key and a device-unique root key generated by the intrinsic hardware fingerprint. By doing so, an RBNN model can only be used per paid user per authorized device, thus benefiting both the user and the model provider.
翻译:尽管深神经网络模型在各种应用中表现出杰出的性能,但其庞大的模型规模和广泛的浮动点操作使移动计算平台的部署成为一项重大挑战,特别是在Times设备的互联网上。一个吸引人的解决办法是模型量化,以减少模型大小,并使用由微控制器共同支持的整数操作。为此,一个1位位数的 DNNN 模型或深二进制神经网络将最大程度的记忆效率,BNN 模型中的每个参数只有1位。在本文中,我们提议重新配置BNN(RBNNN),以进一步扩大资源限制的 IoT 设备的存储效率。一般来说,RBNN可以根据需求进行重新配置,以实现任何M (M>1) 的模型大小和使用由微控制器支持的整数,因此只有单项任务才能决定存储要求。换句话说,记忆的利用率通过M. 我们的广泛实验证实,最多七个常用的模型任务(M的值可以更大。这些任务可以由不同数目的班级来进一步提高资源存储模型的存储模型的效率效率。