Computing at the edge is increasingly important since a massive amount of data is generated. This poses challenges in transporting all that data to the remote data centers and cloud, where they can be processed and analyzed. On the other hand, harnessing the edge data is essential for offering data-driven and machine learning-based applications, if the challenges, such as device capabilities, connectivity, and heterogeneity can be mitigated. Machine learning applications are very compute-intensive and require processing of large amount of data. However, edge devices are often resources-constrained, in terms of compute resources, power, storage, and network connectivity. Hence, limiting their potential to run efficiently and accurately state-of-the art deep neural network (DNN) models, which are becoming larger and more complex. This paper proposes a novel offloading mechanism by leveraging installed-base on-premises (edge) computational resources. The proposed mechanism allows the edge devices to offload heavy and compute-intensive workloads to edge nodes instead of using remote cloud. Our offloading mechanism has been prototyped and tested with state-of-the art person and object detection DNN models for mobile robots and video surveillance applications. The performance shows a significant gain compared to cloud-based offloading strategies in terms of accuracy and latency.
翻译:边缘的计算机越来越重要,因为生成了大量数据。 这在将所有数据传送到远程数据中心和云层方面构成了挑战,可以对其进行处理和分析。 另一方面,如果能够缓解设备能力、连通性和异质性等挑战,则利用边缘数据对于提供数据驱动和机器学习应用程序至关重要。机器学习应用程序的计算密集程度很高,需要处理大量数据。然而,边缘设备在计算资源、电力、存储和网络连接方面往往受到资源限制。因此,限制其高效和准确运行最新、日益扩大和复杂的先进神经网络(DNNN)模型的潜力。本文提出一种新型卸载机制,利用安装在前沿的基地(尖端)计算资源。拟议机制允许边缘设备卸载大量和密集的工作量,以在边缘而不是使用远程云层。我们的卸载机制已经与状态人和对象探测功能同步进行模拟测试,这些模型正在日益扩大和复杂。