The Internet of Things (IoT) and smart city paradigm includes ubiquitous technology to extract context information in order to return useful services to users and citizens. An essential role in this scenario is often played by computer vision applications, requiring the acquisition of images from specific devices. The need for high-end cameras often penalizes this process since they are power-hungry and ask for high computational resources to be processed. Thus, the availability of novel low-power vision sensors, implementing advanced features like in-hardware motion detection, is crucial for computer vision in the IoT domain. Unfortunately, to be highly energy-efficient, these sensors might worsen the perception performance (e.g., resolution, frame rate, color). Therefore, domain-specific pipelines are usually delivered in order to exploit the full potential of these cameras. This paper presents the development, analysis, and embedded implementation of a realtime detection, classification and tracking pipeline able to exploit the full potential of background filtering Smart Vision Sensors (SVS). The power consumption obtained for the inference - which requires 8ms - is 7.5 mW.
翻译:互联网(IoT)和智能城市范例包括无处不在的技术,提取背景信息,以便向用户和公民提供有用的服务。这种情景中的一个基本作用往往是计算机视觉应用,需要从特定设备中获取图像。高端摄像头的需要往往使这一过程受到处罚,因为它们是强盗,需要大量计算资源来处理。因此,提供新的低功率低能感应器,采用硬件动作探测等先进特征,对于在IoT域的计算机视觉至关重要。不幸的是,如果高能效,这些感应器可能会使感知性能恶化(例如分辨率、框架率、颜色)。因此,提供具体域的管道通常是为了充分利用这些照相机的潜力。本文介绍了实时探测、分类和跟踪管道的开发、分析和嵌入实施,以便能够充分利用智能视觉感应器(SVS)的全部潜力。为推断而获得的能量消耗量――需要8米-7.5米W。