In this paper, we consider a wireless network of smart sensors (agents) that monitor a dynamical process and send measurements to a base station that performs global monitoring and decision-making. Smart sensors are equipped with both sensing and computation, and can either send raw measurements or process them prior to transmission. Constrained agent resources raise a fundamental latency-accuracy trade-off. On the one hand, raw measurements are inaccurate but fast to produce. On the other hand, data processing on resource-constrained platforms generates accurate measurements at the cost of non-negligible computation latency. Further, if processed data are also compressed, latency caused by wireless communication might be higher for raw measurements. Hence, it is challenging to decide when and where sensors in the network should transmit raw measurements or leverage time-consuming local processing. To tackle this design problem, we propose a Reinforcement Learning approach to learn an efficient policy that dynamically decides when measurements are to be processed at each sensor. Effectiveness of our proposed approach is validated through a numerical simulation with case study on smart sensing motivated by the Internet of Drones.
翻译:在本文中,我们考虑建立一个由智能传感器(试剂)组成的无线网络,对动态过程进行监测,并将测量结果发送到一个进行全球监测和决策的基础站。智能传感器配备了感测和计算设备,可以发送原始测量数据,或者在传输前处理原始测量数据。受约束的试剂资源可以产生一种基本的潜伏性-准确性交换。一方面,原始测量数据不准确,但生产速度很快。另一方面,资源受限制的平台的数据处理可以以不可忽略的计算延迟性为代价进行准确测量。此外,如果处理的数据也压缩,则由无线通信造成的潜伏对于原始测量而言可能更高。因此,要决定网络中的传感器何时和何处传输原始测量数据,或者利用当地耗时的处理方法。为了解决这一设计问题,我们建议采用强化学习方法,学习一种高效的政策,以动态地决定每个传感器将何时进行测量。我们提议的方法的有效性通过数字模拟得到验证,通过对Dones互联网驱动的智能感测进行个案研究。