The huge research interest in cellular vehicle-to-everything (C-V2X) communications in recent days is attributed to their ability to schedule multiple access more efficiently as compared to its predecessor technology, i.e., dedicated short-range communications (DSRC). However, one of the foremost issues still remaining is the need for the V2X to operate stably in a highly dynamic environment. This paper proposes a way to exploit the dynamicity. That is, we propose a resource allocation mechanism adaptive to the environment, which can be an efficient solution for air interface congestion that a V2X network often suffers from. Specifically, the proposed mechanism aims at granting a higher chance of transmission to a vehicle with a higher crash risk. As such, the channel access is prioritized to those with urgent needs. The proposed framework is established based on reinforcement learning (RL), which is modeled as a contextual multi-armed bandit (MAB). Importantly, the framework is designed to operate at a vehicle autonomously without any assistance from a central entity, which, henceforth, is expected to make a particular fit to distributed V2X network such as C-V2X mode 4.
翻译:近日来,对移动车辆到每件(C-V2X)通信(C-V2X)的极大研究兴趣是由于它们能够比其前身技术,即专用短程通信(DSRC)更有效地安排多重接入,然而,最重要的问题之一仍然是V2X需要在高度动态的环境中进行刀刺操作。本文建议了一种利用动态的方法。也就是说,我们建议了一种适应环境的资源分配机制,这可能是一个V2X网络经常遭受的空气界面拥堵的有效解决办法。具体地说,拟议的机制旨在为高碰撞风险的车辆提供更高的传输机会。因此,频道接入被优先提供给有迫切需要的车辆。拟议框架是根据强化学习(RL)建立的,该强化学习模式以一个背景多武装的多臂乐队(MAB)为模范。 重要的是,框架的设计是在没有中央实体的任何援助的情况下,在汽车上自主运行,因此,预计中央实体将特别适合诸如C-V2X模式4等分布的V2X网络。